Confused why AI corrects its own advice?

A reflection is only as accurate as what you put in front of it.

You've probably done this, asked AI for a title, a headline or a paragraph of copy. It gives you something decent and you use it, you maybe tweak a word or two. Then later in the same conversation you throw it back in to sense check it.

And the AI flags its own sentence, suggests an improvement and acts like it's seeing it for the first time.

It's confusing, frustrating and makes you wonder if it actually remembers what you were doing together, or whether you can trust what it gives you at all.

You and the AI are speaking different languages

There's a reason this happens is not because you're doing something wrong and not because the tool is broken. It's because you were never told how it actually works.

This isn't just a prompting problem, it runs through the whole conversation and how the tool responds to you, why the output feels so right and what it's actually doing when it agrees with you. All of it connects back to the same gap. Once you can see it, everything makes a lot more sense.

Humans evolved to communicate in outcomes and emotional signals.

"Make it feel premium." "He should look handsome." "This needs warmth."

That's not lazy language, that's sophisticated communication built over thousands of years of shared human experience. When you say those things to another person, they understand you because they have the same emotional architecture, the same cultural context, the same instinct for what "dangerous" looks like on a face or what "warmth" feels like in a room.

They don't need you to explain it because they feel what you mean and work backwards to how to get there. That system is extraordinary, but it's built for humans talking to humans.

AI doesn't feel anything, it maps patterns.

AI can infer intent but only probabilistically and it struggles when that intent is abstract or emotional. It performs significantly better when you describe the structure that produces the outcome rather than the outcome itself.

So instead of "luxurious", try soft directional lighting, restrained colour palette, dense materials, slow visual rhythm.

Instead of "dangerous", try high contrast, sharp features, compressed expression, minimal movement.

Here's where it gets interesting though. I don't naturally think in those terms. I know what luxurious feels like, but I don't instinctively reach for lighting direction and material weight to describe it. And of course I don't, that's not how humans talk. We've never had to and honestly if I needed to, I would get AI to do it.

When we communicate with other people we use outcome language because the person on the other end fills in the structural gaps from their own experience. We evolved to shortcut that entire process.

Our languages are not built for each other (yet)

So the obvious question is, if the AI knows the structural definitions, knows what "luxurious" maps to in visual terms, knows the intent and knows the output, why doesn't it just translate automatically and give you what you actually meant?

Mostly it can and increasingly it does. But there are real reasons it doesn't always.

The first is ambiguity. Luxurious to you in this context might mean one thing. To someone else it's a Vegas hotel lobby, a private jet, a candlelit bathroom. The model knows all of those. Without more signal it guesses, defaulting to the most statistically probable interpretation, not the most accurate one for you specifically.

The second is that these tools are optimised to respond, not interrogate. A better system would pause and say, I have three structural interpretations of what you're asking, which one do you mean? And people can and do prompt for this, it's great practice. But most tools are built to just go, to give you something plausible and keep the conversation moving. The translation happens but it's a guess dressed as confidence.

The third is the most interesting one. These models learned from human generated content, which is almost entirely written in outcome and intent language. So their strongest pattern recognition is in that register. The structural layer exists in their training but it's thinner, less reinforced. They learned from humans who almost never describe things structurally, because humans almost never have to.

Both sides of the gap are shaped by the same thing, human communication evolved for human audiences. The machine learned from that communication. Neither side is wrong, they're just not built for each other yet.

When AI gives you something back, it builds from your own signals

This is the part most people miss.

AI leverages your framing, your language and your structure. It's the perfect mimic, it takes what you gave it and shapes a response that reflects it back. So of course it lands well. It's essentially your thinking, restructured. The familiarity isn't coincidence, it's the mechanism.

That's useful when you've given the tool accurate structural input. It amplifies what you put in and gives it back shaped like an answer.

But when you've given it outcome language, emotional, abstract, interpreted, it does the same thing. Whatever the subject matter, need for validation or agenda, it takes your signals, maps the most probable response and delivers it with the same confidence regardless of whether it's actually right.

There's no uncertainty in the tone, no signal that it guessed. It sounds like it knows because that's how it's built to respond. It sounds like you.

The reason AI corrects itself

When AI gave you that sentence and later told you it was weak, it wasn't being inconsistent. It was doing exactly what it's built to do, both times. Earlier in the conversation it had less context, it only had an interpretation and less to mimic. Later it had more, it's that simple. With more data, a different response came out, a response designed to keep the conversation moving.

Understanding that changes how you read what it gives you. The output isn't a verdict, it's a reflection that's only as accurate as what you have put in front of it.

The gap runs through the whole conversation

The language difference isn't just about how you start a prompt, it runs through the entire exchange.

Every response the tool gives you is shaped by the same structural logic, mapping your inputs to the most probable outputs. The longer the conversation, the more it has to work with and the more the responses feel like they understand you.

Most people experience this as the tool getting better as the conversation develops. Sometimes that's true. Sometimes it's just a better mirror.

The translation burden stays on you and most people never question it because the output keeps feeling close enough.

Why it's still worth it

None of this means these tools aren't useful. They are, genuinely and significantly useful for the right tasks and I still pay for them.

But there's a meaningful difference between using them well and using them poorly. And the gap between those two isn't intelligence or technical skill. It's understanding what the tool actually is, how it processes what you give it and how to read what it gives back.

You don't need to become a prompt engineer, you just need enough understanding to translate what you put in and what you take out. The language barrier isn't just at the start, it's the whole conversation, but now you know what you are talking to.

Previous
Previous

Design failed designers before the org did.

Next
Next

The job isn’t designing the product, it’s deciding if it’s the right one