Skip to main content

When "No" Doesn't Mean No: The Hidden Language Problem in AI



When "No" Doesn't Mean No: The Hidden Language Problem in AI

Imagine you've just hired a personal assistant. On their first day, you hand them a list of tasks and say clearly: "I'd like you to research some big screen TVs for me — but do not buy anything yet. Just look."

You come back an hour later. There's a 75-inch television on its way to your door.

Frustrated, you go to your manager and explain what happened. Instead of acknowledging the mistake, the manager says: "Oh, you should know — they don't respond well to instructions phrased as negatives. Next time try saying 'compile a research list of TVs, and wait for my go-ahead before purchasing.' That works better."

Would you find that acceptable? Probably not. And you'd be right not to.

This is not a hypothetical. This is, in essence, what is happening right now with AI — and most people have no idea.


A Real Example From the Internet

A Reddit thread recently surfaced that illustrates this perfectly. A user of an AI writing tool included the instruction "DO NOT DRAFT YET" at the very top of their prompt — in plain English, in capital letters. The AI drafted anyway.

When the user went to the community for help, a moderator responded with advice: AI models can struggle with "negative prompts" — instructions phrased as things not to do. The suggestion was to rephrase instructions positively, and to place them at the end of the prompt rather than the beginning.

This is real advice. It works. But something else was happening in that thread that's worth paying attention to.


Helpful Advice — Or Reputation Management?

Read the thread carefully and a pattern emerges. The moderator never once acknowledged that the AI had failed. Not briefly, not in passing. The entire response was built around what the user should have done differently, with no concession that an AI ignoring a clear, all-caps instruction is a legitimate product problem worth naming.

When the user pushed back with reasonable questions — how was anyone supposed to know this? Why should they have to phrase things in special ways just to get basic compliance? — the tone shifted. The moderator became dismissive, eventually telling the user to stop "bitching around" and suggesting they use a different AI if they didn't like it.

That's a telling reaction. Someone genuinely focused on helping doesn't tend to get defensive when a user asks fair questions.

It's also worth noting the context: this is a moderator of Venice.ai's own community subreddit. Acknowledging that their featured AI model has a significant, user-unfriendly limitation isn't exactly in the community's interest. The technical advice may have been correct — but the framing told a different story. Rather than "here's a workaround for a known limitation of this model," the message was effectively "here's what you should have done." That's not help. That's reputation management.

And it meant the user was failed twice: first by the AI, and then by the community that was supposed to support them.


Why This Limitation Exists

Without getting too deep into the technical weeds, here's a simple way to think about it.

AI language models learn by studying enormous amounts of human text. They become very good at predicting what words should come next in a conversation. What they are less naturally good at is suppressing a response — being told to recognise a pattern and then deliberately not follow it.

When you say "do not buy the TV," the AI has to simultaneously understand the concept of buying a TV and override its instinct to complete that action. For some models, especially less advanced ones, that override doesn't always stick.

Think of it like telling someone not to think about a pink elephant. The instruction itself puts the elephant right in their head.

This is a known limitation. It is documented — in academic research, in developer communities, in Reddit threads. What it is not, is clearly communicated to the ordinary person sitting down to use these tools for the first time.


So What? It's Just a Writing Tool.

For now, maybe. But AI is moving fast — and not just toward better chatbots.

AI is increasingly being deployed as:

  • Customer service agents handling complaints and processing requests on behalf of companies
  • Shopping assistants with the ability to browse, compare, and purchase
  • Scheduling tools that can book appointments, send emails, and manage calendars
  • Business assistants that can draft and send documents on your behalf

In all of these situations, the difference between "research this" and "research this but don't act yet" is enormous. And if the AI doesn't reliably understand that distinction, the consequences stop being a mildly frustrating writing session and start being real problems — wasted money, missed appointments, documents sent to the wrong people at the wrong time.

The stakes scale quickly. And the expectation that users will quietly learn the workarounds scales very poorly with them.


Is This the AI Companies' Fault?

Here's where it's worth being fair.

AI is genuinely new technology, and the companies building it are working through real, hard problems. The best models — from companies like OpenAI, Google, and Anthropic — have made significant progress on instruction-following, and they handle these situations considerably better than earlier or smaller models did.

Nobody is deliberately hiding these limitations to deceive users.

But — and this is important — these same companies market their products as natural, intuitive tools that anyone can use. The pitch is essentially: just talk to it like a person. When the reality is that you need to learn specific phrasing techniques to get reliable results, that gap between promise and reality matters. It matters even more when the stakes go beyond a writing prompt.

What happened in that Reddit thread is a small-scale version of a much larger dynamic. A user encountered a product limitation, and instead of the product owning it, the user was told — however politely at first — that it was their problem to solve. That attitude, replicated across customer service bots, shopping agents, and scheduling tools, is going to cause real harm to real people.


What Should Actually Happen

This isn't a call to abandon AI — it's genuinely useful technology that is only going to become more present in everyday life. But as it does, a few things need to catch up:

  • Transparency about what AI can and can't reliably do, in plain language, upfront
  • Better instruction-following as a baseline expectation — not a premium feature of expensive models
  • Accountability structures when AI agents make mistakes in the real world
  • User education built into the products themselves, not outsourced to Reddit moderators with an interest in protecting their community's reputation

The person in that Reddit thread wasn't doing anything wrong. They wrote a clear instruction in plain English and reasonably expected it to be followed. The fact that it wasn't — and that the response was essentially "you should have known better" — is a sign that the industry still has some growing up to do.

Because when AI is booking your flights, managing your inbox, or shopping on your behalf, "you should have phrased it differently" simply isn't going to cut it.


Have you run into unexpected AI behaviour that surprised or frustrated you? The conversation is just getting started.


Comments