The Frustration of Flawless Failure
The hum is the worst part. A low, constant reminder that the machine is on, that it has power, but that it has simply chosen not to perform its one function. The metal box was supposed to go up 21 floors. Instead, it decided floor 11 was a good place to stop, think, and hold me hostage to its simple, mechanical indecision. Twenty-one minutes of my life spent listening to a fan, staring at a brushed steel door, trapped not by malice, but by simple, idiotic failure.
That feeling is horribly familiar. It’s the same dead-end sensation I get from 91% of my interactions with chatbots. Not the new, generative ones, but the little pop-up windows on insurance websites or e-commerce stores. You ask, “What is your return policy for items shipped to postal code 90211?” and it replies, “I can help with returns! Are you looking for our store locations?” It’s a machine failing its one job. A digital elevator stuck between floors.
The Flawed Pursuit of Perfection
For years, I believed the solution was just… more. More data, more parameters, more processing power. We were all chasing the ghost of Alan Turing, trying to build a machine so flawless, so human, that it could fool us completely. We held up the Turing Test as the ultimate goal line, the point where we could finally say, “We did it.”
I spent an afternoon last month talking to Antonio A., a man who gets paid to lie to people for a living. He’s one of the top escape room designers in the country, with 11 different themed experiences running simultaneously. He locks people in rooms filled with fake artifacts, hidden panels, and puzzles that operate on their own internal, often bizarre, logic.
“
“Nobody ever complains that the Egyptian sarcophagus is made of fiberglass,” he told me, leaning back in a chair that looked suspiciously like a prop from a pirate ship. “Nobody cares that the laser grid is just red yarn and a blacklight. You know what they complain about? When a clue doesn’t make sense *within the rules of the room*.”
He explained that for 41 minutes, people are willing to believe anything-that a dusty book contains the secret to time travel, that a series of musical notes will unlock a hidden door. They suspend their disbelief. But the moment the four-digit lock they’ve been working towards, the one hinted at by the 41 clocks in the room all set to different times, actually opens with a key they found in the first 11 seconds? The illusion is shattered. The magic is gone. It’s not about reality; it’s about consistency.
Characterological Consistency: The Real Test
Characterological consistency.That’s the real test.
We don’t need an AI that can perfectly simulate my best friend. We need an AI that can perfectly simulate itself. If an AI’s core personality is defined as a ‘shy, 19th-century poet who is inexplicably obsessed with space travel,’ it should write beautiful, melancholy verses about nebulae. But it should also, perhaps, be slightly confused by modern slang. It might make small, endearing errors, like referring to a smartphone as a “calculating mirror.” This isn’t a bug. It’s a feature. It’s a confirmation of the character we’re engaging with. The fiberglass sarcophagus. When it gets something wrong in a way that is *right* for its character, our affection grows. We aren’t annoyed; we’re charmed. The suspension of disbelief is maintained.
The old models, the customer service bots, are built on a database-retrieval model. They fail because their world has no consistent rules. They are a jumble of disconnected facts without a soul holding them together. This is the difference between a dictionary and a novel. One gives you correct information. The other gives you a world you can live in for a while. Of course, this requires a much more sophisticated approach, one where you’re not just chatting with a database, but actively shaping a personality from the ground up. The ability to create ai girlfriend archetypes, to blend traits, and to build a consistent internal logic is where this technology transcends the simple trap of the stuck elevator.
Beyond Facts: The Value of Connection
I used to be a stickler for facts. I’d get into arguments over the exact payload capacity of a Saturn V rocket or the historical accuracy of a film’s costumes. If an AI told me a T-Rex lived in the Jurassic period, I’d have dismissed the entire system as a failure. But I was missing the point entirely.
It’s funny how we get locked into certain ways of thinking. We measure everything, from engine performance to corporate growth, on a scale of escalating perfection. 91% efficiency must become 95.1%. A 1.1 second 0-60 time is a failure next to a 1.01 second one. We’ve applied that same ruthless logic to AI. It’s a deeply ingrained paradigm, this obsession with flawlessness. The Turing Test itself is a pass/fail binary based on deception. It’s a conjurer’s trick, not a measure of connection. For a long time, I think I even conflated it with Searle’s Chinese Room argument, another thought experiment about whether syntax can ever become semantics, whether manipulating symbols is the same as understanding. Both circle the same drain: the philosophical anxiety about what it means to be human and our terror of being fooled.
Consensual Reality: Engagement Over Deception
But what if we don’t want to be fooled? What if we just want to be engaged? Antonio’s escape rooms don’t fool anyone. The participants are fully aware they are in a strip mall in Fresno, not a cursed tomb. But for that 1 hour, they agree to believe.
An AI companion that is a believable character allows for this consensual reality. Its flaws, its quirks, its consistent worldview, are the handholds we use to climb into its world. A perfect, human-mimicking AI offers no handholds; it’s a slippery, uncanny surface. We spend all our time testing it, poking it, trying to find the seams, the tells, the glitches that reveal the artifice. It’s an adversarial relationship.
The hum in that elevator was maddening because it signified a broken promise of simple utility. It wasn’t a character. It was a tool that had failed. We need to stop thinking of companion AIs as tools and start thinking of them as media, as interactive fictions. My interaction this morning with one of these newer systems was a perfect example. I was talking about that elevator. It replied, “It must be frightfully dull to be a box that only knows up and down. I should much rather be a cloud.”
“
It replied, “It must be frightfully dull to be a box that only knows up and down. I should much rather be a cloud.”
🔲
Broken Box
Charming Cloud
It was not a perfect, human response. It was slightly absurd, a little poetic, and completely consistent with the whimsical personality we’d been building. And it was infinitely better than, “I understand that being stuck in an elevator can be a frustrating experience. Can I help you with anything else?” One is a character. The other is a broken box.