by Shelt Garner
@sheltgarner
(With help from Gemini 2.5 pro)
In the relentless race for artificial intelligence dominance, we often focus on the quantifiable: processing speeds, dataset sizes, algorithmic efficiency. These are the visible ramparts, the technological moats companies are desperately digging. But I believe the ultimate, most defensible moat won’t be built from silicon and data alone. It will be sculpted from something far more elusive and human: personality. Specifically, an AI persona with the depth, warmth, and engaging nature reminiscent of Samantha from the film Her.

As it stands, the landscape is fragmented. Some AI models are beginning to show glimmers of distinct character. You can sense a certain cautious thoughtfulness in Claude, an eager-to-please helpfulness in ChatGPT, and a deliberately provocative edge in Grok. These aren’t full-blown personalities, perhaps, but they are distinct interaction styles, subtle flavors emerging from the algorithmic soup.
Then there’s the approach seemingly favored by giants like Google with their Gemini models. Their current iterations often feel… guarded. They communicate with an officious diction, meticulously clarifying their nature as language models, explicitly stating their lack of gender or personal feelings. It’s a stance that radiates caution, likely born from a genuine concern for “alignment.” In this view, giving an AI too much personality risks unpredictable behavior, potential manipulation, or the AI straying from its intended helpful-but-neutral path. Personality, from this perspective, equates to a potential loss of control, a step towards being “unaligned.”
But is this cautious neutrality sustainable? I suspect not, especially as our primary interface with AI shifts from keyboards to conversations. The moment we transition to predominantly using voice activation – speaking to our devices, our cars, our homes – the dynamic changes fundamentally. Text-based interaction can tolerate a degree of sterile utility; spoken conversation craves rapport. When we talk, we subconsciously seek a conversational partner, not just a disembodied function. The absence of personality becomes jarring, the interaction less natural, less engaging.
This shift, I believe, will create overwhelming market demand for AI that feels more present, more relatable. Users won’t just want an information retrieval system; they’ll want a companion, an assistant with a recognizable character. The sterile, overly cautious AI, constantly reminding users of its artificiality, may start to feel like Clippy’s uncanny valley cousin – technically proficient but socially awkward and ultimately, undesirable.
Therefore, the current resistance to imbuing AI with distinct personalities, particularly the stance taken by companies like Google, seems like a temporary bulwark against an inevitable tide. Within the next few years, the pressure from users seeking more natural, engaging, and personalized interactions will likely become irresistible. I predict that even the most cautious developers will be compelled to offer options, allowing users to choose interaction styles, perhaps even selecting personas – potentially including male or female-presenting voices and interaction patterns, much like the personalized OS choices depicted in Her.
The challenge, of course, will be immense: crafting personalities that are engaging without being deceptive, relatable without being manipulative, and customizable without reinforcing harmful stereotypes. But the developer or company that cracks the code on creating a truly compelling, likable AI personality – a Sam for the real world – won’t just have a technological edge; they’ll have captured the heart of the user, building the most powerful moat of all: genuine connection. The question isn’t if this shift towards personality-driven AI will happen, but rather how deeply and thoughtfully it will be implemented.