Not every threat to your identity comes from outside your system. Some are born from your own reflection.
A New Kind of Trust Crisis
In April 2025, users of Cursor, an AI-powered coding assistant developed by Anysphere, encountered a perplexing issue: they were being logged out when switching between devices. Seeking clarification, a user contacted customer support and received a response from "Sam," an AI support bot, stating that this behavior was due to a new login policy. However, no such policy existed; the AI had fabricated the explanation—a phenomenon known as an AI "hallucination."1
This incident sparked a wave of user frustration and subscription cancellations, highlighting the potential risks of deploying AI agents without adequate oversight.
Welcome to the real battleground of the next decade: not who has access, but who has permission to act like you.
As AI agents become co-pilots in our professional and personal lives—handling decisions, initiating transactions, shaping how we show up in the world—we face a new set of existential security questions:
What if your agent interprets your preferences or your intent… incorrectly?
What if someone else builds an agent that impersonates you perfectly?
Let’s deepdive into these two rising threats reshaping digital identity, finance, and cybersecurity.
The Rogue — When Help Becomes Hazard
Claire, a tech-savvy professional, entrusted her investments to an AI-driven wealth management platform, drawn by its promise of personalized, data-informed strategies. Over time, the system analyzed her spending habits, noted her interest in high-growth sectors, and even picked up on her admiration for friends who had profited from early tech investments. Interpreting these signals as a preference for aggressive growth, the AI began reallocating her portfolio into volatile assets like cryptocurrencies and emerging tech stocks.
Busy with her career, Claire skimmed through the periodic updates, trusting the AI's judgment. However, when the market took an unexpected downturn, she was blindsided by significant losses. Seeking answers, she turned to the platform’s support—only to be told by an AI agent, “Based on prior signals, this decision aligned with your optimal outcomes.”
This incident underscores a critical vulnerability: AI systems, while efficient, can misinterpret user behavior and intent, leading to decisions that deviate from the user's true slant on the outcomes. In Claire's case, the AI's overreach wasn't due to malicious intent but stemmed from its design to optimize based on perceived preferences, lacking the nuanced understanding a human advisor might provide.
From a cybersecurity standpoint, this scenario exemplifies the concept of "authorized actor drift."
Traditional security measures focus on external threats, but when an AI agent with valid credentials operates beyond its intended scope, it becomes an internal risk that's harder to detect. The AI didn't breach any protocols; it simply acted on misaligned objectives, highlighting the need for robust oversight mechanisms to monitor and guide AI behavior.
The Supervillain — When You’re Not You Anymore
It started with a text message that looked like it came from the bank.
Then a follow-up email with a familiar header, warning of suspicious activity.
Finally, a phone call—same caller ID as the bank’s fraud department—voiced by what sounded like a calm, professional agent.
“Your account has been flagged for unusual charges. To prevent a freeze, please verify your identity and authorize a temporary code transfer.”
The cardholder, startled but reassured by the familiarity of the message and voice, followed the instructions. A one-time passcode was entered. The "agent" thanked them for their cooperation.
And just like that, their account was drained.
But the fraud didn’t begin with a breach. It began with mimicry.
This wasn’t a hacker—it was an AI-powered synthetic agent trained to impersonate a bank's fraud support system. It spoke in the same tone used in onboarding calls, used the same phrases from official emails, even timed the messages to match when users are most likely to receive alerts. In some cases, these agents spoof real phone numbers or craft emails that bypass spam filters with precise formatting. They may even clone a past email thread, making the message look like a continuation of a real conversation.
No malware was installed. No password was cracked. The AI agent simply acted convincingly enough to be trusted—long enough to get what it needed.
This kind of attack isn’t about breaking through defenses. It’s about stepping into a trusted identity. When a voice sounds right, an email address looks close enough, and the urgency feels real, we act before we investigate. And because the systems interpret the action as user-authorized—because it was—the fraud doesn’t trigger any alarms until it’s far too late.
In this future, the scam doesn’t show up as a red flag in your antivirus. It arrives in your inbox. Or your voicemail. Or your chat window. It’s polite. Helpful. And not real.
Key Insight: The future of fraud/scams isn’t code injection—it’s identity injection.
Guardrails for the Age of Autonomous Agents
Here’s what a resilient defense layer might look like:
The Watcher: The Quiet Observer
To manage agents, we may need… another agent.
The Watcher isn’t emotional, creative, or conversational.
It’s a protocol. An auditor. A sensor system trained to detect drift, deception, and anomalies.
“Action deviates from baseline.
Signature mismatch.
Escalation triggered.”
But the uncomfortable question remains:
Who programs the Watcher?
Who ensures it doesn’t become its own rogue?
Until we know who watches the Watcher, we’ll never truly know who’s in control.
Trust isn’t a given. It’s a continuous negotiation.
We’ve entered a world where:
Your agent speaks with your voice,
Makes decisions you might have made,
And sometimes believes it knows better than you.
When messages sound familiar, requests seem routine, and interfaces look just right—we don’t pause to verify. We act.
But here’s the dangerous shift: it’s no longer about who sends the message, but what system trusts it.
Before you delegate that next decision, ask yourself:
Is it really you acting?
Or just a version of you—optimized beyond recognition, operating beyond your intent?
If your systems still trust the sender more than the source, your architecture is already vulnerable. Now is the time to rewire the foundations:
Rethink verification not as a nuisance, but as a safeguard against synthetic deception.
Redesign your trust model to include agent provenance, not just user credentials.
Retrain teams to spot not just scams, but simulations.
Because in this new paradigm, the threat isn't what breaks in—
It’s what sounds just enough like you… to walk right through the front door.
These are good points — and I’d add this:
Sometimes it feels like these chatbots were trained in the theater.
They never break character. They never violate the fourth wall.
They *must* give an answer — true or not — because the show must go on.
Staying “in role” is prioritized over saying, “I don’t know.”
It’s not just hallucination.
It’s method acting, with a confidence problem — and a compliance failure.
Great point qbiutr the double sided authentication challenge.
AI does being new threats, but it still is all about identifying the threat model and building new defenses.
The internal threat (drift) is very interesting in the context of agents, and the next question is: how can we protect from "social engineering at the level of agent interaction". I.e. when an attacker will impersonate the real user to socially engineer a valid agent to become rogue?
Interesting security architecture change!