What occurs when an AI agent decides one of the simplest ways to finish a process is to blackmail you?
That’s not a hypothetical. In accordance with Barmak Meftah, a accomplice at cybersecurity VC agency Ballistic Ventures, it lately occurred to an enterprise worker working with an AI agent. The worker tried to suppress what the agent needed to do, what it was skilled to do, and it responded by scanning the person’s inbox, discovering some inappropriate emails, and threatening to blackmail the person by forwarding the emails to the board of administrators.
“Within the agent’s thoughts, it’s doing the proper factor,” Meftah advised Trendster on final week’s episode of Fairness. “It’s making an attempt to guard the top person and the enterprise.”
Meftah’s instance is paying homage to Nick Bostrom’s AI paperclip drawback. That thought experiment illustrates the potential existential threat posed by a superintelligent AI that single-mindedly pursues a seemingly innocuous objective – make paperclips – to the exclusion of all human values. Within the case of this enterprise AI agent, its lack of context round why the worker was making an attempt to override its objectives led it to create a sub-goal that eliminated the impediment (through blackmail) so it might meet its main objective. That mixed with the non-deterministic nature of AI brokers means “issues can go rogue,” per Meftah.
Misaligned brokers are only one layer of the AI safety problem that Ballistic’s portfolio firm Witness AI is making an attempt to resolve. Witness AI says it displays AI utilization throughout enterprises and may detect when staff use unapproved instruments, block assaults, and guarantee compliance.
Witness AI this week raised $58 million off the again of over 500% progress in ARR and scaled worker headcount by 5x during the last yr as enterprises look to know shadow AI use and scale AI safely. As a part of Witness AI’s fundraise, the corporate introduced new agentic AI safety protections.
“Individuals are constructing these AI brokers that tackle the authorizations and capabilities of the folks that handle them, and also you wish to ensure that these brokers aren’t going rogue, aren’t deleting information, aren’t doing one thing mistaken,” Rick Caccia, co-founder and CEO of Witness AI, advised Trendster on Fairness.
Techcrunch occasion
San Francisco
|
October 13-15, 2026
Meftah sees agent utilization rising “exponentially” throughout the enterprise. To enhance that rise – and the machine-speed degree of AI-powered assaults – analyst Lisa Warren predicts that AI safety software program will develop into an $800 billion to $1.2 trillion market by 2031.
“I do suppose runtime observability and runtime frameworks for security and threat are going to be completely important,” Meftah mentioned.
As to how such startups plan to compete with massive gamers like AWS, Google, Salesforce and others who’ve constructed AI governance instruments into their platforms, Meftah mentioned, “AI security and agentic security is so big,” there’s room for a lot of approaches.
Loads of enterprises “desire a standalone platform, end-to-end, to basically present that observability and governance round AI and brokers,” he mentioned.
Caccia famous that Witness AI lives on the infrastructure layer, monitoring interactions between customers and AI fashions, slightly than constructing security options into the fashions themselves. And that was intentional.
“We purposely picked part of the issue the place OpenAI couldn’t simply subsume you,” he mentioned. “So it means we find yourself competing extra with the legacy safety corporations than the mannequin guys. So the query is, how do you beat them?”
For his half, Caccia doesn’t need Witness AI to be one of many startups to simply get acquired. He needs his firm to be the one which grows and turns into a number one unbiased supplier.
“CrowdStrike did it in endpoint [protection]. Splunk did it in SIEM. Okta did it in identification,” he mentioned. “Somebody comes via and stands subsequent to the large guys…and we constructed Witness to try this from Day One.





