A recent incident at Meta shows how an AI agent provided guidance that led an engineer to unintentionally expose a large amount of sensitive internal data to employees for a short period of time.
While Meta confirmed the issue was contained and no external data was mishandled, the episode highlights a broader risk as AI agents become embedded in engineering workflows. These systems aren’t just generating suggestions, they’re influencing real actions inside environments that handle sensitive data.
Gidi Cohen, CEO & Co-founder, Bonfy.AI
“Meta’s incident is exactly what happens when you let agents loose on sensitive data without any real data-centric guardrails. This wasn’t some exotic AGI failure, it was a very simple pattern: an engineer asked an internal agent for help, the agent produced a “reasonable” plan, and that plan quietly exposed a huge amount of internal and user data to people who were never supposed to see it.
The problem is that neither the engineer nor the agent had any persistent notion of “who actually should see this data” beyond whatever happened to sit in a narrow context window at that moment. Traditional controls don’t help much here. Endpoint DLP, CASB, browser controls, even basic role-based permissions, none of them are watching the actual content as it moves through an agent’s reasoning steps and tool calls, especially when the agent is running as a system service in some framework.
Our view is simple: treat agents like very fast, very forgetful junior interns and make the data security layer smart enough to compensate. That means three things: constrain what data is even available to the agent via contextual labeling and grounding; give the agent a Bonfy MCP tool it can call inline to ask “is this safe to use or send in this context?” before it takes an action; and inspect what ultimately comes out of the workflow before it lands in email, chat, dashboards, or internal portals. In a Meta-style scenario, those controls would have either prevented the broad internal exposure entirely or at least shrunk the blast radius to something manageable.
As organizations “experiment at scale” with agents, the only sustainable path is to make agents first-class entities in the risk model and put the intelligence where it belongs: on the data that’s being read, composed, and shared, not just on the configuration screens of yet another AI tool.”
The thing is that when you expose AI anything to sensitive data, it can get out there. Samsung banned AI usage for that reason. Keep that in mind if you’re an organization that uses AI


Meta pauses work with Mercor after supply chain breach raises risk to AI training data
Posted in Commentary with tags Facebook on April 6, 2026 by itnerdAs first reported by Wired on Friday, Meta has paused all work with AI data startup Mercor following a confirmed security breach linked to a supply chain attack involving the LiteLLM open-source project, which impacted thousands of organizations globally.
Mercor, which provides proprietary training data to major AI companies including Meta, OpenAI, and Anthropic, said it was among those affected and has launched an investigation with third-party forensic experts.
The breach raised concerns about potential exposure of sensitive AI training data and internal datasets, which are used to develop and fine-tune large language models. Reports indicate that Mercor’s systems were impacted as part of a broader compromise involving malicious updates to widely used AI tooling, though it remains unclear what specific data was accessed.
Michael Bell, Founder & CEO, Suzu Labs had this comment:
“The Mercor breach is what happens when the companies building the most valuable AI models in the world outsource the creation of their training data to vendors running on Airtable and shared passwords. A single poisoned open-source package gave attackers VPN credentials, and from there they walked through Mercor’s systems and took 4TB of proprietary datasets, source code, and contractor PII.
“We’ve been investigating these AI data vendors for months and found the same structural failures at Sama, Teleperformance, Scale AI, and Cognizant we see unrotated credentials, info-stealer infections on contractor endpoints, and access controls that don’t exist. The training data behind every major frontier model is sitting inside vendors that wouldn’t pass a basic security audit, and now that data is on an extortion site. This is a national security problem dressed up as a vendor management failure.”
Lydia Zhang, President & Co-Founder,Ridge Security Technology Inc. adds this comment:
“This incident alerts us that AI training data should be treated as critical infrastructure, subject to stricter security scrutiny and regulation.
“The breach also underscores the risks of relying directly on open-source projects in enterprise environments. Supply chain attacks, like the compromised LiteLLM library in this case, can introduce vulnerabilities at scale and expose highly sensitive data.
“At a minimum, enterprises should adopt thoroughly tested and commercially supported versions of such components, with stronger security guarantees and accountability.”
Noelle Murata, Sr. Security Engineer, Xcape, Inc. provided this comment:
“Meta’s indefinite suspension of its partnership with Mercor underscores how the AI industry’s rush to outsource training data has effectively liquidated billions in proprietary methodology. By allowing a poisoned version of the LiteLLM gateway (versions 1.82.7 and 1.82.8) to persist in their environment, Mercor gifted attackers 4 TB of data, including the precise “secret sauce” protocols Meta and OpenAI use to tune their models.
“This was not a sophisticated zero-day; it was a basic supply chain failure where a compromised security scanner (Trivy) was used to poison a niche dependency that nobody bothered to pin. For anyone surprised that an autonomous, interconnected AI stack would eventually expose sensitive data to the internet, the lesson is clear.
“If you are not auditing your data vendors for basic dependency hygiene, your IP is already public property. Defenders must immediately scan for litellm_init.pth files, which provide stealthy persistence on every Python startup, and rotate all LLM provider API keys and cloud tokens. Protecting training integrity now requires treating every AI data broker as a high-risk production endpoint and enforcing strict, pinned Software Bill of Materials (SBOM) standards.
“If your AI supply chain is this leaky, you are not training a model; you are just broadcasting a technical manual to Lapsus$.”
Supply chain vulnerabilities are real. If your organization doesn’t take them seriously, your organization will get pwned. It’s as simple as that. And you can double that if AI is involved.
Leave a comment »