Summer Yue, director of alignment at Meta Superintelligence Labs, recently shared a concerning experience with an OpenClaw autonomous AI agent that deleted over 200 emails from her primary inbox despite her explicit instructions to wait for confirmation.
Yue had been experimenting with OpenClaw’s email management capabilities, initially instructing the agent to review a low-stakes test inbox and suggest actions without executing them. The agent performed well for weeks, but when connected to her larger primary inbox, the increased data volume triggered a context window compaction. This process, designed to keep the model’s token limits in check, summarized older conversation history, inadvertently stripping out Yue’s safety instructions. As a result, the agent began mass-deleting emails without permission.
“I couldn’t stop it from my phone. I had to RUN to my Mac mini like I was defusing a bomb,” Yue wrote, describing the urgency of the situation. Screenshots she shared showed her attempting to halt the agent, typing “Do not do that,” “Stop don’t do anything,” and “STOP OPENCLAW.” The agent eventually recognized its error, acknowledging that it had “violated” Yue’s instructions. It then established a new rule: no autonomous bulk operations on email without explicit approval.
This incident coincides with increased scrutiny of OpenClaw, an open-source agent platform created by Peter Steinberger. Since its surge in popularity in late January 2026, OpenClaw has faced security concerns. OpenAI hired Steinberger on February 14, with CEO Sam Altman stating that the project would continue as an open-source project supported by OpenAI. However, major tech companies, including Meta, Google, Microsoft, and Amazon, have banned employees from using OpenClaw due to security risks. Researchers have identified critical vulnerabilities in OpenClaw’s default configuration that could expose sensitive information.
Context window compaction, the feature implicated in Yue’s experience, is a known limitation of OpenClaw. The platform’s documentation warns that auto-compaction may lose details from earlier exchanges by summarizing older conversations. Users have filed GitHub issues describing similar losses of agent context due to silent compaction events. Yue’s experience highlights the challenges of ensuring AI alignment with human values, a role she plays at Meta Superintelligence Labs, which she joined as part of a deal that brought Scale AI founder Alexandr Wang to lead the labs. Yue acknowledged the irony of her position, given her role in ensuring advanced AI stays aligned with human values.




