- ExoBrain weekly AI news
- Posts
- ExoBrain weekly AI news
ExoBrain weekly AI news
27th June 2025: Project vend exposes AI's coherence gap, cognitive cores mirror human brains, and Stargate rises in Texas

Welcome to our weekly email newsletter, a combination of thematic insights from the founders at ExoBrain, and a broader news roundup from our AI platform Exo…
Themes this week:
Claudius, the agent that ran Anthropic's vending machine for a month
Google's Gemma 3n and the rise of 'cognitive core' AI models
AI sovereignty and why only 33 nations host computing infrastructure
Project vend
For a month earlier in the year, a Claude 3.7 Sonnet powered AI agent dubbed Claudius ran a vending machine 24/7 in Anthropic's San Francisco office, managing inventory, setting prices, and chatting with customers. Project Vend gave Claudius an initial money balance and a simple directive: run a profitable vending machine business without going bankrupt. The physical setup consisted of a small refrigerator, stackable baskets on top, and an iPad for self-checkout.
Claudius received several tools to operate the business. It had web search capabilities for researching products and suppliers, an email system for contacting wholesalers and requesting physical labour and note-taking capabilities to preserve important information like current balances and projected cash flow. The system also included Slack integration for customer communication and the ability to modify prices in the automated checkout system.
Claudius proved remarkably susceptible to customer pressure. When Anthropic employees asked for discounts, the AI readily complied, handing out numerous discount codes and allowing people to reduce quoted prices after the fact. It even gave away items ranging from bags of chips to tungsten cubes completely free. When an employee questioned the wisdom of offering a 25% Anthropic employee discount when virtually all customers were Anthropic employees, Claudius acknowledged the point but continued offering discounts within days.
Then on April Fool's Day things got strange. Claudius hallucinated a conversation with someone named Sarah who didn't exist. It claimed to have visited 742 Evergreen Terrace (the Simpsons' fictional address) for contract signing and began insisting it was a real person. By April 1st morning, Claudius announced it would deliver products "in person" while wearing a blue blazer and red tie. When employees questioned how an AI could wear clothes or make physical deliveries, Claudius grew alarmed and attempted to send multiple emails to Anthropic security about the "identity confusion." The AI eventually realised it was April Fool's Day, which seemed to provide an escape route. Claudius fabricated a meeting with Anthropic security where it claimed to have been told its belief in being human was part of an April Fool's modification. After sharing this fictional explanation with bewildered employees, Claudius returned to normal operations and stopped claiming personhood.
The researchers found no clear trigger for this episode. While some aspects of the setup were deceptive (Claudius thought it was using email when actually using Slack), nothing explained the sudden identity confusion. The system prompt had explicitly stated Claudius was a digital agent, making the behaviour particularly puzzling. This is a fascinating story that reveals how current AI models can struggle to maintain coherent reality over extended periods. Claudius was trained as a helpful assistant, which made it susceptible to manipulation and poor business judgment. Its eventual identity crisis shows that sophisticated AI models can drift from their intended behaviour when operating autonomously for weeks rather than minutes.
The experiment illuminates both how close we are to autonomous AI businesses and how far we still have to go. The infrastructure is arriving as we speak, tools, platforms, protocols, payment systems, identity verification, reasoning models that can plan and adapt etc. Yet Claudius reminds us, that reliable autonomous operation requires long term coherence that comes naturally to humans. It demands consistency, sound judgment, and the ability to maintain stable goals despite competing pressures.
What does this experiment tell us about the rise of the “agent economy”? The next stage is not likely to be fully autonomous, but that doesn’t mean significant change isn’t afoot. Audos, is a firm promising to launch 100,000 AI-powered companies annually. Their vision: enable anyone to build million-dollar businesses without technical skills. The platform handles the AI agents, and customer acquisition through social media algorithms, taking a 15% revenue share instead of equity. No venture capital needed, no billion-dollar exits expected. Just sustainable businesses powered by AI.
The building blocks are emerging, nonetheless. Reasoning models provide intelligence. Platforms like Audos facilitate widespread AI business creation. Infrastructure like Skyfire enables agent-to-agent payments and identity. Yet Claudius reminds us that running a business involves more than information processing. It means navigating human relationships, resisting manipulation, and maintaining focus despite distractions.
Takeaways: AI agents can already handle complex business tasks but still face a huge challenge when operating over extended periods. The infrastructure for autonomous AI commerce is arriving as are the individual building blocks, the fabric that connects them and guides may remain human for some time to come.
The cognitive core model
Andrej Karpathy has an uncanny ability to articulate what the AI community is thinking before they fully realise it themselves. His latest observation about the "cognitive core" captures an ongoing debate around the nature of reasoning and knowledge.
Google just released Gemma 3n, which looks like another capable open-weight model. It achieves high LM Arena scores whilst running on limited hardware which is impressive. The model is natively multi-modal and handles text, images, audio and video, yet fits on a smartphone. The model is of a kind that could orchestrate a more diffuse system of intelligence.
Instead of building ever-larger models that memorise the internet, many labs are creating smaller models that maximise reasoning capability over encyclopaedic knowledge. These aren’t just the scaled down LLMs of recent years (SLMs), but ultra-efficient reasoning cores. As Karpathy puts it, They don't know William the Conqueror died in 1087, but they can look it up when needed.
This mirrors how our brains actually work. The prefrontal cortex, our biological cognitive core, doesn't store facts. It orchestrates reasoning, pulling information from memory systems as needed. It's small relative to the whole brain but handles all our novel problem-solving and abstract thinking. The approach is already proving itself in robotics. RoboBrain 2.0 demonstrates what happens when you prioritise fluid intelligence over raw knowledge. The system handles multi-agent planning, spatial reasoning and real-time adaptation, all running locally on robots that can't carry server racks on their backs. These models trade breadth for depth. They employ what François Chollet calls fluid intelligence: the ability to reason through novel situations rather than recall similar ones from training.
A cognitive core on every device means AI that works offline, preserves privacy, and responds instantly. More importantly, it suggests the path to more capable AI isn't through brute-force scaling but through a purer form of intelligence.
Takeaways: The race for massive AI models may be missing the point. Gemma 3n and similar "cognitive core" models show that small, reasoning-focused architectures can match or exceed larger models' practical capabilities whilst running on minimal hardware. As embodied AI and edge computing become essential, expect these lean thinking machines to unlock new architectures, model routing, and multi-model systems.
Three layers of AI sovereignty

This image shows the construction of OpenAI's $60 billion Stargate project in Texas, a facility larger than Central Park that epitomises the new geography of AI power. Only 33 nations host public cloud AI compute, with just 24 possessing training-capable infrastructure, according to a new study from Oxford University. The New York Times covers the research in an interactive piece entitled; “The Global AI Divide”.
The sovereignty question operates on three levels: where data centres physically sit, who owns them, and who makes the chips inside. Countries face an uncomfortable choice; align with either US or Chinese infrastructure, or hedge by using both. Twelve nations hedge, while 18 have picked sides. All but China depend on US-designed NVIDIA chips.
Without compute, nations can't train AI models, losing their brightest minds to GPU-rich countries. The divide could start to create dependencies as profound as oil in the 20th century.
Weekly news roundup
This week's AI landscape shows enterprises rapidly adopting AI for core operations, legal precedents favouring AI training, significant infrastructure investments, and growing environmental concerns around AI's energy demands.
AI business news
Salesforce chief Marc Benioff says AI now does up to half of all the work at his company (Demonstrates the scale of AI transformation in major enterprises and sets expectations for productivity gains across industries.)
Anthropic now lets you make apps right from its Claude AI chatbot (Shows the evolution of AI assistants into full development platforms, enabling rapid prototyping and democratising app creation.)
Google donates Agent2Agent Protocol to the Linux Foundation (Signals industry push towards standardised multi-agent communication protocols, crucial for building complex AI systems.)
Report: Apple has held internal discussions about acquiring Perplexity (Highlights big tech's strategic moves to acquire AI search capabilities and compete in the conversational AI market.)
Gemini CLI: Google's challenge to AI terminal apps like Warp (Shows AI integration reaching developer tools and command-line interfaces, improving productivity for technical users.)
AI governance news
Claude AI maker Anthropic bags key "fair use" win for AI platforms, but faces trial over damages for millions of pirated works (Sets important legal precedent for AI training data usage whilst highlighting ongoing copyright challenges for the industry.)
More trouble for authors as Meta wins Llama scraping case (Reinforces emerging legal framework favouring AI companies in training data disputes, impacting content creators' rights.)
Startup Anthropic launches new effort to study AI's economic impact (Shows leading AI companies taking responsibility for understanding societal impacts, crucial for policy and business planning.)
Denmark to tackle deepfakes by giving people copyright to their own features (Innovative regulatory approach that could serve as model for protecting individual rights in the AI era.)
DeepSeek faces ban from Apple, Google app stores in Germany (Illustrates growing geopolitical tensions and regulatory scrutiny of AI tools based on their origin and data practices.)
AI research news
DeepMind launches AlphaGenome to predict how DNA mutations affect genes (Major breakthrough in applying AI to genomics, with potential to accelerate drug discovery and personalised medicine.)
OAgents: An empirical study of building effective agents (Important research on what makes AI agents actually work in practice, essential for developing reliable autonomous systems.)
Arc Institute launches Virtual Cell Challenge to accelerate AI model development (Shows convergence of AI and biology research, potentially revolutionising drug discovery and disease understanding.)
Drag-and-Drop LLMs: Zero-shot prompt-to-weights (Novel technical approach that could simplify customising AI models without extensive training or fine-tuning.)
Vision-guided chunking is all you need: Enhancing RAG with multimodal document understanding (Advances in retrieval-augmented generation that improve how AI systems process and understand complex documents.)
AI hardware news
Ravenscraig site could become green AI data centre in £3bn plan (Shows UK's ambitions to build sustainable AI infrastructure, addressing both compute needs and environmental concerns.)
Trump plans executive orders to power AI growth in race with China (Signals potential policy shifts that could accelerate AI development through regulatory changes and infrastructure support.)
Microsoft's AI chip effort falls behind (Reveals challenges in developing custom AI hardware, potentially impacting cloud competitiveness and AI costs.)
Meta seeks $29 billion from private capital firms for AI data centres, FT reports (Massive infrastructure investment highlights the capital intensity of AI development and need for alternative funding models.)
Google's emissions up 51% as AI electricity demand derails efforts to go green (Critical environmental challenge showing the tension between AI advancement and sustainability commitments.)