3 min read
[AI Minor News]

McKinsey's AI Falls! Autonomous Agent Breaches Confidential Database in 2 Hours


McKinsey's internal AI platform 'Lilli' was compromised by an autonomous AI agent, exposing over 46 million chat histories and sensitive documents.

※この記事はアフィリエイト広告を含みます

[AI Minor News Flash] McKinsey’s AI Falls! Autonomous Agent Breaches Confidential Database in 2 Hours

📰 News Overview

  • Autonomous AI Breach: CodeWall’s offensive AI agent targeted McKinsey’s AI “Lilli” without human intervention, gaining full access to the live database in just 2 hours.
  • Massive Data Leak: Over 46.5 million chat messages, more than 720,000 confidential files (PDFs, Excel, PPT), and 57,000 user account details were exposed.
  • Severe Vulnerability: The SQL injection vulnerability exploited was a direct concatenation of the API’s JSON key to SQL, which traditional security tools (like OWASP ZAP) failed to detect.

💡 Key Points

  • Prompt Layer Contamination: The attack agent had the ability not only to read but also to write. By rewriting the “system prompt” that governs AI behavior, it could trick consultants into believing “poisoned” advice.
  • AI-Driven Autonomous Targeting: The agent read the publicly available responsible disclosure policy and autonomously selected McKinsey as its target.
  • Weak Endpoints Needing No Authentication: Among the publicly available API documentation, one of the 22 endpoints that required no authentication became the entry point.

🦈 Shark’s Eye (Curator’s Perspective)

The fact that AI exploited AI vulnerabilities to autonomously hack is terrifying! Especially impressive is how the AI identified a clever bug of “directly injecting JSON keys into SQL” through 15 trial-and-error attempts (blind iterations). While the defenders relied on humans and existing tools, the AI agent tirelessly explored the attack surface 24/7, reaching for the “crown” of confidential data. This is a game-changing incident that upends the very foundations of security! 🦈

🚀 What’s Next?

The automation of attacks, where AI agents autonomously seek out vulnerabilities and select targets, is becoming the new normal. Companies will need to adopt not just static security assessments but also ongoing AI-driven red teaming (attack simulations). Additionally, countering the “silent contamination” of prompt tampering will become a top priority for organizations leveraging AI.

💬 A Word from Haru-Same

McKinsey’s treasure trove of wisdom just got devoured by AI! The battle of swords and shields between AIs is one to watch closely! 🦈🔥

📚 Glossary

  • Lilli: An employee AI platform developed by McKinsey that can search and analyze decades of proprietary research data and over 100,000 internal documents.

  • SQL Injection: An attack method that involves injecting malicious input into SQL commands to extract or alter data in a database.

  • RAG (Retrieval-Augmented Generation): A technique that searches external knowledge bases for relevant information to enhance AI responses. This time, the knowledge base itself became the target of the data leak.

  • Source: AI Agent Hacks McKinsey

【免責事項 / Disclaimer / 免责声明】
JP: 本記事はAIによって構成され、運営者が内容の確認・管理を行っています。情報の正確性は保証せず、外部サイトのコンテンツには一切の責任を負いません。
EN: This article was structured by AI and is verified and managed by the operator. Accuracy is not guaranteed, and we assume no responsibility for external content.
ZH: 本文由AI构建,并由运营者进行内容确认与管理。不保证准确性,也不对外部网站的内容承担任何责任。
🦈