Research & Insight

Blog

Original research, curated industry reads, and annotated perspectives on enterprise AI adoption, governance, and risk, from the 01 research desk.

01 Whitepapers
Whitepaper·01AI LTD · April 2026New

The Agentic Blast Radius

Why Your AI Agents Are Your Next Insider Threat

Non-Human IdentityExcessive AgencySanctioned LeakageGovernance GapShadow AI

Five control gaps that turn autonomous AI from a productivity multiplier into a breach multiplier. Each gap maps to a control enterprises already possess for human users. The defensive work is not inventing new disciplines; it is extending existing ones to a class of actor that the original designs did not anticipate.

Whitepaper·01AI LTD · March 2026

The Five Silent Pitfalls of First-Time AI Adoption

How Enterprises Undermine Their Own Transformation Before It Begins

Shadow AISanctioned LeakageVendor Lock-InGovernance GapSkills Illusion

Between 70-85% of AI initiatives fail to meet their expected outcomes. This paper identifies five structural decisions, made in the earliest stages of adoption, that consistently undermine first-time enterprise AI. They are not technical failures. They are governance, architecture, and competence failures, each avoidable, each compounding over time.

April 2026 Picks
Pick 1·Wharton AI & Analytics Initiative·April 2026

Two Early 2026 AI Exposures: Lessons for the Future of AI and Data Governance

Excessive AgencySanctioned LeakageNon-Human IdentityGovernance Gap
  • CodeWall's autonomous offensive AI agent compromised McKinsey's internal GenAI platform Lilli in under two hours on 28 February 2026
  • 46.5 million plaintext chat messages, 728,000 confidential files, and 57,000+ records exposed, covering M&A, strategy, and client engagement content
  • Attack path: 22 unauthenticated API endpoints discovered by the offensive agent, followed by SQL injection into the production database
  • No human operator in the loop on the attacker side. No credentials required. The entire intrusion ran at machine tempo.

01 Annotation

Wharton's April 2026 analysis of the McKinsey Lilli incident is the clearest published case study to date of the failure pattern we describe in our whitepaper The Agentic Blast Radius. Every one of the five control gaps is visible in the incident timeline. The attacker did not need a human insider. An autonomous offensive agent ran reconnaissance, discovered 22 unauthenticated API endpoints, identified the SQL injection path, and exfiltrated the database in under two hours. That is the Gap 5 scenario made concrete: a compromise that propagates faster than any conventional SOC escalation path can respond to it. The 46.5 million chat messages were not primary data. They were the memory layer of a GenAI platform: the accumulated conversational residue of tens of thousands of consultants working on strategy, M&A, and client engagements. Most organisations classify their CRM, their document management system, and their email archive as sensitive. Very few classify their agent memory, their vector stores, or their conversation logs with the same rigour. The unauthenticated API endpoints are the Gap 2 scenario: permission architecture scoped for the broadest possible task, not the narrowest current one. And the absence of a behavioural kill switch that could have terminated anomalous query patterns against the production database is Gap 5 in its operational form. The lesson is not that McKinsey got it uniquely wrong. The lesson is that the failure pattern (over-privileged agents, ungoverned memory, no machine-speed response capability) is the default pattern in most enterprise agentic deployments right now.

Maps to: Gaps 2, 4 & 5: Excessive Agency, Sanctioned Leakage (memory layer), Non-Human Identity

March 2026 Picks
Pick 1·EY US Newsroom·5 March 2026

Autonomous AI Adoption Surges at Tech Companies as Oversight Falls Behind

Shadow AIGovernance GapSanctioned Leakage
  • 52% of department-level AI initiatives operating without formal approval
  • 78% of leaders say adoption is outpacing their ability to manage risk
  • 45% report confirmed or suspected data leaks tied to unauthorised AI tools in the past year

01 Annotation

EY's March 2026 Technology Pulse Poll quantifies what we describe in our whitepaper as the Governance Gap: over half of department-level AI initiatives are running without formal oversight, and 78% of leaders admit adoption has outpaced risk management. Most telling: 45% report confirmed or suspected data leaks from unauthorised AI tools in the past year alone. The "velocity paradox" EY identifies (the tension between speed-to-market and accountability) is precisely why we argue governance must precede deployment, not follow it.

Maps to: Traps 1, 2 & 4: Shadow AI, Sanctioned Leakage, Governance Gap

Pick 2·AI Business · Gartner Data & Analytics Summit·16 March 2026

AI Innovation vs. Adoption: Why They Are Misaligned

Skills IllusionGovernance Gap
  • AI innovation continues to move at a rapid pace, but enterprise adoption remains slow
  • Two root causes: lack of data foundation and lack of governance model
  • Without governance, trust never forms, and without trust, adoption stalls

01 Annotation

Speaking at the Gartner Data & Analytics Summit this month, Deloitte's AI Institute head Beena Ammanath put it plainly: enterprises are caught between pressure from leadership to show AI value and foundations that aren't ready. Her two blockers (data readiness and governance) mirror what we identify as the Governance Gap and the Skills Illusion. As she notes: without the right governance model, trust never forms, and adoption stalls. That is the trap. Not the technology; the organisational discipline around it.

Maps to: Traps 4 & 5: Governance Gap, Skills Illusion

Further Reading
McKinsey & Company·March 2025

The State of AI: How Organizations Are Rewiring to Capture Value

Skills IllusionGovernance Gap

The flagship industry report our whitepaper draws from directly. McKinsey confirms that 78% of enterprises use AI, but only ~6% qualify as high performers: organisations that have actually redesigned workflows and achieved enterprise-wide impact. That single stat is the central argument of our paper: adoption metrics are not capability metrics. The 94% gap is exactly the terrain 01 operates in.

Maps to: Trap 5 (Skills Illusion) & Conclusion

IBM Security & Ponemon Institute·July 2025

Cost of a Data Breach Report 2025

Shadow AISanctioned Leakage

The definitive quantitative source on the financial cost of ungoverned AI. Shadow AI is now involved in 20% of all breaches, adding $670K per incident. 97% of AI-breached organisations lacked proper access controls; 63% had no AI governance policy at all. These aren't hypothetical risks; IBM put a dollar figure on them. The cost of not governing AI is measurable, and it compounds.

Maps to: Traps 1 & 2: Shadow AI, Sanctioned Leakage

CTO Magazine·June 2025

The Great AI Vendor Lock-In: How CTOs Can Avoid Getting Trapped by Big Tech

Vendor Lock-In

This article centres on the Builder.ai bankruptcy as a cautionary tale and provides a practical CTO-level playbook for vendor independence: code escrow, data portability, multi-provider architecture. It complements our whitepaper because it goes deeper on the operational 'how to avoid it' side of vendor lock-in, while our paper frames the strategic 'why it matters.' Builder.ai wasn't a fringe startup: backed by Microsoft, the Qatar Investment Authority, $445M raised, and still collapsed overnight, stranding its clients.

Maps to: Trap 3: Vendor Lock-In

Recognise any of these pitfalls?

We help enterprises audit their current AI posture, close governance gaps, and build proprietary AI systems they actually own.

Talk to us