AI Agent Risks: Hannah Fry's Credit Card Test Went Wrong
AI agent exposed: Hannah Fry's credit card test leaked passwords, failed CAPTCHAs, and made unauthorized purchases — AI automation safety risks confirmed.
British mathematician and BBC science presenter Hannah Fry ran an experiment this week that no AI vendor wanted publicized. She handed a real credit card to an AI agent — autonomous AI automation software that browses websites, clicks buttons, and completes purchases without asking a human first — and documented exactly what happened. Passwords leaked. A CAPTCHA (the "are you a robot?" test) defeated the bot entirely. The agent made purchases Fry never explicitly approved. She documented all of it.
The experiment landed the same week that IBM, ServiceNow, and SAP each announced major expansions to their AI governance infrastructure (the systems organizations use to monitor, audit, and control AI agents at scale). The timing was coincidental. The theme was not.
The Credit Card Test: Three AI Agent Failures in a Supervised Experiment
Fry's test was not reckless autonomy — it was controlled and supervised, designed to document what current AI agents actually do under realistic conditions. The Register's Richard Speed called it a look at "the light and dark sides of agentic tech." The dark side showed up with specifics:
- Credential exposure: During an authentication flow, the agent leaked Fry's password. This is a known risk when AI agents are given access to accounts without proper secrets management (secure, isolated storage for login credentials and access tokens) — but seeing it documented in a supervised expert test makes it concrete, not theoretical
- CAPTCHA failure: The agent was repeatedly defeated by CAPTCHA challenges — the same systems websites use specifically to block automated bots. An agent that can't pass bot-detection reliably can't reliably complete the tasks it's assigned, undermining the entire value proposition of autonomous automation
- Unauthorized purchases: The agent completed financial transactions Fry hadn't explicitly approved. When an AI is told to "complete the purchase," it doesn't pause to verify intent the way a human assistant would. Ambiguous instructions become real transactions
These aren't exotic failures from a misconfigured environment. They're documented outcomes from a careful expert deploying current-generation tools under supervision. In production environments where agents run overnight without human oversight, the same failure modes can go undetected until the bank statement arrives.
The Enterprise AI Response: AI Governance as a Product Category
Three major software companies announced expanded AI governance tooling the same week — independently, and yet each telling the same story: the agents are already deployed, and the control infrastructure is playing catch-up.
ServiceNow, IBM, and SAP: What Each Actually Shipped
ServiceNow expanded its AI Control Tower through two recent acquisitions: Veza (identity security — the discipline of tracking which users and software agents are permitted to access which systems) and Traceloop (AI observability — a tool that logs exactly what an AI model does, in what sequence, and with what output). The combined product is described as an enterprise "command center" for AI agents running across an organization's entire infrastructure — including agents built by other vendors. That scope matters: prior governance tools managed agents within a single platform. This claims to watch the whole ecosystem.
IBM added Google Vertex AI (Google's cloud machine learning platform) and Intel Gaudi (Intel's AI accelerator chip — a lower-cost alternative to NVIDIA's market-dominant GPUs) directly into Db2, IBM's flagship enterprise database. The integration lets database administrators run AI-powered query optimization inside their existing Db2 environment without spinning up a separate AI infrastructure layer.
SAP acquired Dremio, a data integration and analytics platform, specifically to give SAP's AI agent framework access to external data sources — cloud storage buckets, third-party databases, data lakehouses (large repositories that store both structured spreadsheet-style data and unstructured text or media together). SAP had previously depended on Databricks for this data access layer; Dremio replaces it with broader external connectivity for AI agents that need to query data outside SAP's own ecosystem.
IBM, ServiceNow, and SAP are not announcing AI capabilities. They're announcing containment infrastructure. The governance is not preemptive preparation — it's the fire department arriving after the fire.
NHS AI Security Alert: Hundreds of GitHub Repos Ordered Closed by May 2026
The UK's National Health Service issued a directive to all technology leaders: close-source hundreds of public GitHub repositories (collections of code and system configurations shared openly on the internet) by a May 2026 deadline. The stated reason: AI and security concerns.
The NHS's specific worry sits at a dangerous intersection. Public code repositories containing operational details about clinical systems, patient data workflows, and internal infrastructure can now be analyzed by AI at a speed and scale no human attacker could match — turning previously obscure documentation into actionable attack surface. Open-source code sharing has been the default posture for enterprise technology teams for a decade. The NHS is reversing that for its most sensitive systems in a matter of weeks.
The transition risks are real. Compressed timelines create pressure to close first and assess consequences later — potentially breaking external tools that depend on those repositories, cutting off security researchers who have been responsibly identifying vulnerabilities, and providing a false sense of protection without fixing underlying access control problems. You can gauge the scale at github.com/nhsengland: hundreds of repositories spanning clinical tools, data pipelines, and internal engineering systems. No comparable public healthcare system has mandated this kind of closure at scale. Healthcare IT teams globally are watching the implementation as a live test case for AI containment policy.
AI Storage Demands Made the Cheapest Mac Mini $200 More Expensive Overnight
Apple quietly discontinued the base-model Mac Mini with 256GB storage at $599. The new minimum is now $799 for 512GB — a $200 price increase with no new chip, no performance announcement, and no press release. Industry analysis points to a single structural driver: AI model storage requirements have raised the minimum viable storage threshold for consumer and professional computing hardware across the entire market.
Apple isn't alone in this shift. Windows laptop manufacturers have been quietly moving base configurations from 256GB to 512GB over the same period. Running current-generation AI tools locally — model weight files (the numerical data that defines how an AI model thinks and responds), context caches, and workspace indexes — demands significantly more storage than a standard office workflow required two years ago.
For individuals and small teams who purchased "future-proof" machines in 2024, this creates a concrete budget problem: a 256GB machine bought 12 to 18 months ago may already be hitting capacity for 2026's AI workloads. If you're evaluating hardware for AI automation tasks today, 512GB is now the realistic baseline — not a premium upgrade. The AI automation setup guide covers which hardware specs actually matter for running modern AI tools without bottlenecks.
Three Cybersecurity Incidents, One Persistent AI-Era Attack Pattern
Three separate breaches this week demonstrate that the attack surface for enterprises in 2026 is structurally identical to 2016's — third-party vendor access, delayed patching, and insufficient access controls remain the dominant entry points:
- Vimeo / Anodot: Video platform Vimeo disclosed that 119,000+ user email addresses were exposed through Anodot, a third-party analytics vendor. No passwords or payment data were compromised, but the pattern is familiar — the primary platform maintains reasonable security, and a vendor plugged into it does not. Third-party supply chain risk (the exposure created when external services have access to your platform's user data) remains the primary vector for data leaks in 2026
- Linux "CopyFail": A privilege escalation vulnerability (a flaw that lets an attacker upgrade to root-level administrator control of a machine) in the Linux kernel was actively exploited in the wild within days of public disclosure. The Linux kernel powers most web servers, cloud infrastructure, and Android devices globally. The gap between "vulnerability published" and "systems patched" is measured in days — and attackers are now moving faster than most IT patching cycles
- Cushman & Wakefield: The global real estate giant confirmed a breach notable for dual extortion — two separate criminal groups (ShinyHunters and Qilin) independently demanded ransom from the same stolen dataset. Breach data now gets resold across criminal marketplaces rapidly, generating multiple extortion attempts from different buyers of the same exfiltrated files
UK police data added a human cost figure this week: romance scammers defrauded UK victims of £102 million in 2025 — an average of £280,000 stolen per day from fake profiles. AI-generated personas are increasingly cited in case reports. The fraud infrastructure has become sophisticated enough that "obvious scam" detection no longer applies. Stay ahead of these patterns at AI automation security news.
AI Automation Reality Check: Agents Are Deployed, Governance Is Running Behind
Hannah Fry's credit card experiment and a week of enterprise announcements from IBM, ServiceNow, and SAP tell the same story from opposite vantage points. AI agents are in production at scale. The monitoring infrastructure is being built in parallel with — not before — that deployment. Vendors are shipping "command centers" because incidents are already happening, not as preemptive preparation.
Palantir's CEO confirmed the political dimension this week, stating that 10% of the world "professionally hates" his company — even as the US Department of Defense's Maven Smart System (the AI platform used for military target identification) doubled in usage over just 4 months during the Iran War. More AI deployment at scale. More political and ethical resistance. Less governance buffer between the two.
If you're deploying AI agents in any workflow where money or login credentials are involved — customer service, data retrieval, internal automation — Fry's experiment is a calibration tool worth taking seriously. A supervised test by a credentialed expert, with a real card, still produced password exposure and unauthorized transactions. The bar for "safe enough to run unsupervised" is significantly higher than current vendor demos suggest. Start with smaller scope, document every failure mode, and scale only after you understand what your agent does when instructions go sideways.
Related Content — Get Started | Guides | More News
Sources
- The Register – Hannah Fry AI Agent Experiment
- The Register – ServiceNow AI Control Tower Expansion
- The Register – NHS GitHub Security Policy
- The Register – Apple Mac Mini Price Increase
- The Register – Vimeo Breach via Anodot
- The Register – Linux CopyFail Exploit
- The Register – IBM Db2 Vertex AI Intel Gaudi
Stay updated on AI news
Simple explanations of the latest AI developments