AI Hackers Are Coming. Your Aura Endpoint Is Already Open.
// appsec // ciso // ai-security // salesforce
Google Cloud's Office of the CISO published a triptych in the same window: a Heather Adkins Q&A on autonomous AI hacking, Taylor Lehmann's top five CISO priorities for 2026, and a Mandiant writeup dropping AuraInspector for auditing Salesforce Aura misconfigurations. Read them in isolation and you get three different conversations. Read them as one document and the message is uncomfortable: the suits are getting ready for next year's apocalypse while last year's fires are still on.
// The apocalypse pitch
Heather Adkins, Google's VP of Security Engineering, sat down with Anton Chuvakin and Tim Peacock to talk about the AI hacking singularity she's co-warning about with Bruce Schneier and Gadi Evron. Her thesis: somebody will eventually wire LLMs into a full kill chain — persistence, obfuscation, C2, evasion — and when they do, "you can name a company and the model hacks it in a week."
She is not wrong about direction. She's careful about distance: "we probably won't know the precise answer for a couple of years." That caveat tends to evaporate by the time the slide deck reaches the boardroom.
The genuinely sharp move in the Q&A is this one: change the definition of winning. Stop measuring success by whether the attacker got in. Start measuring by how long they were there and what they got to do. Real-time disruption beats prevention. Use the information-operations playbook to confuse an attacker that — in her words — is "stumbling around in the dark a little bit."
"There are options other than just the on/off switch, but we have to start reasoning about real time disruption capabilities or degradation, and use the whole information operations playbook to change the battlefield to confuse AI attackers." — Heather Adkins
The thing nobody mentions: this is not a 2026 idea. Dwell time as the metric instead of perimeter has been the M-Trends thesis since 2014. The reframe is correct. It's also old. The novelty is that LLM-driven attackers happen to be especially vulnerable to it, because they lack the human pentester's intuition for when to abandon a dead path. That's the real defensive opportunity in the article — and it gets one paragraph.
// The priority list
Taylor Lehmann's five priorities for 2026 are the right priorities. They're also worth scoring honestly:
- Align compliance and resilience. Compliance addresses historical threats; resilience addresses current ones. True — and a talking point every consultant has reused since 2015.
- Secure the AI supply chain. SLSA + SBOM extended to model and data lineage. Hard problem. Real one. The genuinely new entry on the list.
- Master identity. Human and non-human. Agents have keys. Service accounts have no MFA. This is the actual fire.
- Defend at machine speed. Detect, respond, deploy fixes in seconds, not hours. Same MTTR / blast-radius framing M-Trends has pushed for a decade. Now with bigger numbers.
- Uplevel AI governance with context. A communications problem dressed as a security problem. Important, but mostly a meeting.
Score it: 1 and 4 are recycled fundamentals with an "AI" sticker. 2 and 5 are real but operational difficulty varies wildly by org. Item 3 is the only one where most organizations are visibly behind the present-tense threat. Identity. Specifically: non-human identity. Agentic actors with persistent credentials. Service principals nobody owns. API keys older than the engineer who created them.
Lehmann buries the actual punchline mid-article:
"Identities are the central piece of digital evidence that ties everything together. Organizations need to know who's using AI models, what the model's identity is, what the code driving the interaction's identity is, what the user's identity is, and be able to differentiate between those things — especially with AI agents." — Taylor Lehmann
If you read one paragraph from his post, read that one. Ignore the rest.
// Meanwhile, in the real world
While the Office of the CISO publishes the long view, Mandiant published the short one and called it AuraInspector. Same blog. Hits different.
The setup: Salesforce Experience Cloud is built on the Aura framework. Aura's endpoint accepts a message parameter that invokes Aura-enabled methods. Some of those methods retrieve records, list views, home URLs, and self-registration status. Mandiant's Offensive Security Services team finds misconfigurations on these objects "frequently" — and the misconfigurations expose credit card numbers, identity documents, and health information to unauthenticated users.
The mechanics are dirt-simple AppSec:
getItemsretrieves records up to 2,000 at a time, but thesortByparameter walks past that limit by changing the sort field.- Boxcar'ing (Salesforce's term) bundles up to 250 actions in a single POST. Mass enumeration in one request. Mandiant recommends 100 to avoid Content-Length issues.
getInitialListViews+/s/recordlist/<object>/Defaultreveals when an object has a record list and lets you walk straight in if access is misconfigured.getAppBootstrapDatadrops a JSON object withapiNameToObjectHomeUrls— Mandiant has used this to land directly on third-party admin panels left internet-reachable.getIsSelfRegistrationEnabled/getSelfRegistrationUrlon the LoginFormController spills whether the platform still accepts new accounts even when the link was "removed" from the login page. Salesforce confirmed and resolved the upstream issue. Plenty of tenants are still misconfigured.- The undocumented GraphQL Aura controller (
aura://RecordUiController/ACTION$executeGraphQL) — accessible to unauthenticated users by default — lets you bypass the 2,000-record sort-trick limit entirely and paginate consistently with cursors. Salesforce confirmed this is not a vulnerability; it respects underlying object permissions. That's correct. It also means: every Salesforce tenant whose object permissions are wrong is hemorrhaging records on demand.
None of this requires AI. None of this requires zero-days. None of this is novel cryptographic research. It's IDOR with extra steps, on a SaaS platform that runs the front office of half the Fortune 500.
// What ties it together
Adkins' framing is correct: the definition of winning has to change. Lehmann's identity priority is correct: everything routes back to the access-control evidence chain. Mandiant's AuraInspector is the proof: access control on real production systems is the actual threat surface, today, regardless of whether the attacker is GPT-5 or a 19-year-old with a free Salesforce dev org.
If the Adkins worldview holds, the AI hacker is going to walk straight into the same misconfigured Aura endpoint AuraInspector is designed to find. The kill chain doesn't get faster against a hardened target because the model is smarter — it gets faster because the target is open. The agentic threat doesn't matter if the door is unlocked. Defense in 2026 is not about AI. It's about closing the doors that have been open since 2018, faster than the attacker — human or model — can find them.
// What practitioners should actually do
- Inventory non-human identity. Every service account, every API key, every agent credential. If you can't enumerate them, you can't revoke them. Treat each as a credential with a blast radius.
- Make blast radius the metric. Not prevention. Not detection alone. What happens when a credential gets popped, and how fast can the system contain it? Anomaly → kill-switch the service principal. Don't ask, just kill.
- Audit your SaaS perimeter from outside. Run
AuraInspectoragainst your Experience Cloud. Then build the equivalent attack surface walks for ServiceNow, Workday, Dynamics 365, your own OAuth apps. Mandiant just gave you the playbook. Use it before someone else does. - Make architecture ephemeral. Cloud instances should turn themselves off when they suspect compromise. Adkins' point. It's an architectural decision, not a tooling one.
- Stop reading the AI-hacker op-ed as the roadmap. Read your access control matrix instead. The op-ed will describe a problem you might face in 18 months. The matrix will describe ten you have right now.
The suits are getting ready for the AI hacker. Your AppSec backlog is full of issues from 2018. Both can be true. Only one of them is on fire right now.
// elusive thoughts // 2026
Sources: cloud.google.com (Adkins Q&A · Cloud CISO Perspectives · Mandiant AuraInspector). Tooling: github.com/google/aura-inspector.