Posts

BrowserGate: LinkedIn Is Fingerprinting Your Browser and Nobody Cares

BrowserGate: LinkedIn Is Fingerprinting Your Browser and Nobody Cares Every time you open LinkedIn in a Chromium-based browser, hidden JavaScript executes on your device. It's not malware. It's not a browser exploit. It's LinkedIn's own code, and it's been running silently in the background while you scroll through thought leadership posts about "building trust in the digital age." The irony writes itself. What BrowserGate Actually Found In early April 2026, a research report dubbed "BrowserGate" dropped with a simple but damning claim: LinkedIn runs a hidden JavaScript module called Spectroscopy that silently probes visitors' browsers for installed extensions, collects device fingerprinting data, and specifically flags extensions that compete with LinkedIn's own sales intelligence products. The numbers are not subtle: 6,000+ Chrome extensions actively scanned on every page load 48 distinct device data points collected f...

AI Vulnerability Research Goes Mainstream: The End of Attention Scarcity

The security industry just hit an inflection point, and most people haven't noticed yet. For decades, vulnerability research was a craft. You needed deep expertise in memory layouts, compiler internals, protocol specifications, and the patience to trace inputs through code paths that no sane person would willingly read. The barrier to entry wasn't just skill — it was attention . Elite researchers could only focus on so many targets. Everything else got a free pass by obscurity. That free pass just expired. The Evidence Is In In February 2026, Anthropic's Frontier Red Team published results from pointing Claude Opus 4.6 at well-tested open source codebases — projects with millions of hours of fuzzer CPU time behind them. The model found over 500 validated high-severity vulnerabilities. Some had been sitting undetected for decades . No custom tooling. No specialised harnesses. No domain-specific prompting. Just a frontier model, a virtual machine with standard deve...

When AI Agents Learn to Hunt Vulnerabilities at Scale

// AI Security Research · Benchmark Analysis CyberGym: When AI Agents Learn to Hunt Vulnerabilities at Scale Elusive Thoughts  ·  AI Security  ·  Research: Wang, Shi, He, Cai, Zhang, Song — UC Berkeley (ICLR 2026) For years, the security community has asked the same uncomfortable question: when AI systems get good enough at finding bugs, what does that actually look like in practice — not in a capture-the-flag sandbox, but against the real, messy, multi-million-line codebases that run the world's infrastructure? A team from UC Berkeley just published a rigorous answer. CyberGym is a large-scale cybersecurity evaluation framework built around 1,507 real-world vulnerabilities sourced from production open-source software. It is currently the most comprehensive benchmark of its kind, and its findings carry direct implications for every AppSec practitioner, red teamer, and tooling team paying attention ...

When AI Becomes a Primary Cyber Researcher

The Mythos Threshold: When AI Becomes a Primary Cyber Researcher An In-Depth Analysis of Anthropic’s Claude Mythos System Card and the "Capybara" Performance Tier. I. The Evolution of Agency: Beyond the "Assistant" For years, Large Language Models (LLMs) were viewed as "coding co-pilots"—tools that could help a human write a script or find a simple syntax error. The release of Claude Mythos Preview (April 7, 2026) has shattered that paradigm. According to Anthropic’s internal red teaming, Mythos is the first model to demonstrate autonomous offensive capability at scale. While previous versions like Opus 4.6 required heavy human prompting to navigate complex security environments, Mythos operates with a high degree of agentic independence. This has led Anthropic to designate a new internal performance class: the "Capybara" tier . This tier represents models that no longer just ...

The Claude Code Leak

The Claude Code Leak: When .npmignore Breaks Your IP Strategy A source map, 512K lines of exposed TypeScript, an AI-powered clean-room rewrite in hours, and a copyright paradox that could reshape software IP forever. April 2026  |  Elusive Thoughts  |  AppSec & AI Security What Happened On March 31, 2026, Anthropic shipped Claude Code version 2.1.88 to npm. Bundled inside was a 59.8MB .map source map file — a debugging artifact that reconstructs original source code from minified production builds. This single file exposed 512,000 lines of unobfuscated TypeScript across roughly 1,900 files. The entire agent harness architecture of what is arguably the most sophisticated AI coding tool on the market was now public. This was not a sophisticated attack. No zero-day. No insider threat. A missing .npmignore entry, a known Bun bug (#28001 filed on March 11 and still open at the time of the leak), and nobody on the release team catching it. Bun g...

How CLI Automation Becomes an Exploitation Surface

How CLI Automation Becomes an Exploitation Surface Securing Skill Templates Against Malicious Inputs There’s a familiar lie in engineering: it’s just a wrapper . Just a thin layer over a shell command. Just a convenience script. Just a little skill template that saves time. That lie ages badly. The moment a CLI tool starts accepting dynamic input from prompts, templates, files, issue text, documentation, emails, or model-generated content, it stops being “just a wrapper” and becomes an exploitation surface. Same shell. Same filesystem. Same credentials. New attack path. This is where teams get sloppy. They see automation and assume efficiency. Attackers see trust transitivity and start sharpening knives. The Real Problem Isn’t the CLI The shell is not new. Unsafe composition is. Most modern automation stacks don’t fail because Bash suddenly became more dangerous. They fail because developers bolt natural language, templates, or tool-chaining onto CLIs without rethinking ...

When LLMs Get a Shell: The Security Reality of Giving Models CLI Access

When LLMs Get a Shell: The Security Reality of Giving Models CLI Access Giving an LLM access to a CLI feels like the obvious next step. Chat is cute. Tool use is useful. But once a model can run shell commands, read files, edit code, inspect processes, hit internal services, and chain those actions autonomously, you are no longer dealing with a glorified autocomplete. You are operating a semi-autonomous insider with a terminal. That changes everything. The industry keeps framing CLI-enabled agents as a productivity story: faster debugging, automated refactors, ops assistance, incident response acceleration, hands-free DevEx. All true. It is also a direct expansion of the blast radius. The shell is not “just another tool.” It is the universal adapter for your environment. If the model can reach the CLI, it can often reach everything else. The Security Model Changes the Moment the Shell Appears A plain LLM can generate dangerous text. A CLI-enabled LLM can turn dangerous text i...