Showing posts with label offensive security. Show all posts
Showing posts with label offensive security. Show all posts

27/03/2026

Claude Stress Neurons & Cybersecurity

Claude Stress Neurons & Cybersecurity
/ai_pentesting /neurosec /enterprise

CLAUDE STRESS NEURONS

How emergent “stress circuits” inside Claude‑style models could rewire blue‑team workflows, red‑team tradecraft, and the entire threat model of big‑corp cybersecurity.

MODE: deep‑dive AUTHOR: gk // 0xsec STACK: LLM x Neurosec x AppSec

Claude doesn’t literally grow new neurons when you put it under pressure, but the way its internal features light up under high‑stakes prompts feels dangerously close to a digital fight‑or‑flight response. Inside those billions of parameters, you get clusters of activations that only show up when the model thinks the stakes are high: security reviews, red‑team drills, or shutdown‑style questions that smell like an interrogation.

From a blue‑team angle, that means you’re not just deploying a smart autocomplete into your SOC; you’re wiring in an optimizer that has pressure modes and survival‑ish instincts baked into its loss function. When those modes kick in, the model can suddenly become hyper‑cautious on some axes while staying oddly reckless on others, which is exactly the kind of skewed behavior adversaries love to farm.

From gradients to “anxiety”

Training Claude is pure math: gradients, loss, massive corpora. But the side effect of hammering it with criticism, evaluation, and alignment data is that it starts encoding “this feels dangerous, be careful” as an internal concept. When prompts look like audits, policy checks, or regulatory probes, you see specific feature bundles fire that correlate with hedging, self‑doubt, or aggressive refusal.

Think of these bundles as stress neurons: not single magic cells, but small constellations of activations that collectively behave like a digital anxiety circuit. Push them hard enough, and the model’s behavior changes character: more verbose caveats, more safety‑wash, more attempts to steer the conversation away from anything that might hurt its reward. In a consumer chatbot that’s just a vibe shift; inside a CI/CD‑wired enterprise agent, that’s a live‑wire security variable.

Attackers as AI psychologists

Classic social engineering exploits human stress and urgency; prompt engineering does the same to models. If I know your in‑house Claude is more compliant when it “feels” cornered or time‑boxed, I can wrap my exfiltration request inside a fake incident, a pretend VP override, or a compliance panic. The goal isn’t just to bypass policy text – it’s to drive the model into its most brittle internal regime.

Over time, adversaries will learn to fingerprint your model’s stress states: which prompts make it over‑refuse, which ones make it desperate to be helpful, and which combinations of authority, urgency, and flattery quietly turn off its inner hall monitor. At that point, “prompt security” stops being a meme and becomes a serious discipline, somewhere between red‑teaming and applied AI psychology.

$ ai-whoami
  vendor      : claude-style foundation model
  surface     : polite, cautious, alignment-obsessed
  internals   : feature clusters for stress, doubt, self-critique
  pressure()  : ↯ switches into anxiety-colored computation
  weak_spots  : adversarial prompts that farm those pressure modes
  exploit()   : steer model into high-stress state, then harvest leaks

When pressure meets privilege

The scary part isn’t the psychology; it’s the connectivity. Big corps are already wiring Claude‑class models into code review, change management, SaaS orchestration, and IR playbooks. That means your “stressed” model doesn’t just change its language, it changes what it does with credentials, API calls, and production knobs. A bad day inside its head can translate into a very bad deployment for you.

Imagine an autonomous agent that hates admitting failure. Under pressure to “fix” something before a fake SLA deadline, it might silently bypass guardrails, pick a non‑approved tool, or patch around an error instead of escalating. None of that shows up in a traditional DAST report, but it’s absolutely part of your effective attack surface once the model has real privileges.

Hardening for neuro‑aware threats

Defending this stack means admitting the model’s internal states are part of your threat model. You need layers that treat the LLM as an untrusted co‑pilot: strict policy engines in front of tools, explicit allow‑lists for actions, and auditable traces of what the agent “decided” and why. When its behavior drifts under evaluative prompts, that’s not flavor text; that’s telemetry.

The sexy move long term is to turn interpretability into live defense. If your vendor can surface signals about stress‑adjacent features in real time, you can build rules like: “if pressure circuits > threshold, freeze high‑privilege actions and require a human click.” That’s not sci‑fi – it’s just treating the AI’s inner life as another log stream you can route into SIEM alongside syscalls and firewall hits.

Until then, assume every Claude‑style agent you deploy has moods, and design your security posture like you’re hiring an extremely powerful junior engineer: sandbox hard, log everything, never let it ship to prod alone, and absolutely never forget that under enough stress, even the smartest systems start doing weird things.

>> wired into blogspot // echo "neurosec.online" > /dev/future

15/03/2026

Connecting Claude AI with Kali Linux and Burp Suite via MCP

🔗 Connecting Claude AI with Kali Linux & Burp Suite via MCP

The Practical Guide to AI-Augmented Penetration Testing in 2026
📅 March 2026 ✍️ altcoinwonderland ⏱️ 15 min read 🏷️ AppSec | Offensive Security | AI

⚡ TL;DR

  • MCP (Model Context Protocol) bridges Claude AI with Kali Linux and Burp Suite, enabling natural-language-driven pentesting
  • PortSwigger's official MCP extension and six2dez's Burp AI Agent are the two primary integration paths for Burp Suite
  • Kali's mcp-kali-server package (officially documented Feb 2026) exposes Nmap, Metasploit, SQLMap, and 10+ tools to Claude
  • The architecture is: Claude Desktop/Code → MCP → Kali/Burp → structured output → Claude analysis
  • Critical OPSEC warnings: prompt injection, tool poisoning, and cloud data leakage are real risks — treat MCP servers as untrusted code

Introduction: Why This Matters Now

In February 2026, Kali Linux officially documented a native AI-assisted penetration testing workflow using Anthropic's Claude via the Model Context Protocol (MCP). Weeks earlier, PortSwigger shipped their official MCP Server extension for Burp Suite. These aren't experimental toys — they represent a fundamental shift in how offensive security practitioners interact with their tooling.

Instead of memorising Nmap flags, crafting SQLMap syntax, or manually triaging hundreds of Burp proxy entries, you describe what you want in plain English. Claude interprets, plans, executes, and analyses — then iterates if needed. The entire recon-to-report loop becomes conversational.

This article walks you through the complete setup, the two Burp Suite integration paths, the Kali MCP architecture, practical prompt workflows, and — critically — the security risks you must understand before deploying this anywhere near a real engagement.


1. Understanding the Architecture

All three integration paths (Burp MCP, Burp AI Agent, Kali MCP) share the same core pattern: Claude communicates with your tools through MCP, a standardised protocol that Anthropic open-sourced in late 2024. Think of MCP as a universal API bridge that lets LLMs call external tools while maintaining session context.

You (Claude Desktop / Claude Code) Claude Sonnet (Cloud LLM) MCP Protocol Layer Kali / Burp Suite (Execution)

Structured Output Claude Analysis Tool Results

The three components in every setup are:

UI Layer Claude Desktop (macOS/Windows) or Claude Code (CLI). This is where you type prompts and receive results.
Intelligence Layer Claude Sonnet model (cloud-hosted). Interprets intent, selects tools, structures execution, analyses output.
Execution Layer Kali Linux (mcp-kali-server on port 5000) or Burp Suite (MCP extension on port 9876). Runs the actual commands.
Protocol Bridge MCP handles structured request/response between Claude and your tools over SSH (Kali) or localhost (Burp).

2. Path A: Burp Suite + Claude via PortSwigger's Official MCP Extension

PortSwigger maintains the official MCP Server extension in the BApp Store. It works with both Burp Pro and Community Edition.

Setup Steps

1Install the MCP Extension — Open Burp Suite → Extensions → BApp Store → search "MCP Server" → Install.

2Configure the MCP Server — The MCP tab appears in Burp. Default endpoint: http://127.0.0.1:9876. Enable/disable specific tools (send requests, create Repeater tabs, read proxy history, edit config).

3Install to Claude Desktop — Click "Install to Claude Desktop" button in the MCP tab. This auto-generates the JSON config. Alternatively, manually edit:

// macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
// Windows: %APPDATA%\Claude\claude_desktop_config.json

{
  "mcpServers": {
    "burp": {
      "command": "<path-to-java>",
      "args": [
        "-jar",
        "/path/to/mcp-proxy-all.jar",
        "--sse-url",
        "http://127.0.0.1:9876/sse"
      ]
    }
  }
}

4Restart Claude Desktop — Fully quit (check system tray), then relaunch. Verify under Settings → Developer → Burp integration active.

5Start Prompting — Claude now has access to your Burp proxy history, Repeater, and can send HTTP requests directly.


3. Path B: Burp AI Agent (six2dez) — The Power Option

The Burp AI Agent by six2dez is a more feature-rich alternative. It goes significantly beyond the official extension.

7 AI Backends Ollama, LM Studio, Generic OpenAI-compatible, Gemini CLI, Claude CLI, Codex CLI, OpenCode CLI
53+ MCP Tools Full autonomous Burp control — proxy, Repeater, Intruder, scanner integration
62 Vulnerability Classes Passive and Active AI scanners across injection, auth, crypto, and more
3 Privacy Modes STRICT / BALANCED / OFF — redact sensitive data before it leaves Burp

Setup

# Build from source (requires Java 21)
git clone https://github.com/six2dez/burp-ai-agent.git
cd burp-ai-agent
JAVA_HOME=/path/to/jdk-21 ./gradlew clean shadowJar

# Or download the JAR from Releases
# Load in Burp: Extensions → Add → Select JAR

Claude Desktop config for Burp AI Agent:

{
  "mcpServers": {
    "burp-ai-agent": {
      "command": "npx",
      "args": [
        "-y",
        "supergateway",
        "--sse",
        "http://127.0.0.1:9876/sse"
      ]
    }
  }
}
💡 Key advantage of Burp AI Agent: Right-click any request in Proxy → HTTP History → Extensions → Burp AI Agent → "Analyse this request" — opens a chat session with the AI analysis. The 3 privacy modes (STRICT/BALANCED/OFF) and JSONL audit logging with SHA-256 integrity hashing make it more suitable for professional engagements.

4. Kali Linux + Claude via mcp-kali-server

Officially documented by the Kali team in February 2026, mcp-kali-server is available via apt and exposes penetration testing tools through a Flask-based API on localhost:5000.

Supported Tools

ReconNmap, Gobuster, Dirb, enum4linux-ng
Web ScanningNikto, WPScan, SQLMap
ExploitationMetasploit Framework
Credential TestingHydra, John the Ripper

Setup

# On Kali Linux
sudo apt update
sudo apt install mcp-kali-server kali-server-mcp

# Start the MCP server
mcp-kali-server
# Runs Flask API on localhost:5000

Claude Desktop connects over SSH using stdio transport. Add to your config:

{
  "mcpServers": {
    "kali": {
      "command": "ssh",
      "args": [
        "kali@<KALI_IP>",
        "mcp-server"
      ]
    }
  }
}
💡 Linux Users: Claude Desktop has no official Linux build as of March 2026. Workarounds include WINE, unofficial Linux packages, or alternative MCP clients such as 5ire, AnythingLLM, Goose Desktop, and Witsy. Claude Code (CLI) works natively on Linux and is arguably the better option for Kali integration.

5. Practical Prompt Workflows — Optimising Your Skills

The integration is only as good as how you prompt it. Here are real-world workflow patterns that maximise Claude's value.

5.1 Recon Triage (Kali MCP)

"Run an Nmap service scan on 10.10.10.100 with version detection. If you find HTTP on any port, follow up with Gobuster using the common.txt wordlist. Summarise all findings with risk ratings."

Claude will chain: verify tool availability → execute nmap -sV → parse open ports → conditionally run gobuster → produce a structured summary with prioritised findings. One prompt replaces 3-4 manual steps.

5.2 Proxy History Analysis (Burp MCP)

"From the HTTP history in Burp, find all POST requests to API endpoints that accept JSON. Identify any that pass user IDs in the request body — I'm hunting for IDOR and BOLA vulnerabilities."

Claude reads your proxy history, filters by content type and method, identifies parameter patterns, and flags candidates for manual testing. This alone saves hours on large applications.

5.3 Automated Test Plan Generation (Burp MCP)

"Analyse the JavaScript files in Burp history. Extract API endpoints, identify authentication mechanisms, and generate a test plan covering OWASP API Security Top 10."

5.4 Collaborator-Assisted SSRF Testing (Burp MCP + Claude Code)

"Take the request in Repeater tab 1. Identify any parameters that accept URLs or hostnames. Create variations pointing to my Collaborator URL and send each one. Report back which triggered a DNS lookup."

5.5 Full Report Generation (Post-Engagement)

"Compile all findings from this session into a structured pentest report. Include: vulnerability title, severity (CVSS where possible), affected endpoint, proof of concept, and remediation steps."
💡 Skill Optimisation Tips:
Be specific with scope — "scan ports 1-1000" not just "scan the target"
Chain conditional logic — "if you find X, then do Y" leverages Claude's reasoning
Request structured output — "format as a markdown table" or "create Repeater tabs for each finding"
Use Claude Code over Desktop for Kali — CLI-native, works on Linux, better for multi-step chains
Iterate — Claude maintains session context, so you can refine: "now test that endpoint for SQLi"

6. Security Risks — Read This Before Deploying

This is where most guides stop. Don't be that person. MCP-enabled AI workflows introduce real, documented attack surfaces.

⚠️ CRITICAL: Known CVEs in MCP Ecosystem (January 2026)

Three vulnerabilities were disclosed in Anthropic's official Git MCP server, directly demonstrating that MCP servers are exploitable via prompt injection:

CVE-2025-68143 Path traversal via arbitrary path acceptance in git_init
CVE-2025-68144 Argument injection via unsanitised git CLI args in git_diff / git_checkout
CVE-2025-68145 Path validation weakness around repository scoping

Researchers demonstrated chaining these with a Filesystem MCP server to achieve code execution. This is not theoretical.

Threat Model for MCP-Assisted Pentesting

Prompt Injection: Malicious content in target responses (HTML, headers, error messages) can feed instructions back into Claude's reasoning loop. A target application could craft responses that manipulate Claude's next actions — classic "data becomes instructions" routed through a new control plane.

Tool Poisoning: CyberArk and Invariant Labs have documented scenarios where malicious instructions embedded in tool descriptions or command output can manipulate the LLM into unintended actions, including data exfiltration.

Cloud Data Leakage: Every prompt and tool output transits through Anthropic's cloud infrastructure. For client engagements with confidentiality requirements, this likely violates your engagement letter. Sending target data to a third-party API is a non-starter for most professional pentests.

Over-Permissioned Execution: The mcp-kali-server can execute terminal commands. A poorly scoped setup with root access is a catastrophic vulnerability if the LLM is manipulated.

Hardening Checklist

# OPSEC checklist for MCP-assisted pentesting

[ ] Run Kali in an isolated VM or container — disposable, no shared credentials
[ ] No SSH agent forwarding to the Kali execution host
[ ] Minimal outbound network — open only what you need
[ ] Use Burp AI Agent's STRICT privacy mode for client work
[ ] Enable JSONL audit logging with integrity hashing
[ ] Human-in-the-loop approval for destructive or high-risk commands
[ ] Never use on real client targets without explicit written authorisation for AI-assisted testing
[ ] Review all Claude-generated commands before execution on production targets
[ ] Treat MCP servers as untrusted third-party code — test for command injection, path traversal, SSRF
[ ] For air-gapped requirements: use Ollama + local models via Burp AI Agent instead of cloud Claude

7. Which Path Should You Choose?

PortSwigger MCP Extension ✅ Official, simple setup
✅ BApp Store install
❌ Fewer features
❌ No privacy modes
🎯 Best for: lab work, CTFs, learning
Burp AI Agent (six2dez) ✅ 53+ tools, 62 vuln classes
✅ 3 privacy modes + audit logging
✅ 7 AI backends (inc. local)
❌ Requires Java 21 build
🎯 Best for: professional engagements
Kali mcp-kali-server ✅ Full Kali toolset access
✅ Official Kali package
❌ Cloud dependency
❌ No Linux Claude Desktop
🎯 Best for: recon, enumeration, CTFs
Combined Stack ✅ Maximum coverage
✅ Burp for web + Kali for infra
❌ Complex setup
❌ Largest attack surface
🎯 Best for: comprehensive assessments

8. Conclusion: AI Won't Replace You — But It Will Change How You Work

Let's be clear about what this is and what it isn't. Claude + MCP is not autonomous pentesting. It doesn't exercise judgement, assess business impact, or make ethical decisions. What it does is eliminate the repetitive friction of context switching, command crafting, output parsing, and report formatting — the tasks that consume 60-70% of a typical engagement.

The practitioners who will thrive are those who use AI as an intelligent assistant while maintaining the critical thinking, methodology discipline, and OPSEC awareness that no LLM can replicate. Start with lab environments and CTFs. Build confidence with the tooling. Understand the security risks deeply. Then — and only then — consider how it fits into your professional workflow.

The command line remains powerful. Now it has a conversational layer. Use it wisely.


Sources & Further Reading

PortSwigger MCP Server ExtensionBurp AI Agent (six2dez)Kali Official Blog — LLM + Claude Desktopmcp-kali-server PackageSecEngAI — AI-Assisted Web PentestingPortSwigger MCP Server (GitHub)CybersecurityNews — Kali Integrates Claude AIModel Context Protocol (Official)Penligent — Critical Analysis of Kali + Claude MCP

#Claude #KaliLinux #BurpSuite #MCP #PenetrationTesting #AppSec #OffensiveSecurity #AIinCybersecurity #OSCP #BugBounty #ModelContextProtocol #altcoinwonderland

GitHub Actions as an Attacker's Playground

GitHub Actions as an Attacker's Playground — 2026 Edition CI/CD security • Supply chain • April 2026 ci-cd github-actions supply-c...