Posts

Claude Code Hooks: The Deterministic Security Layer Your AI Agent Needs

Claude Code Hooks: The Deterministic Security Layer Your AI Agent Needs > APPSEC_ENGINEERING // CLAUDE_CODE // FIELD_REPORT Claude Code Hooks: The Deterministic Security Layer Your AI Agent Needs CLAUDE.md rules are suggestions. Hooks are enforced gates. exit 2 = blocked. No negotiation. If you're letting an AI agent write code without guardrails, here's how you fix that. // March 2026 • 12 min read • security-first perspective Why This Matters (Or: How Your AI Agent Became an Insider Threat) Since the corporate suits decided to go all in with AI (and fire half of the IT population), the market has changed dramatically, let's cut through the noise. The suits in the boardroom are excited about AI agents. "Autonomous productivity!" they say. "Digital workforce!" they cheer. Meanwhile, those of us who actually hack things for a living are watching these agents get deployed with shell access, API keys, and service-l...

Connecting Claude AI with Kali Linux and Burp Suite via MCP

🔗 Connecting Claude AI with Kali Linux & Burp Suite via MCP The Practical Guide to AI-Augmented Penetration Testing in 2026 📅 March 2026 ✍️ altcoinwonderland ⏱️ 15 min read 🏷️ AppSec | Offensive Security | AI ⚡ TL;DR MCP (Model Context Protocol) bridges Claude AI with Kali Linux and Burp Suite, enabling natural-language-driven pentesting PortSwigger's official MCP extension and six2dez's Burp AI Agent are the two primary integration paths for Burp Suite Kali's mcp-kali-server package (officially documented Feb 2026) exposes Nmap, Metasploit, SQLMap, and 10+ tools to Claude The architecture is: Claude Desktop/Code → MCP → Kali/Burp → structured output → Claude analysis Critical OPSEC warnings : prompt injection, tool poisoning, and cloud data leakage are real risks — treat MCP servers as untrusted code Introduction: Why T...

πŸ’€ JAILBREAKING THE PARROT: HARDENING ENTERPRISE LLMs

The suits are rushing to integrate "AI" into every internal workflow, and they’re doing it with the grace of a bull in a china shop. If you aren't hardening your Large Language Model (LLM) implementation, you aren't just deploying a tool; you're deploying a remote code execution (RCE) vector with a personality. Here is the hardcore reality of securing LLMs in a corporate environment. 1. The "Shadow AI" Black Hole Your devs are already pasting proprietary code into unsanctioned models. It’s the new "Shadow IT." The Fix: Implement a Corporate LLM Gateway . Block direct access to openai.com or anthropic.com at the firewall. The Tech: Force all traffic through a local proxy (like LiteLLM or a custom Nginx wrapper) that logs every prompt, redacts PII/Secrets using Presidio , and enforces API key rotation. 2. Indirect Prompt Injection (The Silent Killer) This is where the real fun begins. If your LLM has access to the web or internal docs (RAG...

πŸ›‘️ Claude Safety Guide for Developers

Claude Safety Guide for Developers (2026) — Securing AI-Powered Development Application Security Guide — March 2026 πŸ›‘️ Claude Safety Guide for Developers Securing Claude Code, Claude API & MCP Integrations in Your SDLC πŸ“‘ Contents Why This Guide Exists The AI Developer Threat Landscape in 2026 Real-World CVEs: Claude Code Vulnerabilities Understanding Claude Code's Permission Model Prompt Injection: Attack Vectors & Defences MCP (Model Context Protocol) Security AI Supply Chain Risks Claude API Safety Best Practices Claude Code Hardening Checklist Integrating Claude Security into CI/CD Compliance Considerations (SOC 2, GDPR, AI Act) Resources & References 1. Why This Guide Exists AI-powered development tools have moved from novelty to necessity. Anthropic's Claude ecosystem — spanning Claude Code (terminal-based agentic coding), Claude API (programmatic integration), and the broader Model Context Protocol (MCP) integrati...