<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Research on Rex Coleman</title><link>https://rexcoleman.dev/categories/research/</link><description>Securing AI from the architecture up. Research, tools, and methodology for AI security. Creator of govML.</description><image><title>Rex Coleman</title><url>https://rexcoleman.dev/images/og-default.png</url><link>https://rexcoleman.dev/images/og-default.png</link></image><generator>Hugo</generator><language>en-us</language><lastBuildDate>Tue, 31 Mar 2026 12:00:00 +0000</lastBuildDate><atom:link href="https://rexcoleman.dev/categories/research/index.xml" rel="self" type="application/rss+xml"/><item><title>AI Security Research → OWASP, NIST, and MITRE Standards Mapping</title><link>https://rexcoleman.dev/posts/ai-security-standards-mapping/</link><pubDate>Tue, 31 Mar 2026 12:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/ai-security-standards-mapping/</guid><description>Cross-reference between 8 original AI security research projects and OWASP LLM Top 10, OWASP Agentic Apps, NIST AI RMF, and MITRE ATLAS. Start from any framework, find relevant research.</description></item><item><title>Our Simulation Was Wrong by 37 Percentage Points — What Real LLM Agents Taught Us About Multi-Agent Cascade</title><link>https://rexcoleman.dev/posts/multi-agent-security/</link><pubDate>Fri, 20 Mar 2026 12:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/multi-agent-security/</guid><description>Simulation predicted 97% cascade poison. Real Claude agents: 60%. Topology matters (simulation said it didn&amp;#39;t). The simulation-to-real gap changes everything.</description></item><item><title>Your AI Makes SQL Injection Worse: CWE-Stratified Patch Safety for LLM Code Generation</title><link>https://rexcoleman.dev/posts/llm-patch-correctness/</link><pubDate>Fri, 20 Mar 2026 12:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/llm-patch-correctness/</guid><description>LLM-generated patches have a 42% fix rate and 10% regression rate. SQL injection patches are net-negative — 0% fix, 50% regression.</description></item><item><title>How Many Rewrites to Strip a Watermark? Empirical Paraphrase-Removal Curves for LLM Watermarks</title><link>https://rexcoleman.dev/posts/llm-watermark-robustness/</link><pubDate>Fri, 20 Mar 2026 11:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/llm-watermark-robustness/</guid><description>Cross-model paraphrasing drops watermark detection from 100% to 60% in one pass, then plateaus at 40% after 10 passes. Kirchenbauer green-list watermarks are partially robust — but not enough for adversarial settings.</description></item><item><title>Privilege Escalation Cascades at 98% While Domain-Aligned Attacks Are Invisible</title><link>https://rexcoleman.dev/posts/agent-semantic-resistance/</link><pubDate>Fri, 20 Mar 2026 10:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/agent-semantic-resistance/</guid><description>First taxonomy of why real LLM agents resist cascade poisoning — and which attacks bypass each resistance pattern.</description></item><item><title>Your AI Can't Beat EPSS at Vulnerability Triage (But the Ensemble Might)</title><link>https://rexcoleman.dev/posts/agent-vuln-triage/</link><pubDate>Fri, 20 Mar 2026 10:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/agent-vuln-triage/</guid><description>An LLM agent achieves 92% precision@10 on vulnerability triage but underperforms EPSS (100%). The ensemble reaches 98% with lower variance.</description></item><item><title>We Built a Multi-Agent Defense and It Failed — Here's Why That Matters More</title><link>https://rexcoleman.dev/posts/verified-delegation-protocol/</link><pubDate>Thu, 19 Mar 2026 22:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/verified-delegation-protocol/</guid><description>We proposed a 3-layer defense for multi-agent cascade. Real agent experiments refuted 5 of 7 hypotheses. The simulation was wrong by 48 percentage points.</description></item><item><title>A CFA Charterholder Built an ML Fraud Detector: Here's What the Models Miss</title><link>https://rexcoleman.dev/posts/financial-anomaly-detection/</link><pubDate>Thu, 19 Mar 2026 00:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/financial-anomaly-detection/</guid><description>CFA-informed rule-based scoring achieves 0.898 AUC on its own, and 8 of the top 20 predictive features come from domain expertise, not raw data.</description></item><item><title>I Built a PQC Migration Scanner: Here's What Your Codebase Is Hiding</title><link>https://rexcoleman.dev/posts/pqc-migration-analyzer/</link><pubDate>Thu, 19 Mar 2026 00:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/pqc-migration-analyzer/</guid><description>70% of the crypto in your codebase isn&amp;#39;t yours to change — and classical exploit risk matters more than quantum vulnerability for deciding what to fix first.</description></item><item><title>Beyond Prompt Injection: RL Attacks on AI Agent Decision-Making</title><link>https://rexcoleman.dev/posts/rl-agent-vulnerability/</link><pubDate>Mon, 16 Mar 2026 22:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/rl-agent-vulnerability/</guid><description>I trained RL agents on security tasks, then attacked their reward functions, observations, and policies. Observation perturbation is 20-50x more effective than reward poisoning. Policy extraction achieves 72% agreement with just 500 queries.</description></item><item><title>Antivirus for AI Models: Behavioral Fingerprinting Detects What Static Analysis Misses</title><link>https://rexcoleman.dev/posts/model-fingerprinting/</link><pubDate>Mon, 16 Mar 2026 00:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/model-fingerprinting/</guid><description>How do you know a model downloaded from Hugging Face hasn&amp;#39;t been backdoored? I built a behavioral fingerprinting system that uses unsupervised anomaly detection to answer that question.</description></item><item><title>I Red-Teamed AI Agents: Here's How They Break (and How to Fix Them)</title><link>https://rexcoleman.dev/posts/agent-redteam/</link><pubDate>Mon, 16 Mar 2026 00:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/agent-redteam/</guid><description>19 attack scenarios against LangChain and CrewAI agents. 100% success with reasoning chain hijacking. 7 attack classes systematized — 5 not in OWASP or MITRE ATLAS.</description></item><item><title>One Principle, Six Domains: Adversarial Control Analysis for AI Security</title><link>https://rexcoleman.dev/posts/adversarial-control-analysis/</link><pubDate>Mon, 16 Mar 2026 00:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/adversarial-control-analysis/</guid><description>The same security principle — classify inputs by who controls them — works across network IDS, vulnerability management, AI agents, post-quantum crypto, fraud detection, and AI supply chains.</description></item><item><title>Adversarial ML on Network Intrusion Detection: What Adversarial Control Analysis Reveals</title><link>https://rexcoleman.dev/posts/adversarial-ids/</link><pubDate>Sat, 14 Mar 2026 00:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/adversarial-ids/</guid><description>I built and red-teamed an ML-based intrusion detection system. The key finding: which features an attacker controls matters more than which model you choose.</description></item><item><title>Why CVSS Gets It Wrong: ML-Powered Vulnerability Prioritization</title><link>https://rexcoleman.dev/posts/cvss-gets-it-wrong/</link><pubDate>Sat, 14 Mar 2026 00:00:00 +0000</pubDate><guid>https://rexcoleman.dev/posts/cvss-gets-it-wrong/</guid><description>I trained an ML model on 338,000 CVEs to find out what actually predicts exploitation. CVSS scores severity. Attackers measure opportunity. The model reveals what they look for.</description></item></channel></rss>