Chain-of-Verification meets stress testing and tri-modal reasoning.
Catch what normal reviews miss. Know what's genuinely robust.
A structured pipeline with tri-modal reasoning and stress testing
Detect the verification domain, collect artifacts, and establish ground truth from tests, specs, and production state.
Break artifacts into verifiable claims. Classify each by reasoning mode: deductive, inductive, or abductive.
Generate counter-questions, then abstract into recurring failure categories across the codebase.
Trace actual code paths. Apply hidden behavior probing and the modular adversarial scaffold.
Existence Question, Scale Shift, Time Travel, Requirement Inversion. Produce Survived: yes/no verdicts.
Findings with reasoning-aware confidence, anti-fabrication discipline, and Survived verdicts. Hypotheses reported separately.
Propose updates to TODO.md, SPEC.md, PLAN.md based on findings. Waits for user confirmation.
Every artifact type gets domain-specific verification with appropriate ground truth
Source changes, logic flows, edge cases, concurrency issues
Ground truth: tests, type system, specDesign decisions, spec coverage, dependency risks, breaking changes
Ground truth: requirements, constraints, patternsSchemas, migrations, contracts, backward compatibility
Ground truth: production schema, validation rulesTechnical docs, process docs, user-facing content, install guides
Ground truth: actual codebase, current API, git historyTest suite integrity, mock honesty, coverage accuracy, flaky detection
Ground truth: production code, requirements, coverage reportsAgent outputs, performance reports, research summaries
Ground truth: source material, cited referencesSeven research-backed methodologies working together
Decompose artifacts into verifiable claims, generate adversarial counter-questions, then verify each claim independently against ground truth.
Dhuliawala et al., 2023Identify recurring failure categories across the codebase: frequency assumptions, implicit ordering, stale state, missing completeness, silent fallthrough, assumed environment.
Anthropic, 2026Four probing strategies: indirect (trace execution), scaffolded (chain findings), cross-reference (claims vs reality), and absence (what's NOT there).
AuditBench — Anthropic, 2026Five modules: suspicion modeling, attack selection (blast radius × probability), plan synthesis, execution, and subtlety detection.
Anthropic, 2025Four techniques with forced variety: Existence Question (should this exist?), Scale Shift (10x? zero?), Time Travel (6 months later?), Requirement Inversion (exact opposite?). Produces Survived: yes/no verdicts.
Principles of Chaos EngineeringEach claim classified as deductive (verify against source), inductive (generalize from 3+ instances), or abductive (best explanation from observations). Abductive findings reported as hypotheses with alternatives, never as facts.
v3.5Confidence tied to reasoning mode: 80–100 requires cited source, 60–79 requires 3+ instances, 40–59 requires hypothesis with alternatives. Can't score >79 without citing where you looked. Absence claims must state what was checked.
v3.5Real bug categories discovered across production codebases
Built on 6 research references from Anthropic, academia, and industry. Not heuristics — proven methodology.
Verifies code, architecture, data, documentation, tests, and analysis — not just syntax or style.
Reasoning-aware confidence scoring. Can't claim absence without citing where you looked. Abductive findings are hypotheses, not facts.
Goes beyond individual bugs to find systemic failure categories that recur across your codebase.
Built-in agent trust scoring: tracks confirmed bugs, false positives, and clean reviews for multi-agent workflows.
Extracted from real game development. Caught critical bugs including timer resets, patrol flip-flops, and missing collision bounds.
The research foundation behind every verification step
Dhuliawala et al., 2023 — The core methodology: decompose, question, verify independently.
arxiv.org/abs/2309.11495 ↗ AuditingAnthropic, 2025 — Systematic approaches to automated code and behavior auditing.
alignment.anthropic.com ↗ Red-TeamingAnthropic, 2026 — Finding failure categories, not just individual issues.
alignment.anthropic.com ↗ ProbingAnthropic, 2026 — Benchmarks for hidden behavior probing and agent meta-verification.
alignment.anthropic.com ↗ ScaffoldAnthropic, 2025 — Modular adversarial scaffolding for systematic security testing.
alignment.anthropic.com ↗ StressHypothesis-driven resilience testing adapted for code review: formulate steady-state, vary conditions, observe, conclude.
principlesofchaos.org ↗git clone https://github.com/fullo/claude-adversarial-skill.git
cp -r claude-adversarial-skill/skills/adversarial-verify ~/.claude/skills/
claude marketplace add fullo-plugins https://github.com/fullo/claude-plugins-marketplace
claude plugin install adversarial-verify@fullo-plugins
Works with any tool that supports the Agent Skills format. Requires git for diff analysis.
Trigger the skill with natural language or the slash command
/adversarial-verify
run an adversarial review on my recent changes
CoV check the last commit
verify this code with total skepticism
verify the PLAN.md against the SPEC.md
adversarial check on this migration
verify this agent's analysis report
look for systemic failure patterns in the codebase
probe this function for hidden behaviors
verify the README matches the actual install process
stress test the auth module
what happens at 10x scale?
check if the planning agent's output is biased