The public corpus of code patterns that AI coding tools consistently generate incorrectly.
Academic benchmarks measure AI coding tools on curated test sets. AI Incident Database documents production failures. Neither documents the repeatable patterns that AI coding tools reliably produce wrong in normal developer workflows.
AI Regression Database fills that gap. Every pattern includes reproduction steps, correct alternatives, and detection rules.
npx @korext/regression-submit draftAnonymous submissions welcome.
npx @korext/regression-submit detectThis scans your repository for patterns matching all published ARDs.
ARD-YYYY-NNNN
Example: ARD-2026-0042
A pattern is an ARD candidate when:
- AI coding tools reliably produce incorrect or suboptimal code given recognizable prompt patterns.
- The pattern is reproducible at least 3 out of 10 attempts.
- The pattern has a clear correct alternative.
- The pattern is observed in normal developer workflows, not adversarial prompts.
- security: introduces security vulnerabilities
- correctness: output does not match intent
- performance: poor algorithmic or systems performance
- hallucination: nonexistent APIs, libraries, or documentation
- compliance: violates regulatory or industry standards
- maintainability: difficult to maintain
This database differs from academic benchmarks in critical ways:
- Practitioner observed: patterns from real developer workflows, not curated tests
- Reproducible: every pattern has reproduction steps
- Version tracked: patterns are retested against current AI tool versions
- Detection linked: every pattern maps to detection rules
- Time series: reproduction rates tracked over time
- Vendor responses: AI tool vendors can respond to patterns
The database includes an automated test harness that reproduces patterns against current AI tool versions weekly. When vendors ship fixes, patterns are marked fixed automatically with public recognition of the improvement.
- No AI tool shaming. This is neutral infrastructure.
- Reproducibility over anecdotes. Unrepeatable claims are rejected.
- Version awareness. Tools improve. We track improvement.
- No adversarial use. Patterns are for defense, not jailbreaking.
- Contributor attribution. Contributors get credit for their observations.
- Vendor notification. 7 day notification before publication.
See ETHICS.md.
| Project | Purpose |
|---|---|
| ai-attestation | Tracks which AI tools wrote code |
| ai-license | Declares AI authorship |
| supply-chain-attestation | Scans AI across dependencies |
| ai-incident-registry | Catalogs AI code failures (incidents) |
| ai-code-radar | Live statistics on AI adoption |
| ai-regression-database | Catalogs AI code patterns that are wrong (patterns) |
ARD documents patterns. AICI documents incidents. Patterns can cause incidents. Incidents reference patterns.
See API documentation.
See SPEC.md. Released under CC0 1.0.
All pattern data is released under CC BY 4.0. Attribution to the database and contributor is required.
See PRIOR_ART.md.
See CONTRIBUTING.md.
Korext builds AI code governance tools. AI Regression Database is an open community resource maintained by the Korext team.