Environment assumptions
Repository and stale-repository defaults update from the developer count using Arnica's benchmark cohorts: SMB under 100 developers, Midmarket from 100 to 999, and Enterprise at 1,000 or more. Active and stale repositories can use different full-scan cadences.
Select up to 3 models
Prices match the model list from the reference page. Select up to three models here, then compare their estimated scan costs in section 3.
Select up to 3 models to compare in the estimate table below.
Annual provider cost
This is the estimated amount paid to the AI model provider for the selected workload. It excludes engineering time, orchestration infrastructure, storage, CI minutes, alert routing, and triage operations.
How the estimate is calculated
The calculator separates workload volume from model token economics so finance and engineering can test each assumption independently.
Workload
Developers set the default repository and stale-repository assumptions by company segment. Manual repository and stale overrides are preserved until reset.
Per event cost
Per-event scan costs are estimated from calibrated provider benchmarks and each model's published input/output pricing.
Annual cost
The general-purpose estimate lets active and stale repositories use separate full-scan cadences, while PR scan volume follows developer activity.
Provider tokens are only one part of operating AI code scanning.
This calculator exists because executives are watching models like Claude Mythos demonstrate source-visible vulnerability discovery and asking a reasonable question: what unknown risk is already sitting in the backlog, and how do we prevent the next wave of AI-detectable risk from entering the codebase?
Backlog discovery
Frontier security models raise the urgency of finding latent vulnerabilities before attackers or auditors do. The immediate question is no longer whether AI can find buried issues; it is how often you can afford to look across all repositories.
Forward prevention
Backlog scans are only half the problem. New pull requests and active repositories need continuous coverage so fresh risk is caught while developers still have context and can fix it quickly.
Model choice
Public follow-up testing of Mythos showcase cases found that multiple cheaper and smaller models recovered the same vulnerability analyses. That makes orchestration, context, and scan design as important as always choosing the most expensive model.
Caching and dedupe
Repeatedly scanning unchanged code pays the model provider again. Arnica's offering is designed around reusing prior analysis and reducing repeat token spend on unchanged code.
Active vs stale repos
Stale repositories still need periodic coverage for newly disclosed vulnerability classes, but active repositories drive most continuous scan activity. Treating both the same can inflate spend.
Operational controls
Provider limits, key management, monthly budgets, retry behavior, failed scans, and silent workflow breaks all need owners. Arnica provides the control plane to operate this safely without turning it into another platform burden.
Developer workflow
Raw model output still needs routing, prioritization, suppression, fix context, and developer-native comments. Otherwise token spend can turn into alert volume instead of resolved risk.
Model upgrades
Frontier models, pricing, and deprecation timelines change quickly. Arnica can help route the right work to the right model as price and quality shift.
Cost reduction path
If this estimate is higher than expected, Arnica can help reduce the provider bill through scan orchestration, caching, dedupe, active-repo targeting, and model operations built for application security.
Arnica helps make AI code scanning better, faster, and cheaper by reducing repeated work and bringing the results into developer workflows.
Talk to Arnica- Arnica benchmark defaults: 13 PRs per developer per month; SMB under 100 developers with repository ratio 2.7x and stale 45%; Midmarket 100-999 developers with repository ratio 4.0x and stale 60%; Enterprise at 1,000+ developers with repository ratio 4.5x and stale 47%.
- Claude PR benchmark: Anthropic's Claude Code Review announcement references reviews averaging $15-$25 per PR, billed on token usage.
- Claude full-audit benchmark: Insight Services APAC's security audit cost experiment used Opus 4.7 with a 1M context window and reported substantial per-project audit spend. This calculator uses an editable $250 SOTA full-scan planning anchor.
- Mythos and model-choice context: Anthropic's Claude Mythos Preview describes source-visible vulnerability discovery, while AISLE's Mythos jagged-frontier reproduction materials report that multiple smaller or cheaper models recovered the same public showcase vulnerability analyses.
- Provider pricing references from the source page: Anthropic Claude API pricing, OpenAI API pricing, and Google Gemini API pricing as of May 2026.