Simple pricing
One plan. Cancel anytime. No long-term contract.
Free preview
$0
No card required
- ✓ Scan the first 100 URLs of any site
- ✓ Top 50 highest-risk pages shown
- ✓ Flag breakdown
- ✗ No downloadable report file
- ✗ No full-site crawl
CrawlAudit
$9.99/month
Cancel anytime
- ✓ Up to 5,000 URLs per scan
- ✓ 5 full-site scans per month
- ✓ Unlimited Markdown + CSV downloads
- ✓ Per-pattern risk roll-up
- ✓ Re-run scans to track fixes
- ✓ Email support
What you get on a paid plan
- • Up to 5,000 URLs per scan (sitemap discovery + chunked crawl).
- • 5 full-site scans every 30 days — re-run after fixes to verify the delta.
- • Baseline diff vs your previous scan: which findings are new, which got resolved.
- • Top fixes ranked by total risk reduction — bigger leverage first.
- • Fetch-fidelity card on every report (so you know exactly what was inspected).
- • Markdown report (human-readable) plus CSV (spreadsheet-ready) on every scan.
- • Per-pattern roll-up so you can see which section is dragging the site down.
Common questions
Can I cancel any time?
Yes. Cancellation takes effect at the end of the current billing month. You keep access until then and can re-subscribe whenever.
Do unused scans carry over?
No — the 5-scan allotment resets monthly. Most sites scan once after a content push, then once more after fixes ship.
What if my site has more than 5,000 URLs?
Rotate scans across sections by pointing CrawlAudit at sub-sitemaps (e.g. /sitemaps/posts.xml separately from /sitemaps/products.xml). Email us if you need a custom higher cap.
How is this different from Screaming Frog or Sitebulb?
Those tools cover technical SEO (broken links, status codes). CrawlAudit specifically scores content quality against Helpful Content signals — thin pages, AI footprints, templated intros, soft 404s, broken schema. We score what they don't.
Will the bot hammer my server?
No. We fetch at 20 concurrent requests, 15-second timeout. A 5,000-URL scan is roughly one normal browsing session spread over 10-15 minutes. The user-agent is CrawlAuditBot/1.0 so you can identify or block us if needed.
Do you execute JavaScript?
No — static HTML only, by design. This mirrors what Googlebot uses for its first-pass index, which is the index your traffic actually depends on. JS-only content is invisible to both us and Google's initial crawl.