Critical — Affects Indexing
Discovered — Currently Not Indexed
Google found your URL but hasn't crawled it — the page is stuck in the crawl queue. This typically signals a crawl budget bottleneck: Google knows the page exists but doesn't think it's worth spending resources on.
Scans 50+ pages for crawl budget signals. No account required.
What Does This Mean?
Unlike "Crawled — currently not indexed" where Google saw your content and rejected it, "Discovered — currently not indexed" means Google hasn't even visited the page yet. Your URL was found through sitemaps, internal links, or external links, but Googlebot decided it's not worth the crawl budget to fetch it right now.
This is especially common on large sites (100k+ pages) where Google allocates limited crawl budget. Pages competing for crawl quota with thousands of low-value URLs often get stuck in this state for weeks or months.
Common Causes
- Crawl budget exhaustion. Too many URLs in your sitemap relative to the crawl budget Google allocates for your site. Large faceted navigation, infinite scroll pagination, and parameter URLs are the usual culprits.
- Slow server response. If your server takes >500ms to respond, Google reduces crawl rate. Check TTFB under load.
- Weak link signals. Pages that are only reachable through sitemaps (no internal links) get deprioritized in the crawl queue.
- Recent bulk URL additions. If you just added thousands of new pages (e.g., pSEO launch), Google will process them gradually over weeks.
- Sitemap bloat. Submitting sitemaps with many low-quality URLs teaches Google that your sitemaps aren't reliable signals — causing it to deprioritize all URLs.
How to Fix It
- Audit your sitemap. Remove low-value URLs (parameter variants, empty categories, login/cart pages). Only submit URLs you want indexed.
- Improve internal linking. Link to priority pages from your highest-traffic and highest-authority pages. Orphan pages get deprioritized.
- Optimize server performance. Target <200ms TTFB. Use caching layers (CDN, reverse proxy). Google crawls faster sites more aggressively.
- Use URL Inspection. For the most important stuck pages, manually request crawling via GSC's URL Inspection tool.
- Consolidate thin content. Merge similar thin pages into fewer, more comprehensive pages via 301 redirects.
How SEODiff Detects This
SEODiff's Spider Trap Architecture Map identifies crawl budget waste — infinite scroll traps, faceted navigation explosions, and pagination bloat that prevent Google from discovering your valuable content. The Indexing Predictions module flags pages likely to enter the "Discovered" purgatory based on link depth, content quality, and crawl efficiency metrics.