SEO & Marketing

Page Discovered but Not Indexed: Why It Happens and How to Fix It

E
Equip editorial Posicionament-Web
06 May 2026 9 min 7 views

Page Discovered but Not Indexed: Why It Happens and How to Fix It

You open Google Search Console, go to the indexation section and find a bunch of pages with the status "Discovered: pending indexation". The first temptation is to click «Request indexation» and wait. Wrong path. I've seen this situation dozens of times—in a clothing store in Sabadell, in a law firm in Barcelona, in a dental clinic in Tarragona—and it almost never resolves on its own. What does work is understanding why it happens and acting in the correct order.

What exactly does this status mean?

Google does two separate things: discover a URL and index it. Discovery means Googlebot has seen the address—via sitemap, via an internal or external link—but hasn't yet gone to read its content. Indexing means it has read it, evaluated it and decided to include it in search results.

When a page appears as «discovered but not indexed», Googlebot knows it exists but has decided to postpone its visit. It's not a technical error per se: it's a priority decision. And this is where you need to understand why Google makes this decision.

Status in Search ConsoleDiscovered: pending indexation
EnglishDiscovered – currently not indexed
Where to find itSearch Console → Indexation → Pages → filter by reason
Most common causeInsufficient crawl budget or low quality signal
Typical resolution time1–4 weeks once corrections are applied

Notice the difference with the status «Crawled: pending indexation»: in that case Googlebot has visited the page and decided not to index it—content quality problem. In the case of «Discovered», it hasn't even gone to read it. Often it's an architecture and priority problem, not a text problem.

Why doesn't Google crawl the page?

Rarely is there just one cause. What I typically find is a combination of two or three simultaneous factors. Here are the main ones, ordered by actual frequency:

Crawl budget exhausted by unnecessary URLs

The crawl budget is the number of pages Googlebot is willing to crawl on your domain in a period of time. When this budget is exhausted on pages without value, important pages remain uncrawled.

A typical case: a fashion store in Sabadell with 800 real products but 4,000 indexable URLs because size, color and price filters generate infinite combinations without noindex or canonical. Googlebot reaches the limit before reaching the products that matter to index. How you detect it: in Search Console, go to Settings → Crawl statistics. If the number of pages crawled per day is much lower than the total number of pages on the site, you have a crawl budget problem.

Insufficient domain authority

Google prioritizes crawling sites with good reputation. A new restaurant in Gràcia without any external backlinks can take weeks to see all its pages indexed, even if the content is good. It's not a penalty: it's that Googlebot has millions of sites to crawl and prioritizes those it considers more trustworthy. The solution here is not technical; it's getting first quality links.

Weak internal link structure

If the affected page receives no internal links from other pages on the site, Google considers it less relevant. The most common case I see: a clinic in Tarragona creates a new service page—for example, «invisible orthodontics»—but doesn't link it from the menu, from the homepage or from any blog article. Googlebot finds it in the sitemap but sees no internal signal of importance.

Content that Google doesn't consider a priority

Since the content usefulness updates, Google crawls thin pages or very similar to others on the same domain with less urgency. A service page of 150 words without structure, without a clear answer to any real user question, has a high probability of ending up in this status. And if you have two pages covering the same topic with slightly different words, Google may decide not to crawl either of them with priority.

Diagnosis first, solution later: Applying the same recipe to all cases is the most common mistake. An online store in Girona with 200 affected pages needs a different strategy than a law firm in Barcelona with 3 affected pages. Identify the real cause before acting.

How to fix it: real action order

This is where most articles fail: they tell you «improve content and request indexation» without telling you in what order or where to start. This is the process I follow:

  1. Clean up the XML sitemap (day 1). Check that it doesn't contain URLs with 404 errors, 301 redirects, noindex tags or pagination pages. Tools like Screaming Frog or Sitebulb do this in minutes. A messy sitemap is the silent saboteur of indexation.

  2. Block low-quality URLs (days 2–5). Add noindex to filter pages, tags, internal search results and deep pagination. In e-commerce, this step usually has the biggest impact because it frees up crawl budget at once. If you're unsure which pages to block, ask yourself: «Would someone search for exactly this URL on Google?» If the answer is no, apply noindex.

  3. Strengthen internal links (days 3–7). Add at least 2–3 links from pages with authority—homepage, main categories, blog articles with traffic—to each affected page. It's not optional: it's the most direct signal you give Google about the importance of a page. If no page on your site points to the affected URL, for Google it's as if it doesn't exist.

  4. Improve the content of affected pages (days 5–14). Make sure each important page has at least 400–500 useful words, answers a real user question and is clearly different from other pages on your site. If two pages cover the same topic, merge them into one and redirect the old URL.

  5. Request indexation manually (once the previous steps are done). In Search Console, use the URL inspection tool and click «Request indexation». Do this after improvements, never before. For large volumes of pages, update the sitemap date and resubmit it from Search Console.

  6. Work on external backlinks (parallel or later). One or two quality links to the affected pages—from a sector directory, a local news article or a sector collaborator—can significantly accelerate crawling, especially if the domain is relatively new.

Day 1
Clean up sitemap and detect unnecessary URLs
Days 3–7
Improve internal links to affected pages
Day 7+
Request indexation manually once everything is clean

Mistakes that make it worse

These are the mistakes I see repeatedly, regardless of whether the business is a store in Girona or a consultancy in Eixample:

  • Requesting indexation without fixing the cause. The most frequent. Google receives the request, checks that nothing has changed and ignores it. After a few days, the page returns to the same status. It's like calling the doctor without explaining your symptoms.
  • Sitemap inflated with worthless URLs. Pages like «thank you for your purchase», user profiles, internal search results or URLs with UTM parameters. Each unnecessary URL in the sitemap is noise that makes crawling the important ones harder.
  • Unmanaged pagination. Categories with /page/2, /page/3… up to /page/47 that consume crawl budget without adding value. The usual solution is noindex from page 2 onwards.
  • Incorrect or missing canonical on product variants. In e-commerce, each variant (size S, M, L; color red, blue) can generate an almost identical URL. Without canonical pointing to the main URL, Google sees duplicate content and crawls the entire set with less priority.
  • Confusing «discovered» with «crawled» not indexed. The causes and solutions are not the same. If you apply the wrong solution, you waste time and don't move forward.

When should you act today?

SituationUrgencyFirst action
Main service or product pages affected🔴 HighInternal links + manual request today
Dozens or hundreds of pages affected🔴 HighCrawl budget audit and sitemap cleanup
New site (less than 6 months) with few pages🟡 MediumBacklinks + quality content + patience
Secondary pages (old blog, archives)🟢 LowWait 3–4 weeks; if it persists, review quality

The situation that requires the most urgent action is when the affected pages are the main services or products. I've seen it with an aesthetic clinic in Tarragona that had treatment pages in this status for two months without knowing it. When we fixed the internal architecture and cleaned up the sitemap, organic traffic improved noticeably in the following weeks. Every week a key page isn't indexed is traffic going to the competition.

If you want to know exactly how many pages on your site are in this status and which ones need urgent attention, ask us for a free initial Search Console review. In less than 48 hours we'll tell you where the real problem is and where you should start.

Frequently Asked Questions

How long does it take Google to index a new page?

On established sites with good authority, it can be a matter of days. On new sites or with little traffic, it can take between 2 and 8 weeks. Requesting indexation manually in Search Console can speed up the process, but it's only effective if the content and structure are in order. If you request indexation of a page with problems, Google simply puts it back in the queue.

Can I force Google to index a page?

Not entirely. You can request indexation via Search Console, but Google decides whether to index it and when. What you can control is improving the conditions: useful and differentiated content, good internal links, clean sitemap and, if possible, some quality external backlink pointing to the page.

What's the difference between «discovered» and «crawled» but not indexed?

«Discovered» means Google knows the URL exists but hasn't visited it—crawl priority problem. «Crawled» means Googlebot has read the page and decided not to index it, often because the content isn't useful enough or is very similar to another page on the same site. The first is solved with architecture; the second, with content improvement.

Does crawl budget affect small sites?

On sites with fewer than 500–1,000 pages, crawl budget is rarely the main problem. If you have a small site with many unindexed pages, the problem is usually content quality, domain authority or weak internal link structure. Start by reviewing these three factors before worrying about crawl budget.

Should I worry if my site is new?

It's normal for a new site to take time to index all its pages. What you can do is focus on quality content, get the first backlinks and make sure the internal structure is clear and the sitemap is clean. With good practices, indexation improves progressively in 3–6 months. If after 6 months you still have important pages in this status, you need to review it in depth.

Want to improve your SEO in Catalonia?

Free SEO analysis: we tell you exactly where to start.

Free analysis


Share:
E
Equip editorial Posicionament-Web

L'equip editorial de Posicionament-Web publica continguts SEO pensats per a negocis de Catalunya.

Comments

No comments yet. Be the first to comment!

Leave a comment