NETWORK TECHNOLOGY
Comprehensive IT solutions, support, security and automation for your business
Tech support: 24/7
+359 2 958 6535

OpenCart: How we stop search URLs and clean crawl budget in Google

How to stop OpenCart search and parameter URLs from being indexed, reduce “Crawled – not indexed” pages, and free crawl budget for categories and products.

OpenCart SEO cleanup: We stop the indexation of “junk” URLs (search, filters, parameters), clean up “Crawled – currently not indexed” pages and free crawl budget, so Google crawls categories and products more often. Remote setup (Zoom/AnyDesk) with live testing.

OpenCart: How to remove search pages from Google and free crawl budget

If you see a lot of “Crawled – currently not indexed” pages in Search Console and examples like index.php?route=product/search&tag=..., the problem is almost always the same. Google discovers a huge number of dynamic URLs with no SEO value. Below is the exact method to fix this, without breaking the store and without blocking important pages.

⚠ This is not “ranking magic”. This is SEO hygiene. If Google wastes resources on parameters and search combinations, real categories and products are crawled less often, indexing becomes slower and you get noise instead of SEO results.

Google Search Console – many Crawled currently not indexed pages

With OpenCart, it is very common for URLs to be crawled (or even indexed) that should never exist for Google. These include search pages, tag results, filter parameters, sorting, limits, pagination and endless combinations. The result is a massive number of “pages” that bring no traffic but consume crawl budget.

1) What “Crawled – currently not indexed” means

It means Google has crawled the page, analyzed its content and decided not to index it. By itself, this is not a problem. It becomes a problem when the number grows large, because Google spends time crawling pages that will never bring traffic.

In practice, Google is saying: “I see it, but I don’t need it”. Our goal is to make Google stop crawling it, or receive a clear signal that this page is not meant for search results.

2) Why this happens in OpenCart

The most common sources of noise are:

Important: these URLs may be useful for users as functionality, but they are toxic for SEO because they create duplicates and thin pages.

3) The goal: what we keep and what we restrict

The goal is not to hide the site from Google. The goal is for Google to see only valuable pages: categories, products, static pages, blog posts, landing pages. Everything that is an internal tool (search, filters, sorting) must be controlled.

We keep for Google
  • Categories
  • Products
  • Information pages
  • Blog / articles
  • Important landing pages
We restrict for Google
  • Search (route=product/search)
  • sort, order, limit
  • filter parameters
  • meaningless combinations and duplicates

4) Action plan (robots, meta, htaccess, sitemap)

We do this on multiple levels. One alone is not enough, the combination matters.

4.1 robots.txt, minimal and clean

robots.txt does not remove already indexed URLs, but it stops future crawling and reduces noise. A typical OpenCart example that works in 90% of cases:

User-agent: *
Disallow: /admin/
Disallow: /system/
Disallow: /storage/
Disallow: /vendor/

Disallow: /*?route=product/search
Disallow: /*?sort=
Disallow: /*&sort=
Disallow: /*?order=
Disallow: /*&order=
Disallow: /*?limit=
Disallow: /*&limit=

Allow: /

Sitemap: https://YOUR-DOMAIN.com/sitemap.xml

4.2 Meta robots for search and result pages

When we want the page to function for users but not be indexed, the cleanest solution is noindex, follow for search and result pages, implemented via controller logic or OCMOD.

4.3 Canonical for parameters

For category pages with sort, order, limit, canonical must point to the clean URL without parameters.

4.4 htaccess cleanup of tracking parameters

UTM, gclid, fbclid and similar parameters should be redirected to the clean URL with 301 to reduce duplication.

4.5 Sitemap with value-only URLs

The sitemap must contain categories and products only, not search, filters or parameters.

5) When 410 Gone makes sense

410 Gone is useful only for URLs that should not exist at all and are already indexed, for example mass-indexed search URLs. It should be used selectively, not as a blunt tool.

6) How we verify the effect

7) FAQ

Want us to clean OpenCart indexation and crawl budget?

  • Identify real indexed noisy URLs
  • Configure robots.txt, canonical, noindex and sitemap
  • Apply controlled 410 where needed
  • Verify live in Search Console

Email office@ntg.bg or request an SEO consultation.

The rule is simple: Google should crawl value, not noise. That’s how indexation becomes stable and predictable.

Comments

Loading…
Only registered and logged-in users can comment.