Disallowed URL in XML Sitemaps

No Comments

Quick Reference

Element Code: XM-003

Issue: Sitemap includes URLs blocked by robots.txt

Impact: Conflicting signals, wasted crawl attempts

Fix: Remove disallowed URLs from sitemap

Detection: Screaming Frog, Google Search Console

What Is This Issue?

Including robots.txt blocked URLs in sitemaps sends contradictory signals. The sitemap says "index this" but robots.txt says "do not crawl."

Why This Matters for Your Website

Google may report these as errors in Search Console. It wastes crawl budget on URLs that cannot be accessed.

How to Fix This Issue

  1. Remove from sitemap: Exclude disallowed URLs
  2. Or unblock: If URLs should be crawled
  3. Configure generator: Respect robots.txt in sitemap generation

Tools for Detection

  • Google Search Console: Reports sitemap errors

TL;DR (The Simple Version)

Your sitemap includes URLs blocked by robots.txt. Either remove them from the sitemap or unblock them.

About SEO ProCheck

Technical SEO consulting and GEO strategy with 20 years of enterprise experience. Case studies, resources, and tools for search and AI visibility.

Work With Me

Technical SEO audits, GEO strategy, site migrations, and international SEO. Hourly consulting for teams who need hands-on support, not just reports.

Subscribe to our newsletter!

More from our blog