highDistribution Readiness

Sitemap XML Reachable

A reachable sitemap.xml is a baseline crawler requirement. AI and search crawlers rely on it to discover the pages on your site. SaaSalyst fetches your sitemap and verifies it parses as valid XML with the right root element.

What SaaSalyst Checks

SaaSalyst checks the sitemap declared in robots.txt (via the Sitemap: directive); if none is declared, it falls back to /sitemap.xml. The check passes when the URL returns 200 and the response parses as XML with <urlset> or <sitemapindex> as the root element. It fails on 404, malformed XML, or wrong root element.

Why This Matters

Without a sitemap, crawlers must rely on internal links to discover your pages — which means new content, deep pages, and pages with few inbound links may take days or weeks to appear in search and AI results.

The XML format matters too. Crawlers reject sitemaps that don't conform to the sitemaps.org schema (root <urlset> or <sitemapindex>). A common failure mode: server returns 200 but the body is HTML (often a soft 404 or a Vue/React app catching the route) — the crawler ignores it.

This is a high-severity check because the cost of a missing or broken sitemap is silent: pages just don't get indexed, and you don't see the error.

How to Fix It

  1. Generate a sitemap.xml using your framework's built-in tools (Next.js: app/sitemap.ts; Gatsby: gatsby-plugin-sitemap; WordPress: most SEO plugins).
  2. Verify the URL returns 200 and Content-Type: application/xml (not text/html).
  3. The root element must be <urlset> or <sitemapindex>. <urlset> contains direct <url> entries; <sitemapindex> references child sitemaps.
  4. Reference the sitemap URL in robots.txt via a Sitemap: directive — see robots_txt_sitemap_directive.

Frequently Asked Questions

What if my sitemap is at a non-standard path?

Declare it in robots.txt via a Sitemap: directive. SaaSalyst reads that directive first and falls back to /sitemap.xml only when no directive is declared. Crawlers behave the same way.

Why does sitemap.xml need <urlset> or <sitemapindex>?

Those are the two root elements defined by the sitemaps.org spec. Crawlers conform to the spec, so anything else (e.g., <html>, <feed>, custom tags) is ignored. SaaSalyst checks the actual root element to catch the case where the URL returns 200 with the wrong content.

How does this check affect my Business Readiness Score?

SaaSalyst rates a missing or invalid sitemap as high severity in Distribution Readiness. Sitemap accessibility is foundational for crawler-based discovery — broken sitemap = broken organic acquisition.

Check Your SaaS Now | Free

SaaSalyst scans your website in 30 seconds and checks for Sitemap XML Reachable along with 101+ other business readiness signals.

Scan Your App

Related Checks SaaSalyst Runs