The Crawl-only pages signal shows URLs that Siteimp discovered during the crawl but did not find in the sitemap captured for this scan.
This is a scan-level signal. It answers: “Which crawled pages appear to be missing from the sitemap?”
What this signal means
A crawl-only page is a page that Siteimp found by following links or crawl discovery paths, but that was not listed in the sitemap data for this snapshot.
That means the crawl and sitemap do not fully agree.
This can be intentional, accidental, or a sign that the sitemap is stale.
How to read the table
The table lists the URLs found for this signal.
Each row is one URL that appeared in the crawl but not in the sitemap captured for the scan.
Why it matters
If a page should be part of the public, machine-visible site, it should usually appear in the sitemap.
Crawl-only pages can reveal:
- stale sitemap generation
- pages linked on the site but omitted from the sitemap
- pages that should not be publicly linked
- content that exists in navigation but not in the official site map
- mismatches between your site structure and your sitemap tooling
What to check next
Review the listed URLs.
For each URL, decide whether the page should be in the sitemap.
If yes, update the sitemap generator, front matter, CMS settings, or publishing rules.
If no, check whether the page should be linked from crawled pages.
If the sitemap was stale during the scan, regenerate it and run another scan.
Signal vs issue
This is the signal view. It lists crawl-only pages across the scan.
A Sitemap issue is the page-level version. It appears on an individual page when that page was crawled but missing from the sitemap.
Why this signal may be empty
If the page says No URLs were found for this signal in the current scan, Siteimp did not find crawl-only URLs for this signal in this snapshot.
That means the crawl and sitemap matched for this particular check, or no rows were available for this signal.