block 2 · online
Robots.txt + Sitemap + Meta Analyzer · dev-docs

Common Mistakes the Analyzer Surfaces

A checklist of the most frequent configuration errors this tool will catch.

status
Published
slug
common-mistakes
published
Apr 20, 2026

Common Mistakes the Analyzer Surfaces

These are the patterns curators see most often when running this tool on real public sites.

robots.txt

  • A Disallow: /admin rule that inadvertently documents the admin path to everyone.
  • A sitemap declared in robots.txt that 404s or points to a stale domain.
  • Contradictory User-agent: * and bot-specific rules.

Sitemaps

  • Stale lastmod values that haven't changed in years.
  • Inclusion of noindex pages — the site is asking Google to crawl pages it has also asked Google not to index.
  • Single sitemap files over 50,000 URLs (the protocol limit) or 50 MB uncompressed.

Meta tags

  • A canonical pointing at the wrong scheme (http:// vs. https://).
  • A <meta name="robots" content="noindex"> left in place after launch.
  • Open Graph images that are missing or smaller than the 1200×630 minimum.
last published Apr 20, 2026