We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The robots.txt standard allows for declaring the location of sitemaps (plural!), e.g. for https://www.nytimes.com/robots.txt :
# .... User-Agent: omgili Disallow: / User-agent: ia_archiver Disallow: / Sitemap: https://www.nytimes.com/sitemaps/new/news.xml.gz Sitemap: https://www.nytimes.com/sitemaps/new/sitemap.xml.gz Sitemap: https://www.nytimes.com/sitemaps/new/collections.xml.gz Sitemap: https://www.nytimes.com/sitemaps/new/video.xml.gz Sitemap: https://www.nytimes.com/sitemaps/new/cooking.xml.gz Sitemap: https://www.nytimes.com/sitemaps/new/recipe-collects.xml.gz Sitemap: https://www.nytimes.com/sitemaps/new/regions.xml Sitemap: https://www.nytimes.com/sitemaps/new/best-sellers.xml Sitemap: https://www.nytimes.com/sitemaps/www.nytimes.com/2016_election_sitemap.xml.gz Sitemap: https://www.nytimes.com/elections/2018/sitemap Sitemap: https://www.nytimes.com/wirecutter/sitemapindex.xml
It would be great if sitemapper allowed to process URLs to robots.txt in order to transiently return all Sitemap URLs.
robots.txt
The text was updated successfully, but these errors were encountered:
No branches or pull requests
The robots.txt standard allows for declaring the location of sitemaps (plural!), e.g. for https://www.nytimes.com/robots.txt :
It would be great if sitemapper allowed to process URLs to
robots.txt
in order to transiently return all Sitemap URLs.The text was updated successfully, but these errors were encountered: