diff options
author | Bryan Newbold <bnewbold@archive.org> | 2021-04-29 16:13:27 -0700 |
---|---|---|
committer | Bryan Newbold <bnewbold@archive.org> | 2021-04-29 16:13:27 -0700 |
commit | f65f8ffb559826fa9d2ec5fd34d630735519b5dd (patch) | |
tree | cbb1a6d32a0c2b0a5720ad776f7b15e5cc3574ca /fatcat_scholar/static/robots.allow.txt | |
parent | 6b1f87c12f7d40a3016910b214579a368c747df4 (diff) | |
download | fatcat-scholar-f65f8ffb559826fa9d2ec5fd34d630735519b5dd.tar.gz fatcat-scholar-f65f8ffb559826fa9d2ec5fd34d630735519b5dd.zip |
web: sitemap.xml serving
Diffstat (limited to 'fatcat_scholar/static/robots.allow.txt')
-rw-r--r-- | fatcat_scholar/static/robots.allow.txt | 16 |
1 files changed, 14 insertions, 2 deletions
diff --git a/fatcat_scholar/static/robots.allow.txt b/fatcat_scholar/static/robots.allow.txt index 3c40454..35f13a3 100644 --- a/fatcat_scholar/static/robots.allow.txt +++ b/fatcat_scholar/static/robots.allow.txt @@ -1,5 +1,17 @@ -# Allow most queries, but please don't crawl search results (or at least do so very politely, eg no parallel requests) +# Hello friends! +# If you are considering large or automated crawling, you may want to look at +# our catalog API (https://api.fatcat.wiki) or bulk database snapshots instead. +# by default, can crawl anything on this domain. HTTP 429 ("backoff") status +# codes are used for rate-limiting instead of any crawl delay specified here. +# Up to a handful concurrent requests should be fine. User-Agent: * -Disallow: /search Allow: / + +# crawling search result pages is expensive, so we do specify a long crawl delay for those +User-agent: * +Allow: /search +Crawl-delay: 5 + +Sitemap: /sitemap.xml +Sitemap: /sitemap-index-works.xml |