Crawl delay tells the robot about the number of seconds to wait between successive requests to the same server.
Is robots.txt crawl-delay also for sub-pages?
-
25-07-2022 - |
Question
I wrote a crawler for a certain webpage. This webpage has a robots.txt which looks like this:
User-Agent: *
Crawl-delay: 30
So it's only allowed to crawl the site every 30 seconds (right?). But what about sub-pages? E.g. can I crawl the following sites without a delay of 30 seconds because they are different sub-pages:
www.mysite.com
www.mysite.com/category_1
www.mysite.com/category_2
Do I need to make a break of 30 seconds between these requests?
Thank you, Thomas.
Solution
Licensed under: CC-BY-SA with attribution
Not affiliated with StackOverflow