robots.txt
robots.txt copied to clipboard
:robot: robots.txt as a service. Crawls robots.txt files, downloads and parses them to check rules through an API
Results
7
robots.txt issues
Sort by
recently updated
recently updated
newest added
As Google says in his documentation: >A robots.txt request is generally cached for up to one day, but may be cached longer in situations where refreshing the cached version is...
Use docker-compose v2.4 and use limit_mem directive
Like `ftp://ftp.example.com/robots.txt`
Like `dissallow`, `disalow` etc
Differentiate URLs sent from the crawl API and others sent by users (when its query is not found). In this case, the priority is to download the robots file of...
enhancement