Provides functions to download and parse 'robots.txt' files.
Ultimately the package makes it easy to check if bots
(spiders, crawler, scrapers, ...) are allowed to access specific
resources on a domain.
Version: |
0.7.15 |
Depends: |
R (≥ 3.0.0) |
Imports: |
stringr (≥ 1.0.0), httr (≥ 1.0.0), spiderbar (≥ 0.2.0), future.apply (≥ 1.0.0), magrittr, utils |
Suggests: |
knitr, rmarkdown, dplyr, testthat, covr, curl |
Published: |
2024-08-29 |
DOI: |
10.32614/CRAN.package.robotstxt |
Author: |
Pedro Baltazar [aut, cre],
Peter Meissner [aut],
Kun Ren [aut, cph] (Author and copyright holder of list_merge.R.),
Oliver Keys [ctb] (original release code review),
Rich Fitz John [ctb] (original release code review) |
Maintainer: |
Pedro Baltazar <pedrobtz at gmail.com> |
BugReports: |
https://github.com/ropensci/robotstxt/issues |
License: |
MIT + file LICENSE |
URL: |
https://docs.ropensci.org/robotstxt/,
https://github.com/ropensci/robotstxt |
NeedsCompilation: |
no |
Materials: |
NEWS |
In views: |
WebTechnologies |
CRAN checks: |
robotstxt results |