Project Status: Active – The project has reached a stable, usable state and is being actively developed. Signed by Signed commit %Linux build Status Windows build status Coverage Status cran checks CRAN status Minimal R VersionLicense

Parse and Test Robots Exclusion Protocol Files and Rules

Description

The ‘Robots Exclusion Protocol’ https://www.robotstxt.org/orig.html documents a set of standards for allowing or excluding robot/spider crawling of different areas of site content. Tools are provided which wrap The ‘rep-cpp’ https://github.com/seomoz/rep-cpp C++ library for processing these ‘robots.txt’ files.

What’s Inside the Tin

The following functions are implemented:

  • can_fetch: Test URL paths against a robxp robots.txt object
  • crawl_delays: Retrive all agent crawl delay values in a robxp robots.txt object
  • print.robxp: Custom printer for ’robxp“ objects
  • robxp: Parse a ‘robots.txt’ file & create a ‘robxp’ object
  • sitemaps: Retrieve a character vector of sitemaps from a parsed robots.txt object

Installation

remotes::install_git("https://git.rud.is/hrbrmstr/spiderbar.git")
# or
remotes::install_git("https://git.sr.ht/~hrbrmstr/spiderbar")
# or
remotes::install_gitlab("hrbrmstr/spiderbar")
# or
remotes::install_bitbucket("hrbrmstr/spiderbar")
# or
remotes::install_github("hrbrmstr/spiderbar")

NOTE: To use the ‘remotes’ install options you will need to have the {remotes} package installed.

Usage

library(spiderbar)
library(robotstxt)

# current verison
packageVersion("spiderbar")
## [1] '0.2.2'

# use helpers from the robotstxt package

rt <- robxp(get_robotstxt("https://cdc.gov"))

print(rt)
## <Robots Exclusion Protocol Object>

# or 

rt <- robxp(url("https://cdc.gov/robots.txt"))

can_fetch(rt, "/asthma/asthma_stats/default.htm", "*")
## [1] TRUE

can_fetch(rt, "/_borders", "*")
## [1] FALSE

gh_rt <- robxp(robotstxt::get_robotstxt("github.com"))

can_fetch(gh_rt, "/humans.txt", "*") # TRUE
## [1] TRUE

can_fetch(gh_rt, "/login", "*") # FALSE
## [1] FALSE

can_fetch(gh_rt, "/oembed", "CCBot") # FALSE
## [1] FALSE

can_fetch(gh_rt, c("/humans.txt", "/login", "/oembed"))
## [1]  TRUE FALSE FALSE

crawl_delays(gh_rt)
##                agent crawl_delay
## 1             yandex          -1
## 2         twitterbot          -1
## 3              teoma          -1
## 4            httrack          -1
## 5          googlebot          -1
## 6         telefonica          -1
## 7        ia_archiver          -1
## 8                  *          -1
## 9         etaospider          -1
## 10            dotbot          -1
## 11          rogerbot          -1
## 12       duckduckbot          -1
## 13            coccoc          -1
## 14             ccbot          -1
## 15            msnbot          -1
## 16             slurp          -1
## 17            daumoa          -1
## 18       mail.ru_bot          -1
## 19           bingbot          -1
## 20  intuitgsacrawler          -1
## 21          naverbot          -1
## 22         seznambot          -1
## 23          swiftbot          -1
## 24 red-app-gsa-p-one          -1
## 25        sanddollar          -1

imdb_rt <- robxp(robotstxt::get_robotstxt("imdb.com"))

crawl_delays(imdb_rt)
##   agent crawl_delay
## 1     *          -1

sitemaps(imdb_rt)
## character(0)

spiderbar Metrics

Lang # Files (%) LoC (%) Blank lines (%) # Lines (%)
C++ 9 0.38 1763 0.78 257 0.55 258 0.38
C/C++ Header 7 0.29 395 0.18 152 0.33 280 0.42
R 7 0.29 68 0.03 26 0.06 101 0.15
Rmd 1 0.04 23 0.01 31 0.07 33 0.05

Code of Conduct

Please note that this project is released with a Contributor Code of Conduct. By participating in this project you agree to abide by its terms.