Struct robotparser::RobotFileParser [] [src]

pub struct RobotFileParser<'a> { /* fields omitted */ }

robots.txt file parser

Methods

impl<'a> RobotFileParser<'a>
[src]

Returns the time the robots.txt file was last fetched.

This is useful for long-running web spiders that need to check for new robots.txt files periodically.

Sets the time the robots.txt file was last fetched to the current time.

Sets the URL referring to a robots.txt file.

Reads the robots.txt URL and feeds it to the parser.

Reads the HTTP response and feeds it to the parser.

Parse the input lines from a robots.txt file

We allow that a user-agent: line is not preceded by one or more blank lines.

Using the parsed robots.txt decide if useragent can fetch url

Returns the crawl delay for this user agent as a Duration, or None if no crawl delay is defined.

Returns the sitemaps for this user agent as a Vec<Url>.

Returns the request rate for this user agent as a RequestRate, or None if not request rate is defined

Trait Implementations

impl<'a> Debug for RobotFileParser<'a>
[src]

Formats the value using the given formatter.

impl<'a> Eq for RobotFileParser<'a>
[src]

impl<'a> PartialEq for RobotFileParser<'a>
[src]

This method tests for self and other values to be equal, and is used by ==. Read more

This method tests for !=.

impl<'a> Clone for RobotFileParser<'a>
[src]

Returns a copy of the value. Read more

Performs copy-assignment from source. Read more