spider 1.3.1

Multithreaded Web spider crawler written in Rust.
Documentation
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
extern crate spider;

use spider::website::Website;

fn main() {
  let mut website: Website = Website::new("https://rsseau.fr");
  website.configuration.blacklist_url.push("https://rsseau.fr/resume".to_string());
  website.configuration.respect_robots_txt = true;
  website.configuration.verbose = true; // Defaults to false
  website.configuration.delay = 2000; // Defaults to 250 ms
  website.configuration.concurrency = 10; // Defaults to 4
  website.configuration.user_agent = "myapp/version"; // Defaults to spider/x.y.z, where x.y.z is the library version
  website.crawl();
}