sitescraper-0.1.52 has been yanked.
Scraping Websites in Rust!
Examples:
Get InnerHTML:
let html = "Hello World!";
let dom = sitescraper::parse_html(html).unwrap();
let filtered_dom = sitescraper::filter!(dom, "body");
println!("{}", filtered_dom.get_inner_html());
//Output: Hello World!
Get Text:
let html = "Hello World!";
let dom = sitescraper::parse_html(html).unwrap();
let filtered_dom = sitescraper::filter!(dom, "body");
println!("{}", filtered_dom.get_text());
//Output: Hello World!
Get Text from single Tags:
use sitescraper;
let html = "Hello World!";
let dom = sitescraper::parse_html(html).unwrap();
let filtered_dom = sitescraper::filter!(dom, "div");
println!("{}", filtered_dom.tag[0].get_text());
//Output: Hello World!
Works also with
get_inner_html()
You can also leave arguments out by passing "*" or "":
use sitescraper;
let html = "Hello World!";
let dom = sitescraper::parse_html(html).unwrap();
let filtered_dom = sitescraper::filter!(dom, "*", "id", "hello");
println!("{}", filtered_dom.tag[0].get_text());
//Output: Hello World!
or
use sitescraper;
let html = "Hello World!";
let dom = sitescraper::parse_html(html).unwrap();
let filtered_dom = sitescraper::filter!(dom, "", "", "hello");
println!("{}", filtered_dom.tag[0].get_text());
//Output: Hello World!
Get Website-Content:
use sitescraper;
let html = sitescraper::http::get("http://example.com/).await.unwrap();
let dom = sitescraper::parse_html(html).unwrap();
let filtered_dom = sitescraper::filter!(dom, "div");
println!("{}", filtered_dom.get_inner_html());