WebSites Parser
This website parser library allows asynchronous fetching and extracting data from web pages in multiple formats.
Key features include:
- Reading an HTML document from a given URL with a randomized user agent (User::random()).
- Selecting elements via CSS selectors and retrieving their attributes and contents.
- Fetching the entire page as plain text.
- Fetching and parsing page content as JSON, with integration for handling it via serde_json.
This tool is well-suited for web scraping and data extraction tasks, supporting flexible parsing of HTML, plain text, and JSON, thereby enabling comprehensive data retrieval from various web sources.
Examples:
use ;
async
Licensing:
Distributed under the MIT license.
Feedback:
You can contact me via GitHub or send a message to my Telegram @fuderis.
This library is constantly evolving, and I welcome your suggestions and feedback.