Skip to content

v1.98.3

Compare
Choose a tag to compare
@j-mendez j-mendez released this 19 Jun 12:24
· 290 commits to main since this release

Whats Changed

  • Fix sitemap regex compile.
  • Whitelisting routes to only get the paths you want can now be done with website.with_whitelist_url example:
  • Fix robot disallow respect
use spider::{tokio, website::Website};
use tokio::io::AsyncWriteExt;

#[tokio::main]
async fn main() {
    let mut website: Website = Website::new("https://rsseau.fr/en/");

    website.with_whitelist_url(Some(vec!["/books".into()]));

    let mut rx2: tokio::sync::broadcast::Receiver<spider::page::Page> =
        website.subscribe(0).unwrap();
    let mut stdout = tokio::io::stdout();

    let join_handle = tokio::spawn(async move {
        while let Ok(res) = rx2.recv().await {
            let _ = stdout
                .write_all(format!("- {}\n", res.get_url()).as_bytes())
                .await;
        }
        stdout
    });

    let start = std::time::Instant::now();
    website.crawl().await;
    website.unsubscribe();
    let duration = start.elapsed();
    let mut stdout = join_handle.await.unwrap();

    let _ = stdout
        .write_all(
            format!(
                "Time elapsed in website.crawl() is: {:?} for total pages: {:?}",
                duration,
                website.get_links().len()
            )
            .as_bytes(),
        )
        .await;
}

Full Changelog: v1.97.14...v1.98.3