57 lines
1.3 KiB
Rust
57 lines
1.3 KiB
Rust
use scraper::{Html,Selector};
|
|
use url::{Url,Host};
|
|
use futures::future;
|
|
|
|
type SiteStat = (Url, Vec<Url>);
|
|
|
|
#[tokio::main]
|
|
async fn main() -> Result<(), Box<dyn std::error::Error>> {
|
|
let addr = std::env::args().nth(1)
|
|
.unwrap_or_else(|| "www.tmplab.org".to_string());
|
|
|
|
let links = get_links("https://tmplab.org").await?;
|
|
let links : Vec<Url> = links.into_iter()
|
|
.filter(| url | url.host() != Some(Host::Domain(&addr)))
|
|
.collect();
|
|
|
|
let joins = links.into_iter()
|
|
.map(| url | {
|
|
tokio::spawn(async move { fetch_site(url).await })
|
|
});
|
|
|
|
let results = future::join_all(joins).await
|
|
.into_iter()
|
|
.filter_map(| r | r.ok());
|
|
|
|
for (url, links) in results {
|
|
println!("{url} : {} links", links.len())
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
async fn fetch_site(url: Url) -> SiteStat {
|
|
println!("Start fetching {url}");
|
|
let links = get_links(url.as_ref()).await
|
|
.unwrap_or_else(|_| vec![]);
|
|
println!("Got {url} => {}", links.len());
|
|
(url, links)
|
|
}
|
|
|
|
async fn get_links(
|
|
url: &str
|
|
) -> Result<Vec<Url>, reqwest::Error> {
|
|
let a_selector = Selector::parse("a[href]").unwrap();
|
|
|
|
let body = reqwest::get(url)
|
|
.await?
|
|
.text()
|
|
.await?;
|
|
|
|
Ok(Html::parse_document(&body)
|
|
.select(&a_selector)
|
|
.filter_map(| link | link.value().attr("href")
|
|
.and_then(| href | Url::parse(href).ok()))
|
|
.collect())
|
|
}
|