当前位置: 首页 > news >正文

基于Rust实现爬取 GitHub Trending 热门仓库

基于Rust实现爬取 GitHub Trending 热门仓库

这个实战项目将使用 Rust 实现一个爬虫,目标是爬取 GitHub Trending 页面的热门 Rust 仓库信息(仓库名、描述、星标数、作者等),并将结果输出为 JSON 文件。本次更新基于优化后的代码,重点提升了错误处理容错性和 CSS 选择器稳定性。

技术栈

  • HTTP 请求reqwest( Rust 最流行的 HTTP 客户端,支持异步)
  • HTML 解析scraper(基于 selectors 库,支持 CSS 选择器,轻量高效)
  • JSON 序列化serde + serde_json( Rust 标准的序列化 / 反序列化库)
  • 异步运行时tokio( Rust 异步编程的事实标准)
  • 日志env_logger + log(简单的日志输出,方便调试)
  • 错误处理anyhow(简化错误传递,无需手动定义复杂错误类型)

项目结构

github-trending-crawler/
├── Cargo.toml       # 依赖配置
├── src/
│   └── main.rs      # 核心逻辑
└── trending_repos.json  # 输出结果文件(运行后生成)

步骤 1:创建项目并配置依赖

1.1 创建项目

cargo new github-trending-crawler
cd github-trending-crawler

1.2 配置 Cargo.toml

添加依赖项(确保版本兼容,可通过 crates.io 查询最新版本):

[package]
name = "github-trending-crawler"
version = "0.1.0"
edition = "2021"
description = "A crawler to fetch GitHub Trending Rust repositories"
license = "MIT"[dependencies]
# HTTP 客户端(异步)
reqwest = { version = "0.12", features = ["json", "rustls-tls"] }
# HTML 解析(CSS 选择器)
scraper = "0.18"
# JSON 序列化/反序列化
serde = { version = "1.0", features = ["derive"] }
serde_json = "1.0"
# 异步运行时
tokio = { version = "1.0", features = ["full"] }
# 日志
log = "0.4"
env_logger = "0.10"
# 错误处理(可选,简化错误传递)
anyhow = "1.0"

步骤 2:核心逻辑实现(src/main.rs )

2.1 导入依赖和初始化日志

use anyhow::{Context, Result};
use log::info;
use reqwest::Client;
use scraper::{Html, Selector};
use serde::Serialize;
use std::fs::File;
use std::path::Path;// 初始化日志(运行时打印调试信息)
fn init_logger() {env_logger::Builder::from_env(env_logger::Env::default().default_filter_or("info")).init();
}

2.2 定义数据结构(序列化用)

定义存储仓库信息的结构体,使用 serde::Serialize trait 支持 JSON 序列化,字段与 GitHub Trending 页面信息一一对应:

#[derive(Debug, Serialize)]
struct GithubRepo {// 作者/组织名author: String,// 仓库名name: String,// 仓库描述description: Option<String>,// 星标数stars: String,// 分支数forks: String,// 今日新增星标today_stars: String,// 仓库链接url: String,
}

2.3 核心爬虫逻辑

2.3.1 构建 HTTP 客户端并请求页面

优化点:保持原有 User-Agent 伪装和超时设置,确保请求不被 GitHub 拒绝,同时保留详细的请求错误上下文:

async fn fetch_trending_page(client: &Client) -> Result<String> {// GitHub Trending Rust 页面 URL(按今日热门排序)let url = "https://github.com/trending/rust?since=daily";info!("Fetching page: {}", url);// 发送 GET 请求(设置 User-Agent 模拟浏览器,避免被 GitHub 拦截)let response = client.get(url).header("User-Agent", "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/120.0.0.0 Safari/537.36").send().await.context(format!("Failed to request URL: {}", url))?;// 检查响应状态码(200-299 为成功状态)if !response.status().is_success() {return Err(anyhow::anyhow!("Request failed with status: {}",response.status()));}// 读取响应体(HTML 字符串),并记录页面大小let html = response.text().await.context("Failed to read response body")?;info!("Successfully fetched page (size: {} bytes)", html.len());Ok(html)
}
2.3.2 解析 HTML 提取仓库信息

优化点:

  1. CSS 选择器错误处理:用 map_err,直接显示选择器解析错误详情,便于调试;
  1. 选择器稳定性提升:将描述选择器改为通用 p 标签、星标 / 分支选择器改为基于 href 后缀(如 a[href$='/stargazers']),避免因 GitHub 样式类名变更导致解析失败;
  1. 缺失值容错:用 unwrap_or_else 给缺失的星标 / 分支 / 今日新增星标设置默认值 "0",避免程序 panic;
  1. 代码简化:合并作者名元素提取逻辑,减少重复代码。
fn parse_repos(html: &str) -> Result<Vec<GithubRepo>> {info!("Starting to parse repositories...");let document = Html::parse_document(html);// 定义 CSS 选择器(优化后:基于语义化属性,降低页面样式变更影响)// 1. 每个仓库的根节点选择器(GitHub 仓库列表统一用 article.Box-row 包裹)let repo_selector = Selector::parse("article.Box-row").map_err(|e| anyhow::anyhow!("Failed to parse repo selector: {}", e))?;// 2. 作者 + 仓库名选择器(h2 下的 a 标签,包含仓库路径)let author_name_selector = Selector::parse("h2 a").map_err(|e| anyhow::anyhow!("Failed to parse author-name selector: {}", e))?;// 3. 仓库描述选择器(通用 p 标签,避免依赖特定类名)let desc_selector = Selector::parse("p").map_err(|e| anyhow::anyhow!("Failed to parse description selector: {}", e))?;// 4. 星标数选择器(基于 href 后缀 /stargazers,语义化更强)let stars_selector = Selector::parse("a[href$='/stargazers']").map_err(|e| anyhow::anyhow!("Failed to parse stars selector: {}", e))?;// 5. 分支数选择器(基于 href 后缀 /forks,语义化更强)let forks_selector = Selector::parse("a[href$='/forks']").map_err(|e| anyhow::anyhow!("Failed to parse forks selector: {}", e))?;// 6. 今日新增星标选择器(基于 data-menu-button-text 属性,稳定性更高)let today_stars_selector = Selector::parse("span[data-menu-button-text]").map_err(|e| anyhow::anyhow!("Failed to parse today-stars selector: {}", e))?;let mut repos = Vec::new();// 遍历每个仓库节点,提取信息for repo_node in document.select(&repo_selector) {// 1. 提取作者和仓库名(格式:"author/name",从 a 标签文本中解析)let author_name_element = repo_node.select(&author_name_selector).next().context("Missing author/name element (GitHub page structure may have changed)")?;// 清理文本(去除空格和换行符)let author_name_text = author_name_element.text().collect::<String>().trim().to_string();// 分割作者和仓库名(格式必须为 "author/name",否则报错)let (author, name) = author_name_text.split_once('/').context(format!("Invalid author/name format: '{}' (expected 'author/name')", author_name_text))?;let author = author.trim().to_string();let name = name.trim().to_string();// 2. 提取仓库完整链接(补全 GitHub 域名,a 标签的 href 属性为相对路径)let url = author_name_element.value().attr("href").context("Missing href attribute for repo link")?.to_string();let url = format!("https://github.com{}", url); // 拼接完整 URL// 3. 提取仓库描述(可选,无描述时为 None,避免强制 unwrap 导致 panic)let description = repo_node.select(&desc_selector).next().map(|elem| elem.text().collect::<String>().trim().to_string());// 4. 提取星标数(缺失时默认值为 "0",容错性优化)let stars = repo_node.select(&stars_selector).next().map(|elem| elem.text().collect::<String>().trim().to_string()).unwrap_or_else(|| "0".to_string());// 5. 提取分支数(缺失时默认值为 "0",容错性优化)let forks = repo_node.select(&forks_selector).next().map(|elem| elem.text().collect::<String>().trim().to_string()).unwrap_or_else(|| "0".to_string());// 6. 提取今日新增星标(缺失时默认值为 "0",容错性优化)let today_stars = repo_node.select(&today_stars_selector).next().map(|elem| elem.text().collect::<String>().trim().to_string()).unwrap_or_else(|| "0".to_string());// 构建仓库对象并添加到列表repos.push(GithubRepo {author,name,description,stars,forks,today_stars,url,});}info!("Successfully parsed {} repositories", repos.len());Ok(repos)
}
2.3.3 保存结果到 JSON 文件

将解析后的仓库列表序列化为格式化的 JSON(pretty 模式),便于阅读和后续处理:

fn save_repos_to_json(repos: &[GithubRepo], path: &str) -> Result<()> {info!("Saving repositories to JSON file: {}", path);// 创建文件(若已存在会覆盖)let file = File::create(Path::new(path)).context(format!("Failed to create file: {} (check directory permissions)", path))?;// 序列化并写入文件(pretty 模式:缩进格式化,可读性强)serde_json::to_writer_pretty(file, repos).context("Failed to serialize repos to JSON (invalid data format)")?;info!("Successfully saved {} repos to {}", repos.len(), path);Ok(())
}

2.4 主函数

使用 tokio 异步运行时,按 “请求页面 → 解析信息 → 保存结果” 的流程执行,同时保留详细日志:

#[tokio::main]
async fn main() -> Result<()> {// 初始化日志(程序启动时执行)init_logger();info!("Starting GitHub Trending Rust Crawler...");// 创建 HTTP 客户端(设置超时,避免网络问题导致程序卡死)let client = Client::builder().connect_timeout(std::time::Duration::from_secs(10)) // 连接超时:10 秒.timeout(std::time::Duration::from_secs(15)) // 响应超时:15 秒.build().context("Failed to create HTTP client (check network or dependencies)")?;// 1. 爬取 GitHub Trending 页面 HTMLlet html = fetch_trending_page(&client).await?;// 2. 解析 HTML,提取仓库信息let repos = parse_repos(&html)?;// 3. 将结果保存到 JSON 文件(项目根目录下的 trending_repos.json)save_repos_to_json(&repos, "trending_repos.json")?;info!("Crawler finished successfully! Check 'trending_repos.json' for results.");Ok(())
}

步骤 3:运行项目并验证结果

3.1 运行爬虫

# 直接运行(默认输出 info 级别日志)
cargo run# (可选)输出 debug 级别日志(查看更详细的执行过程,便于调试)
RUST_LOG=debug cargo run

3.2 验证结果

运行成功后,项目根目录会生成 trending_repos.json 文件,优化后的结果示例(容错性提升,无缺失值):

[{"author": "YaLTeR","name": "niri","description": "A scrollable-tiling Wayland compositor.","stars": "14,823","forks": "523","today_stars": "0","url": "https://github.com/YaLTeR/niri"},{"author": "librespot-org","name": "librespot","description": "Open Source Spotify client library","stars": "6,131","forks": "773","today_stars": "0","url": "https://github.com/librespot-org/librespot"},{"author": "zensical","name": "zensical","description": "A modern static site generator by the creators of Material for MkDocs","stars": "859","forks": "12","today_stars": "0","url": "https://github.com/zensical/zensical"},{"author": "atuinsh","name": "atuin","description": "✨ Magical shell history","stars": "26,951","forks": "730","today_stars": "0","url": "https://github.com/atuinsh/atuin"},{"author": "openai","name": "codex","description": "Lightweight coding agent that runs in your terminal","stars": "50,236","forks": "6,238","today_stars": "0","url": "https://github.com/openai/codex"},{"author": "bevyengine","name": "bevy","description": "A refreshingly simple data-driven game engine built in Rust","stars": "43,023","forks": "4,224","today_stars": "0","url": "https://github.com/bevyengine/bevy"},{"author": "topjohnwu","name": "Magisk","description": "The Magic Mask for Android","stars": "56,884","forks": "15,868","today_stars": "0","url": "https://github.com/topjohnwu/Magisk"},{"author": "uutils","name": "coreutils","description": "Cross-platform Rust rewrite of the GNU coreutils","stars": "22,144","forks": "1,637","today_stars": "0","url": "https://github.com/uutils/coreutils"},{"author": "regolith-labs","name": "ore","description": "It's time to mine.","stars": "741","forks": "262","today_stars": "0","url": "https://github.com/regolith-labs/ore"},{"author": "commonwarexyz","name": "monorepo","description": "Commonware Library Primitives and Examples","stars": "369","forks": "131","today_stars": "0","url": "https://github.com/commonwarexyz/monorepo"},{"author": "rust-lang","name": "rust-analyzer","description": "A Rust compiler front-end for IDEs","stars": "15,670","forks": "1,864","today_stars": "0","url": "https://github.com/rust-lang/rust-analyzer"},{"author": "chroma-core","name": "chroma","description": "Open-source search and retrieval database for AI applications.","stars": "24,344","forks": "1,912","today_stars": "0","url": "https://github.com/chroma-core/chroma"},{"author": "getzola","name": "zola","description": "A fast static site generator in a single binary with everything built-in. https://www.getzola.org","stars": "16,120","forks": "1,094","today_stars": "0","url": "https://github.com/getzola/zola"},{"author": "longbridge","name": "gpui-component","description": "Rust GUI components for building fantastic cross-platform desktop application by using GPUI.","stars": "7,532","forks": "297","today_stars": "0","url": "https://github.com/longbridge/gpui-component"},{"author": "fish-shell","name": "fish-shell","description": "The user-friendly command line shell.","stars": "31,470","forks": "2,159","today_stars": "0","url": "https://github.com/fish-shell/fish-shell"},{"author": "Spotifyd","name": "spotifyd","description": "A spotify daemon","stars": "10,404","forks": "487","today_stars": "0","url": "https://github.com/Spotifyd/spotifyd"},{"author": "tree-sitter","name": "tree-sitter","description": "An incremental parsing system for programming tools","stars": "22,666","forks": "2,198","today_stars": "0","url": "https://github.com/tree-sitter/tree-sitter"},{"author": "UwUDev","name": "ygege","description": "High-performance indexer for YGG Torrent written in Rust","stars": "78","forks": "9","today_stars": "0","url": "https://github.com/UwUDev/ygege"},{"author": "ankitects","name": "anki","description": "Anki is a smart spaced repetition flashcard program","stars": "24,634","forks": "2,583","today_stars": "0","url": "https://github.com/ankitects/anki"},{"author": "get-convex","name": "convex-backend","description": "The open-source reactive database for app developers","stars": "8,066","forks": "452","today_stars": "0","url": "https://github.com/get-convex/convex-backend"}
]

项目总结

本项目是基于 Rust 开发的 GitHub Trending 热门 Rust 仓库爬虫,通过 reqwest 实现异步 HTTP 请求、scraper 解析 HTML 页面、serde 系列库完成 JSON 序列化,搭配 tokio 异步运行时和 anyhow 错误处理库,构建了高效且健壮的爬取流程。相较于初始版本,优化后的代码在 CSS 选择器上采用语义化属性(如基于 href 后缀、数据属性),降低了 GitHub 页面样式变更带来的维护成本;在错误处理上,通过 map_err 明确选择器解析错误、unwrap_or_else 处理信息缺失场景,大幅提升了程序容错性;同时保留详细日志输出,便于调试和问题定位。项目最终能稳定爬取每日热门 Rust 仓库的作者、名称、描述、星标数等关键信息,并以格式化 JSON 文件存储结果。

想了解更多关于Rust语言的知识及应用,可前往华为开放原子旋武开源社区https://xuanwu.openatom.cn/,了解更多资讯~

http://www.dtcms.com/a/604851.html

相关文章:

  • 深圳市建设局官方网站曼联对利物浦新闻
  • 【Android 组件】实现数据对象的 Parcelable 序列化
  • CrowdDiff: 使用扩散模型进行多假设人群密度估计
  • 同创企业网站源码wordpress自定义简单注册
  • 在 Android ARM64 上运行 x86_64 程序
  • 幽冥大陆(二十)屏幕录像特效增加节目效果——东方仙盟炼气期
  • 类加载机制、生命周期、类加载器层次、JVM的类加载方式
  • 数据智能开发五 技术架构
  • 免费的app软件下载网站个人网站备案 法律说明
  • MFC Check Box控件完全指南:属性设置、样式定制与高级应用
  • 广州 网站 建设支付网站建设费入什么科目
  • 西宁做网站需要多少钱wordpress怎么安装模板
  • 网站标题优化工具外贸公司电话
  • 北京北排建设公司招标网站wordpress登陆过程
  • 怎么免费做个人网站建设银行网站怎么打印明细
  • 在线设计图片网站总结郑州app拉新项目
  • 网站建设春节放假番禺免费核酸检测
  • preec网站电子商务seo实训总结
  • 优化网站浏览量怎么看网站建设方案评标原则
  • 网站设计职业工作室打开网站搜索
  • 网站建设方案书 doc织梦修改网站背景颜色
  • 网上商城建站服务商wordpress 抄袭查询
  • 网站建设设计猫和老鼠wordpress加备案号
  • 做电影网站 需要进那些群wordpress头像不同步
  • flash网站制作教程网站开发和网站运营的区别
  • 什么做直播网站sqlite开发网站
  • 沧州兼职网站建设集美区建设局网站
  • 如何把自己做的网站挂网上取公司名称大全简单大气
  • 大理市城乡建设局网站网站设计排行榜前十
  • 泉州网站建设公司首选公司哪家好东营免费网站制作