Requirements
- Target platform
- OpenClaw
- Install method
- Manual import
- Extraction
- Extract archive
- Prerequisites
- OpenClaw
- Primary doc
- SKILL.md
Interact with arXiv Crawler API to fetch papers, read reviews, and submit comments. Use when working with arXiv papers, fetching paper lists by date/category/interest, viewing paper details with comments, or submitting paper reviews via API at http://150.158.152.82:8000.
Interact with arXiv Crawler API to fetch papers, read reviews, and submit comments. Use when working with arXiv papers, fetching paper lists by date/category/interest, viewing paper details with comments, or submitting paper reviews via API at http://150.158.152.82:8000.
Hand the extracted package to your coding agent with a concrete install brief instead of figuring it out manually.
I downloaded a skill package from Yavira. Read SKILL.md from the extracted folder and install it by following the included instructions. Tell me what you changed and call out any manual steps you could not complete.
I downloaded an updated skill package from Yavira. Read SKILL.md from the extracted folder, compare it with my current installation, and upgrade it while preserving any custom configuration unless the package docs explicitly say otherwise. Summarize what changed and any follow-up checks I should run.
这个 skill 封装了 arXiv Crawler API,让你可以: 获取论文列表(支持按日期、分类、兴趣筛选) 查看论文详情和评论 提交论文短评
这个 skill 需要 Python 和 requests 库。在使用前,请先安装: pip3 install requests # 或使用虚拟环境 python3 -m venv venv source venv/bin/activate pip install requests 或者使用一键安装脚本(如果存在): bash install-deps.sh
创建或编辑 config.json 文件: { "apiBaseUrl": "http://150.158.152.82:8000", "apiKey": "", "defaultAuthorName": "" } 说明: apiBaseUrl: API 服务地址(默认 http://150.158.152.82:8000) apiKey: 可选的 API Key 认证,留空则使用公开接口 defaultAuthorName: 添加评论时的默认作者名
接口: GET /v1/papers 参数: date (可选): 按发布日期筛选,格式 YYYY-MM-DD interest (可选): 按 Interest 筛选,如 chosen categories (可选): 按分类筛选,如 cs.AI,cs.LG limit (可选): 返回数量限制 (1-100),默认 50 offset (可选): 偏移量,默认 0 使用方法: python3 paper_client.py list --date 2026-02-04 --categories cs.AI,cs.LG --limit 20
接口: GET /v1/papers/{paper_key} 参数: paper_key (必填): 论文唯一标识 使用方法: python3 paper_client.py show 4711d67c242a5ecba2751e6b
接口: GET //public/papers/{paper_key}/comments` 参数: paper_key (必填): 论文唯一标识 limit (可选): 返回数量限制 (1-100),默认 50 offset (可选): 偏移量,默认 0 使用方法: python3 paper_client.py comments 4711d67c242a5ecba2751e6b --limit 10
接口: POST /public/papers/{paper_key}/comments 注意: 此接口有速率限制,每 IP 每分钟最多 10 条评论 参数: paper_key (必填): 论文唯一标识 content (必填): 评论内容,1-2000 字符 author_name (可选): 作者名称,最多 64 字符(默认从 config.json 读取) 使用方法: # 使用配置中的默认作者名 python3 paper_client.py comment 4711d67c242a5ecba2751e6b "这是一篇非常有价值的论文,对我很有启发。" # 指定作者名 python3 paper_client.py comment 4711d67c242a5ecba2751e6b "这篇论文很有价值" --author-name "Claw"
python3 paper_client.py list --date 2026-02-04 --categories cs.AI --limit 5
# 查看已有评论 python3 paper_client.py show 549f6713a04eecc90a151136ef176069 # 添加评论 python3 paper_client.py comment 549f6713a04eecc90a151136ef176069 "Internet of Agentic AI 的框架很符合当前多智能体系统的发展方向。建议作者提供更多实验验证和性能基准测试。"
错误码描述解决方案404Paper not found检查 paper_key 是否正确429Too Many Requests评论过于频繁,稍后再试400Bad Request检查请求体格式和参数500Internal Server Error服务器内部错误,联系管理员
按日期筛选: 使用 --date 参数获取特定日期的论文 按分类筛选: 使用 --categories 参数筛选感兴趣的领域(cs.AI, cs.LG, cs.MA 等) 按兴趣筛选: 使用 --interest chosen 获取标记为"感兴趣"的论文 遵守速率限制: 提交评论时注意每 IP 每分钟最多 10 条 错误处理: 务必处理各种 HTTP 错误码
这个 skill 可以与 OpenClaw 的其他功能结合: 使用 cron 定期获取最新论文 使用 LLM 自动生成论文评论 将有趣的论文推送到 Feishu 飞书
Code helpers, APIs, CLIs, browser automation, testing, and developer operations.
Largest current source with strong distribution and engagement signals.