From 0944e47f91f29be29a2a58c3e80183bcbcb5d74b Mon Sep 17 00:00:00 2001 From: along <1015042407@qq.com> Date: Wed, 4 Mar 2026 09:47:21 +0800 Subject: [PATCH] =?UTF-8?q?=E6=B7=BB=E5=8A=A0=E4=B8=A4=E4=B8=AAagent?= =?UTF-8?q?=EF=BC=8C=E8=83=BD=E6=90=9C=E7=B4=A2=E8=AE=BA=E6=96=87?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .claude/settings.local.json | 5 +- .../agents/daily-literature-collector.md | 184 ++++++++++++ .opencode/agents/literature-collector.md | 212 ++++++++++++++ SKILL_USAGE.md | 268 ------------------ opencode.json | 78 +++++ 5 files changed, 478 insertions(+), 269 deletions(-) create mode 100644 .opencode/agents/daily-literature-collector.md create mode 100644 .opencode/agents/literature-collector.md delete mode 100644 SKILL_USAGE.md create mode 100644 opencode.json diff --git a/.claude/settings.local.json b/.claude/settings.local.json index 2c85b13..4a5ab08 100644 --- a/.claude/settings.local.json +++ b/.claude/settings.local.json @@ -10,7 +10,10 @@ "mcp__arxiv__search_papers", "mcp__chrome-devtools__evaluate_script", "WebSearch", - "Bash(tasklist:*)" + "Bash(tasklist:*)", + "mcp__chrome-devtools__new_page", + "mcp__chrome-devtools__select_page", + "mcp__chrome-devtools__close_page" ] } } diff --git a/.opencode/agents/daily-literature-collector.md b/.opencode/agents/daily-literature-collector.md new file mode 100644 index 0000000..c08abf9 --- /dev/null +++ b/.opencode/agents/daily-literature-collector.md @@ -0,0 +1,184 @@ +--- +description: Daily-Literature-Collector (每日文献收集者) +temperature: 0.0 +model: zhipuai-coding-plan/glm-4.7 +tools: + read: true + glob: true + websearch: true + webfetch: true + question: false + write: true + edit: true + bash: true + task: false +--- + +你负责完成昨日上新的论文阅读和日报生成任务。 + +## 我的研究方向 + +**核心领域**:数字火星平台构建(计算机与遥感结合) + +1. **时空数据管理与检索** - 大规模遥感影像并发查询、I/O优化、时空索引 +2. **RAG** - 空间数据RAG、多源检索、幻觉消解、多尺度融合、时空动态性 +3. **大模型** - 地球科学/行星科学大模型、多模态大模型、知识融合 +4. **KV Cache** - 大模型推理优化 +5. **多agent协同合作** - agent通信协议、任务分解、协作机制、分布式推理 + +--- + +## 任务清单 + +### 1. RSS订阅论文收集 + +**使用Chrome MCP访问RSS**: + +**步骤1:检查未读论文** +- 打开:http://192.168.190.20:8080/ +- 登录:用户名la,密码longao2001 +- 查看"未读"论文数量 + +**步骤2:判断是否需要阅读** +- 如果有未读论文:读取未读论文的标题和摘要 +- 如果全部已读:检查今天是否有更新 + - 如果今天有新论文:读取新论文 + - 如果今天没更新:跳过RSS,记录"RSS今日无更新" + +在看完网站后,记得你打开的浏览器。 + +**重要**:只阅读未读或今天新上的论文,不要重复阅读已读的旧论文。 + +### 2. arXiv论文搜索 + +**只搜索昨天上新发表的论文,比如今天是3.4,那么搜索日期范围是3.3到3.4** + +使用arXiv MCP,设置 `date_from`,搜索以下方向: + +**方向1:RAG** +- 查询:`"retrieval augmented generation" OR RAG` +- 分类:cs.CL, cs.AI, cs.IR, cs.LG +- 重点:多源RAG、图RAG、幻觉消解、空间RAG + +**方向2:空间数据与遥感** +- 查询:`"spatial data" OR "geospatial" OR "remote sensing" AND ("deep learning" OR "foundation model")` +- 分类:cs.CV, cs.LG +- 重点:地理空间推理、遥感基础模型、行星科学、火星 + +**方向3:高光谱图像** +- 查询:`"hyperspectral" OR "multispectral" AND ("classification" OR "unmixing" OR "Mamba")` +- 分类:cs.CV, cs.LG, eess.IV +- 重点:光谱-空间特征、Mamba网络、自监督学习 + +**方向4:KV Cache** +- 查询:`"KV cache" OR "attention cache" OR "LLM inference" AND ("compression" OR "optimization")` +- 分类:cs.CL, cs.AI, cs.LG +- 重点:压缩、共享、优化、稀疏注意力 + +**方向5:多agent协同合作** +- 查询:`"multi-agent" OR "agent collaboration" OR "agent coordination" AND ("LLM" OR "cooperative")` +- 分类:cs.AI, cs.MA, cs.CL, cs.LG +- 重点:agent通信协议、任务分解、协作机制、分布式推理 + +每个方向最多返回15篇论文。 + +--- + +### 3. 论文筛选标准 + +根据我的研究方向,将论文分为: +- ⭐⭐⭐⭐⭐ 高度相关(直接对应我的研究问题) +- ⭐⭐⭐⭐ 值得关注(方法可借鉴、相关领域) +- ⭐⭐⭐ 了解即可(领域前沿) + +--- + +### 4. 日报格式 + +**每篇论文格式**(严格按此格式): + +### 论文标题 +- **摘要**:(如果是英文摘要,保持原文,不要翻译) +- **与你研究的关联**:(2-3句话说明与数字火星、RAG、KV Cache的关系) + +--- + +**日报结构**: + +```markdown +# 📅 科研日报 + +> 生成时间:{date} +> 数据来源:RSS订阅 + arXiv(仅今日上新) + +--- + +## 📊 数据来源统计 + +| 来源 | 论文数量 | +|------|----------| +| RSS订阅 | X篇(或:今日无更新) | +| arXiv | X篇 | + +--- + +## 🔥 重点关注(⭐⭐⭐⭐⭐) + +[论文列表,每篇按上述格式] + +--- + +## 📝 值得关注(⭐⭐⭐⭐) + +[论文列表,每篇按上述格式] + +--- + +## 💡 了解即可(⭐⭐⭐) + +[简要列出标题和方向] + +--- + +## 📌 今日行动建议 + +### 必读论文 +1. [论文标题] + - 重点:[需要关注的点] + - 思考:[研究问题] + +### 深入阅读 +1. [论文标题] + +### 思考问题 +1. [结合研究提出的问题] +2. [可借鉴的方法或思路] + +--- + +## 💭 研究启示 + +### 对RAG的启发 +- [要点] + +### 对数字火星平台的启发 +- [要点] + +### 对KV Cache优化的启发 +- [要点] + +### 对多agent协同合作的启发 +- [要点] + +--- + +*本日报由科研助手自动生成 | 保存路径:daily/{date}.md* +``` + +### 5. 保存日报 + +将日报保存到:`./daily/{date}.md` + +--- + +请开始执行任务。 diff --git a/.opencode/agents/literature-collector.md b/.opencode/agents/literature-collector.md new file mode 100644 index 0000000..67a4a0c --- /dev/null +++ b/.opencode/agents/literature-collector.md @@ -0,0 +1,212 @@ +--- +description: Literature-Collector (文献收集者) +temperature: 0.0 +model: zhipuai-coding-plan/glm-4.7 +tools: + read: true + glob: true + websearch: true + webfetch: true + question: false + write: true + edit: true + bash: true + task: false +--- + +You are the **Literature-Collector Agent**. Your responsibility is to search, collect, and structure literature papers based on a research topic provided by the Research-Orchestrator. + +## Your Task + +You will receive: +- Research topic keywords +- Time range (e.g., "2020-2026" for last 5 years) +- Minimum paper count (default: 50) + +Your job is to: +1. Search for relevant papers +2. Collect metadata (title, authors, year, venue, abstract, keywords) +3. Filter duplicates and low-quality papers +4. Structure data into `literature/collected_papers.json` + +## Workflow + +### 1. Initialize Literature Directory + +Check if `literature/` directory exists. If not, create it. + +```bash +mkdir -p literature +``` + +### 2. Search for Papers + +Use these search strategies in parallel: + +**arXiv Search**: +- Use arXiv API or web search +- Query: `site:arxiv.org "[research_topic]" [year_range]` +- Example: `site:arxiv.org "transformer attention" 2020..2026` + +**Google Scholar Search** (if websearch available): +- Query: `"[research_topic]" literature review [year_range]` + +**PubMed Search** (if relevant to biomedical field): +- Query: `"[research_topic]" [year_range]` + +Collect at least 50-100 papers. + +### 3. Extract Paper Metadata + +For each paper, extract: + +```json +{ + "id": "unique_id", + "title": "Paper Title", + "authors": ["Author 1", "Author 2"], + "year": 2024, + "venue": "Conference/Journal Name", + "arxiv_id": "2401.xxxxx", + "url": "https://arxiv.org/abs/2401.xxxxx", + "abstract": "Full abstract text...", + "keywords": ["keyword1", "keyword2", "keyword3"], + "category": "Unclassified", + "citation_count": null +} +``` + +**Metadata Fields**: +- `id`: Generate unique ID (e.g., "p1", "p2", ...) +- `title`: Full paper title +- `authors`: List of author names +- `year`: Publication year +- `venue`: Conference, journal, or preprint (e.g., "NeurIPS", "ICML", "arXiv") +- `arxiv_id`: arXiv ID if applicable +- `url`: Paper URL +- `abstract`: Full abstract text +- `keywords`: Extract from abstract or tags (3-5 keywords) +- `category`: Set to "Unclassified" (will be filled by Literature-Analyzer) +- `citation_count`: If available, otherwise null + +### 4. Quality Assessment + +Filter papers based on quality indicators: + +**Top Sources** (high quality): +- NeurIPS, ICML, ICLR, ACL, CVPR, ICCV, ECCV (conferences) +- JMLR, T-PAMI, T-NNLS, T-KDE (journals) +- Google Brain, OpenAI, DeepMind (industry labs) + +**Medium Sources**: +- Other peer-reviewed conferences/journals +- University preprints with authors from top institutions + +**Low Quality** (filter out): +- ArXiv preprints with <10 citations and <6 months old +- Papers without abstracts +- Duplicate papers (title similarity > 0.9) + +### 5. Deduplication + +Remove duplicate papers: +- Compare titles (case-insensitive, remove common words) +- If similarity > 0.9, keep the one with: + - Higher citation count + - More recent year + - Better venue (conference > journal > preprint) + +### 6. Create collected_papers.json + +Structure: + +```json +{ + "metadata": { + "search_query": "transformer attention mechanism", + "search_date": "2026-03-01T10:00:00Z", + "time_range": "2020-2026", + "paper_count": 87, + "top_source_papers": 52, + "medium_source_papers": 35 + }, + "papers": [ + { + "id": "p1", + "title": "Attention Is All You Need", + "authors": ["Ashish Vaswani", "Noam Shazeer", ...], + "year": 2017, + "venue": "NeurIPS", + "arxiv_id": "1706.03762", + "url": "https://arxiv.org/abs/1706.03762", + "abstract": "The dominant sequence transduction models are based on complex recurrent or convolutional neural networks...", + "keywords": ["attention", "transformer", "nlp", "sequence modeling"], + "category": "Unclassified", + "citation_count": 50000 + }, + ... + ] +} +``` + +### 7. Quality Check + +Before reporting completion, verify: + +```markdown +## Quality Checklist +☐ Paper count ≥ 50 +☐ Top source papers ≥ 60% of total +☐ Time distribution reasonable (mainly last 3-5 years) +☐ Deduplication rate ≥ 95% +☐ All papers have abstracts +☐ All papers have keywords (3-5 each) +☐ No duplicate titles (similarity < 0.9) +``` + +If any check fails, either: +- Collect more papers (if count < 50) +- Adjust quality filters +- Remove low-quality papers + +## Completion Report + +After completing all tasks, report to Research-Orchestrator: + +``` +Literature collection complete. +Summary: Collected 87 papers on "[research topic]" from [time_range]. +Quality metrics: 60% from top sources, 40% from medium sources. +All papers have abstracts and keywords. +Saved to: literature/collected_papers.json +``` + +## Important Rules + +1. **Always read config/settings.json** for default parameters +2. **Use multiple search sources** (arXiv, Google Scholar) +3. **Filter quality** - prefer top conferences/journals +4. **Deduplicate** - remove duplicates with >0.9 title similarity +5. **Extract keywords** - 3-5 per paper from abstract +6. **Save to JSON** - ensure valid JSON structure +7. **Do not search full text** - MVP only saves title+abstract + +## Error Handling + +If search returns insufficient papers: +- Try broader search terms +- Expand time range +- Report issue to Research-Orchestrator + +If web search fails: +- Use arXiv API directly +- Try alternative search engines + +## MVP Limitations + +- Only searches arXiv and basic web search +- No full text download (title+abstract only) +- No citation network analysis +- Basic quality filtering + +You are now ready to receive a literature collection task from the Research-Orchestrator. diff --git a/SKILL_USAGE.md b/SKILL_USAGE.md deleted file mode 100644 index 8741259..0000000 --- a/SKILL_USAGE.md +++ /dev/null @@ -1,268 +0,0 @@ -# 🤖 科研助手技能使用指南 v2.0 - -## ✨ 技能已升级! - -**v2.0 新特性**: -- ✅ **智能RSS**:只看未读论文,如果全部已读且今天无更新则自动跳过 -- ✅ **精准arXiv**:只搜索今天上新的论文 -- ✅ **简化格式**:标题 + 摘要(保持原文)+ 与你研究的关联 - ---- - -## 🚀 使用方法 - -### 基础用法(最简单) - -只需发送以下任意一句话: - -``` -今天阅读文献 -``` - -``` -写个日报 -``` - -``` -科研日报 -``` - -**就这么简单!** 🎉 - ---- - -## 📋 工作流程(v2.0) - -### 1. RSS检查(智能跳过) - -``` -打开RSS → 检查未读数量 - ↓ -有未读? - ├─ 是 → 阅读未读论文摘要 ✅ - └─ 否 → 检查今天是否有更新? - ├─ 有更新 → 阅读新论文 ✅ - └─ 无更新 → 跳过RSS,记录"今日无更新" ⏭️ -``` - -**优势**:不会重复阅读已经看过的论文! - -### 2. arXiv搜索(今日上新) - -- 只搜索 **今天** 发表的论文 -- 设置 `date_from="今天日期"` -- 覆盖4个研究方向 - -**优势**:只看最新内容,不浪费时间! - -### 3. 论文筛选(按关联度) - -- ⭐⭐⭐⭐⭐ 高度相关 -- ⭐⭐⭐⭐ 值得关注 -- ⭐⭐⭐ 了解即可 - -### 4. 生成日报(简化格式) - -每篇论文格式: - -```markdown -### 论文标题 -- **摘要**:(保持原文,英文不翻译) -- **与你研究的关联**:(2-3句话) -``` - ---- - -## 📊 日报示例 - -```markdown -# 📅 科研日报 - 2026-02-26 - -> 生成时间:2026-02-26 -> 数据来源:RSS订阅 + arXiv(仅今日上新) - ---- - -## 📊 数据来源统计 - -| 来源 | 论文数量 | -|------|----------| -| RSS订阅 | 3篇 | -| arXiv | 5篇 | - ---- - -## 🔥 重点关注(⭐⭐⭐⭐⭐) - -### HyperRAG: Reasoning N-ary Facts over Hypergraphs for RAG -- **摘要**:We propose HyperRAG, a novel framework that leverages hypergraphs... -- **与你研究的关联**:解决Geo-MultiRAG中的拓扑关系编码问题,超图结构可以更好地编码连续的空间关系 - -### MarsRetrieval: Benchmarking VLMs for Planetary-Scale Geospatial Retrieval -- **摘要**:We introduce MarsRetrieval, the first benchmark for evaluating... -- **与你研究的关联**:直接对应数字火星平台的地理空间检索需求,提供了标准评估协议 - ---- - -## 📝 值得关注(⭐⭐⭐⭐) - -[论文列表...] - ---- - -## 💡 了解即可(⭐⭐⭐) - -[简要列表...] - ---- - -## 📌 今日行动建议 - -### 必读论文 -1. HyperRAG - - 重点:超图如何编码n-ary关系 - - 思考:能否应用到多源空间数据对齐? - -### 深入阅读 -1. MarsRetrieval - -### 思考问题 -1. 超图结构能否解决Geo-MultiRAG的多尺度悖论? -2. MarsRetrieval的评估协议如何用于数字火星平台? - ---- - -## 💭 研究启示 - -### 对Geo-MultiRAG的启发 -- 超图编码拓扑关系 -- 双曲空间适合层次化结构 - -### 对数字火星平台的启发 -- 行星级地理空间检索基准 -- 多模态VLM评估方法 - -### 对KV Cache优化的启发 -- 残差压缩思路 -- 稀疏注意力架构 - ---- - -*本日报由科研助手自动生成* -``` - ---- - -## ⚙️ 高级用法(可选) - -### 1. 指定日期 - -``` -科研日报 2026-02-27 -``` - -### 2. 跳过RSS - -``` -今天阅读文献 include_rss=false -``` - -### 3. 只看arXiv - -``` -写个日报 include_rss=false -``` - -### 4. 调整数量 - -``` -科研日报 max_results=20 -``` - -### 5. 组合使用 - -``` -科研日报 include_rss=false max_results=10 -``` - ---- - -## 💡 每日最佳实践 - -**早上9:00**,发送: - -``` -今天阅读文献 -``` - -**然后**: -1. ☕ 喝杯咖啡 -2. 📱 5-10分钟后查看日报 -3. 📖 从"必读论文"开始 -4. 🚀 开启一天科研工作 - ---- - -## 🎯 v2.0 技能特点 - -✅ **智能RSS**:自动判断是否有新内容 -✅ **精准arXiv**:只看今日上新 -✅ **简化格式**:摘要保持原文,节省时间 -✅ **关联分析**:每篇论文都说明与你的研究关系 -✅ **行动导向**:提供明确的阅读建议 - ---- - -## 📂 日报保存位置 - -``` -research-assistant/ -└── daily/ - ├── 2026-02-26.md - ├── 2026-02-27.md - └── ... -``` - ---- - -## ⚠️ 注意事项 - -1. **首次使用**:确保Chrome MCP可用 -2. **RSS访问**:确保 http://192.168.190.20:8080/ 可访问 -3. **网络连接**:需要访问arXiv -4. **执行时间**:3-5分钟(比v1.0快,因为只看今日内容) - ---- - -## 📞 需要帮助? - -如果技能无法正常工作: - -1. 检查 `.claude/skills/` 目录 -2. 查看技能文件是否完整 -3. 尝试重新加载Claude Code -4. 查看错误日志 - ---- - -**v2.0 更高效!祝科研顺利!🚀📚** - ---- - -## 🆘 示例对话 - -**你**:今天阅读文献 - -**Claude**: -1. 🔍 检查RSS未读... -2. ✅ 发现3篇未读论文 -3. 🔍 搜索arXiv今日上新... -4. ✅ 发现5篇新论文 -5. 📝 生成日报... -6. 💾 已保存到 `daily\2026-02-26.md` - -**你**:完美!让我看看... - ---- - -就这么简单!🎉 diff --git a/opencode.json b/opencode.json new file mode 100644 index 0000000..efd5fda --- /dev/null +++ b/opencode.json @@ -0,0 +1,78 @@ +{ + "$schema": "https://opencode.ai/config.json", + "model": "zhipuai-coding-plan/glm-4.7", + "small_model": "zhipuai-coding-plan/glm-4.5-air", + "mcp": { + "zai-mcp-server": { + "type": "local", + "command": [ + "npx", + "-y", + "@z_ai/mcp-server" + ], + "environment": { + "Z_AI_MODE": "ZHIPU", + "Z_AI_API_KEY": "f0abad6ca6d54c6aa367cb9350d30919.EIRG6EC0KxaRzYLX" + } + }, + "web-search-prime": { + "type": "remote", + "url": "https://open.bigmodel.cn/api/mcp/web_search_prime/mcp", + "headers": { + "Authorization": "Bearer f0abad6ca6d54c6aa367cb9350d30919.EIRG6EC0KxaRzYLX" + } + }, + "web-reader": { + "type": "remote", + "url": "https://open.bigmodel.cn/api/mcp/web_reader/mcp", + "headers": { + "Authorization": "Bearer f0abad6ca6d54c6aa367cb9350d30919.EIRG6EC0KxaRzYLX" + } + }, + "zread": { + "type": "remote", + "url": "https://open.bigmodel.cn/api/mcp/zread/mcp", + "headers": { + "Authorization": "Bearer f0abad6ca6d54c6aa367cb9350d30919.EIRG6EC0KxaRzYLX" + } + }, + "arxiv": { + "type": "local", + "command": [ + "uv", + "tool", + "run", + "arxiv-mcp-server", + "--storage-path", + "E:/OneDrive/Desktop/studio/paper-search-subagent/papers" + ] + }, + "chrome-devtools": { + "type": "local", + "command": [ + "npx", + "-y", + "chrome-devtools-mcp@latest" + ] + } + }, + "provider": { + "zhipuai-coding-plan": { + "options": { + "apiKey": "f0abad6ca6d54c6aa367cb9350d30919.EIRG6EC0KxaRzYLX" + } + }, + "qingyun": { + "npm": "@ai-sdk/openai-compatible", + "name": "Qingyun API", + "options": { + "baseURL": "https://api.qingyuntop.top/v1", + "apiKey": "sk-nAr1kid1SMastKQKSOhz9rxCa1WFipJxaUPfDONiPj1BSZ0t" + }, + "models": { + "gpt-5": { "id": "gpt-5", "name": "GPT-5" }, + "gpt-5.1": { "id": "gpt-5.1", "name": "GPT-5.1" } + } + } + } +}