返回顶部
d

deep-scout

Multi-stage deep intelligence pipeline (Search → Filter → Fetch → Synthesize). Turns a query into a structured research report with full source citations.

作者: admin | 来源: ClawHub
源自
ClawHub
版本
V 0.1.4
安全检测
已通过
389
下载量
0
收藏
概述
安装方式
版本历史

deep-scout

# deep-scout Multi-stage deep intelligence pipeline (Search → Filter → Fetch → Synthesize). ## 🛠️ Installation ### 1. Ask OpenClaw (Recommended) Tell OpenClaw: *"Install the deep-scout skill."* The agent will handle the installation and configuration automatically. ### 2. Manual Installation (CLI) If you prefer the terminal, run: ```bash clawhub install deep-scout ``` ## 🚀 Usage ``` /deep-scout "Your research question" [--depth 5] [--freshness pw] [--country US] [--style report] ``` ### Options | Flag | Default | Description | |------|---------|-------------| | `--depth N` | 5 | Number of URLs to fully fetch (1–10) | | `--freshness` | `pw` | `pd`=past day, `pw`=past week, `pm`=past month, `py`=past year | | `--country` | `US` | 2-letter country code for Brave search | | `--language` | `en` | 2-letter language code | | `--search-count` | 8 | Total results to collect before filtering | | `--min-score` | 4 | Minimum relevance score to keep (0–10) | | `--style` | `report` | `report` \| `comparison` \| `bullets` \| `timeline` | | `--dimensions` | `auto` | Comparison dimensions (comma-separated, for `--style comparison`) | | `--output FILE` | stdout | Write report to file | | `--no-browser` | — | Disable browser fallback | | `--no-firecrawl` | — | Disable Firecrawl fallback | --- ## 🛠️ Pipeline — Agent Loop Instructions When this skill is invoked, execute the following four-stage pipeline: --- ### Stage 1: SEARCH Call `web_search` with: ``` query: <user query> count: <search_count> country: <country> search_lang: <language> freshness: <freshness> ``` Collect: title, url, snippet for each result. If fewer than 3 results returned, retry with `freshness: "py"` (relaxed). --- ### Stage 2: FILTER Load `prompts/filter.txt`. Replace template vars: - `{{query}}` → the user's query - `{{freshness}}` → freshness param - `{{min_score}}` → min_score param - `{{results_json}}` → JSON array of search results Call the LLM with this prompt. Parse the returned JSON array. Keep only results where `keep: true`. Sort by score descending. Take top `depth` URLs as the fetch list. **Deduplication:** Max 2 results per root domain (already handled in filter prompt). --- ### Stage 3: FETCH (Tiered Escalation) For each URL in the filtered list: **Tier 1 — web_fetch (fast):** ``` Call web_fetch(url) If content length >= 200 chars → accept, trim to max_chars_per_source ``` **Tier 2 — Firecrawl (deep/JS):** ``` If Tier 1 fails or returns < 200 chars: Run: scripts/firecrawl-wrap.sh <url> <max_chars> If output != "FIRECRAWL_UNAVAILABLE" and != "FIRECRAWL_EMPTY" → accept ``` **Tier 3 — Browser (last resort):** ``` If Tier 2 fails: Call browser(action="open", url=url) Call browser(action="snapshot") Load prompts/browser-extract.txt, substitute {{query}} and {{max_chars_per_source}} Call LLM with snapshot content + extraction prompt If output != "FETCH_FAILED:..." → accept ``` **If all tiers fail:** Use the original snippet from Stage 1 search results. Mark as `[snippet only]`. Store: `{ url: extracted_content }` dict. --- ### Stage 4: SYNTHESIZE Choose prompt template based on `--style`: - `report` / `bullets` / `timeline` → `prompts/synthesize-report.txt` - `comparison` → `prompts/synthesize-comparison.txt` Replace template vars: - `{{query}}` → user query - `{{today}}` → current date (YYYY-MM-DD) - `{{language}}` → language param - `{{source_count}}` → number of successfully fetched sources - `{{dimensions_or_auto}}` → dimensions param (or "auto") - `{{fetched_content_blocks}}` → build as: ``` [Source 1] (url1) <content> --- [Source 2] (url2) <content> ``` Call LLM with the filled prompt. The output is the final report. If `--output FILE` is set, write the report to that file. Otherwise, print to the channel. --- ## ⚙️ Configuration Defaults are in `config.yaml`. Override via CLI flags above. --- ## 📂 Project Structure ``` skills/deep-scout/ ├── SKILL.md ← This file (agent instructions) ├── config.yaml ← Default parameter values ├── prompts/ │ ├── filter.txt ← Stage 2: relevance scoring prompt │ ├── synthesize-report.txt ← Stage 4: report/bullets/timeline synthesis │ ├── synthesize-comparison.txt← Stage 4: comparison table synthesis │ └── browser-extract.txt ← Stage 3: browser snapshot extraction ├── scripts/ │ ├── run.sh ← CLI entrypoint (emits pipeline actions) │ └── firecrawl-wrap.sh ← Firecrawl CLI wrapper with fallback handling └── examples/ └── openclaw-acquisition.md ← Example output: OpenClaw M&A intelligence ``` --- ## 🔧 Error Handling | Scenario | Handling | |----------|----------| | All fetch attempts fail | Use snippet from Stage 1; mark `[snippet only]` | | Search returns 0 results | Retry with `freshness: py`; error if still 0 | | Firecrawl not installed | `firecrawl-wrap.sh` outputs `FIRECRAWL_UNAVAILABLE`, skip silently | | Browser tool unavailable | Skip Tier 3; proceed with available content | | LLM synthesis exceeds context | Trim sources proportionally, prioritize high-score sources | | Rate limit on Brave API | Wait 2s, retry once | --- ## 📋 Example Outputs See `examples/openclaw-acquisition.md` for a full sample report. --- *Deep Scout v0.1.0 · OpenClaw Skills · clawhub: deep-scout*

标签

skill ai

通过对话安装

该技能支持在以下平台通过对话安装:

OpenClaw WorkBuddy QClaw Kimi Claude

方式一:安装 SkillHub 和技能

帮我安装 SkillHub 和 deep-scout-1776292086 技能

方式二:设置 SkillHub 为优先技能安装源

设置 SkillHub 为我的优先技能安装源,然后帮我安装 deep-scout-1776292086 技能

通过命令行安装

skillhub install deep-scout-1776292086

下载 Zip 包

⬇ 下载 deep-scout v0.1.4

文件大小: 17.51 KB | 发布时间: 2026-4-16 18:44

v0.1.4 最新 2026-4-16 18:44
Added simplified installation instructions to SKILL.md and README.md.

Archiver·手机版·闲社网·闲社论坛·羊毛社区· 多链控股集团有限公司 · 苏ICP备2025199260号-1

Powered by Discuz! X5.0   © 2024-2025 闲社网·线报更新论坛·羊毛分享社区·http://xianshe.com

p2p_official_large
返回顶部