返回顶部
🇺🇸 English
🇨🇳 简体中文
🇨🇳 繁體中文
🇺🇸 English
🇯🇵 日本語
🇰🇷 한국어
🇫🇷 Français
🇩🇪 Deutsch
🇪🇸 Español
🇷🇺 Русский
s

stripfeed

Fetch any URL as clean, AI-ready Markdown with token counts and caching. Strip ads, nav, scripts, and noise from web pages.

作者: admin | 来源: ClawHub
源自
ClawHub
版本
V 1.0.0
安全检测
已通过
334
下载量
免费
免费
0
收藏
概述
安装方式
版本历史

stripfeed

# StripFeed Convert any URL to clean Markdown optimized for LLM consumption. Strips ads, navigation, scripts, and noise. Returns token counts so you know exactly how much context you're using. ## When to use this skill Use StripFeed whenever you need to read a web page, article, documentation, or any URL content. It produces much cleaner output than raw HTML fetching and tells you the token cost. ## Authentication All requests require the `STRIPFEED_API_KEY` environment variable. Pass it as a Bearer token: ``` Authorization: Bearer $STRIPFEED_API_KEY ``` Get a free API key at https://www.stripfeed.dev (200 requests/month, no credit card). ## Fetch a single URL ```bash curl -s "https://www.stripfeed.dev/api/v1/fetch?url=URL_HERE" \ -H "Authorization: Bearer $STRIPFEED_API_KEY" ``` This returns clean Markdown directly as the response body (Content-Type: text/markdown). ### Parameters | Parameter | Required | Description | | ------------ | -------- | ----------------------------------------------------------------------------------------- | | `url` | Yes | The URL to fetch (must be http or https) | | `format` | No | Output format: `markdown` (default), `json`, `text`, `html`. Pro only except markdown. | | `selector` | No | CSS selector to extract specific content (e.g. `article`, `.content`, `#main`). Pro only. | | `cache` | No | Set to `false` to bypass cache and force a fresh fetch | | `ttl` | No | Cache TTL in seconds (default: 3600, max: 86400 for Pro) | | `max_tokens` | No | Truncate output to fit within this token budget | | `model` | No | AI model ID for cost tracking (e.g. `claude-sonnet-4-6`, `gpt-4o`) | ### JSON format (recommended for structured responses) When you need metadata alongside the content, use `format=json`: ```bash curl -s "https://www.stripfeed.dev/api/v1/fetch?url=URL_HERE&format=json" \ -H "Authorization: Bearer $STRIPFEED_API_KEY" ``` JSON response includes: ```json { "markdown": "# Page Title\n\nClean content...", "url": "https://example.com", "title": "Page Title", "tokens": 1250, "originalTokens": 15000, "savingsPercent": 91.7, "cached": false, "fetchMs": 430, "format": "json", "truncated": false, "selector": null, "model": null } ``` ### Response headers Every response includes these headers: - `X-StripFeed-Tokens` - Token count of the stripped content - `X-StripFeed-Original-Tokens` - Token count of the raw HTML (before stripping) - `X-StripFeed-Savings-Percent` - Percentage of tokens saved - `X-StripFeed-Cache` - `HIT` or `MISS` - `X-StripFeed-Fetch-Ms` - Time to fetch the URL (0 if cached) - `X-StripFeed-Format` - Output format used ## Batch fetch (Pro plan) Fetch up to 10 URLs in a single request: ```bash curl -s -X POST "https://www.stripfeed.dev/api/v1/batch" \ -H "Authorization: Bearer $STRIPFEED_API_KEY" \ -H "Content-Type: application/json" \ -d '{"urls": ["https://example.com", "https://example.org"]}' ``` URLs can also include selectors: ```json { "urls": [ { "url": "https://example.com", "selector": "article" }, { "url": "https://example.org" } ], "model": "claude-sonnet-4-6" } ``` Batch response: ```json { "results": [ { "url": "...", "markdown": "...", "tokens": 1250, "status": 200 }, { "url": "...", "error": "Failed to fetch", "status": 502 } ], "total": 2, "success": 1, "failed": 1 } ``` Individual URL errors don't break the batch. Check each result's `status` field. ## Error handling | Status | Meaning | | ------ | ---------------------------------------- | | 401 | Missing or invalid API key | | 403 | Feature requires Pro plan | | 422 | Invalid URL, format, or parameter | | 429 | Rate limit or monthly quota exceeded | | 502 | Target URL unreachable or returned error | | 504 | Target URL timed out (9s limit) | ## Tips - Default format is `markdown` which returns raw text. Use `format=json` when you need token counts and metadata in the response body. - Responses are cached for 1 hour by default. Use `cache=false` for real-time content. - The `max_tokens` parameter is useful to fit content within your context window budget. - Use `selector` to grab only the main content (e.g. `selector=article` or `selector=.post-content`) and skip sidebars/footers. - Free plan: 200 requests/month, markdown format only. Pro plan ($19/mo): 100K requests, all formats, selectors, batch endpoint.

标签

skill ai

通过对话安装

该技能支持在以下平台通过对话安装:

OpenClaw WorkBuddy QClaw Kimi Claude

方式一:安装 SkillHub 和技能

帮我安装 SkillHub 和 stripfeed-1776282136 技能

方式二:设置 SkillHub 为优先技能安装源

设置 SkillHub 为我的优先技能安装源,然后帮我安装 stripfeed-1776282136 技能

通过命令行安装

skillhub install stripfeed-1776282136

下载

⬇ 下载 stripfeed v1.0.0(免费)

文件大小: 2.6 KB | 发布时间: 2026-4-16 15:56

v1.0.0 最新 2026-4-16 15:56
Update skill

Archiver·手机版·闲社网·闲社论坛·羊毛社区· 多链控股集团有限公司 · 苏ICP备2025199260号-1

Powered by Discuz! X5.0   © 2024-2025 闲社网·线报更新论坛·羊毛分享社区·http://xianshe.com

p2p_official_large
返回顶部