rss-agent-viewer

AI agent-friendly CLI RSS/Atom feed viewer with discovery, search, and OPML import/export. Use when an agent needs to manage RSS feeds, read articles, or run feed discovery from the terminal. Triggered by: "read RSS feeds", "view RSS", "RSS reader CLI", "rss-viewer", "discover RSS feeds", "import OPML", "search feeds", or "discover-search feeds".

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "rss-agent-viewer" with this command: npx skills add brooksy4503/rss-agent-viewer/brooksy4503-rss-agent-viewer-rss-agent-viewer

RSS Agent Viewer

CLI RSS/Atom feed viewer with built-in discovery, caching, and search.

Quick start

npx -y rss-agent-viewer init
npx -y rss-agent-viewer discover https://vercel.com
npx -y rss-agent-viewer add https://vercel.com/atom
npx -y rss-agent-viewer read

Core workflow

# Initialize config + database
rss-viewer init

# Discover feeds for a site
rss-viewer discover https://example.com

# Subscribe to a feed (optional: --timeout for slow feeds)
rss-viewer add https://example.com/feed.xml
rss-viewer add https://slow-site.com/feed.xml --timeout 30000

# List feeds and read articles
rss-viewer feeds
rss-viewer read

Common commands

rss-viewer init
rss-viewer add <url>
rss-viewer discover <url>
rss-viewer feeds
rss-viewer remove <url>
rss-viewer read [url]
rss-viewer search <query>
rss-viewer discover-search <query>
rss-viewer import <file>
rss-viewer export
rss-viewer cache <action>
rss-viewer cleanup

Usage patterns

Discover and subscribe

rss-viewer discover https://example.com
rss-viewer add https://example.com/rss.xml

Read a single feed

rss-viewer read https://example.com/rss.xml

Read all feeds (fetches fresh by default)

# Fetches fresh articles from all feeds in parallel
rss-viewer read

# Use cached data only (skip network requests)
rss-viewer read --cached

# Limit results
rss-viewer read --limit 10

# One latest article per feed (feed-diverse results)
rss-viewer read --latest-per-feed --limit 10

# Timeouts (for slow feeds or many feeds)
rss-viewer read --timeout 15000 --overall-timeout 180000

Clean up invalid feeds

# Remove broken and duplicate feeds
rss-viewer cleanup

# Preview what would be removed
rss-viewer cleanup --dry-run

# Only remove broken feeds
rss-viewer cleanup --broken

# Only remove duplicates
rss-viewer cleanup --duplicates

Search across all feeds

# Local database search (enhanced with full-text search)
rss-viewer search "open source"

# Web search + discovery + add + search in one command
rss-viewer discover-search "Rust programming" --auto-add --read

# Using Exa API (BYOK)
rss-viewer discover-search "AI safety" \
  --provider exa \
  --max-results 5 \
  --auto-add \
  --read

Import OPML

rss-viewer import feeds.opml

Export feeds

rss-viewer export

Search Options

Local Database Search

rss-viewer search "React 19" --limit 10
rss-viewer search "TypeScript" --author "Dan" --since "2024-01-01"

Web Search with Discovery

# Use agent's built-in search (default)
rss-viewer discover-search "micro-frontends" --auto-add --read

# Use Exa API (requires EXA_API_KEY)
export EXA_API_KEY="your-api-key"
rss-viewer discover-search "WebGPU" \
  --provider exa \
  --max-results 5 \
  --category Development \
  --auto-add \
  --read \
  --limit 20

Configuration

Environment Variables:

EXA_API_KEY="your-api-key"                    # Exa API key (optional)
RSS_FEED_TIMEOUT=10000                        # Per-feed fetch timeout (ms)
RSS_VIEWER_SEARCH_PROVIDER="agent|exa"        # Search provider (default: agent)
RSS_VIEWER_MAX_WEB_RESULTS=10                 # Max web search results
RSS_VIEWER_SEARCH_LIMIT=20                    # Max article results
RSS_VIEWER_BOOST_RECENT=false                 # Boost recent articles in search
EXA_API_URL="https://api.exa.ai/search"       # Custom Exa endpoint (optional)

Config File (~/.config/rss-viewer/config.json):

{
  "webSearchProvider": "exa",
  "exaApiKey": "your-api-key",
  "feedTimeout": 10000,
  "overallTimeout": 120000,
  "maxWebResults": 10,
  "searchResultsLimit": 20,
  "boostRecentSearch": false
}

When to use this tool

  • Manage and read RSS/Atom feeds from the terminal
  • Discover feed URLs from a website
  • Search across multiple subscriptions with full-text search
  • Discover new feeds from web search queries
  • Automate feed discovery and subscription workflow
  • Import or export subscriptions via OPML
  • Clean up broken or duplicate feeds from database

More information

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

arxiv-paper-writer

Use this skill whenever the user wants Claude Code to write, scaffold, compile, debug, or review an arXiv-style academic paper, especially survey papers with LaTeX, BibTeX citations, TikZ figures, tables, and PDF output. This skill should trigger for requests like writing a full paper, creating an arXiv paper project, turning a research topic into a LaTeX manuscript, reproducing the Paper-Write-Skill-Test agent-survey workflow, or setting up a Windows/Linux Claude Code paper-writing loop.

Archived SourceRecently Updated
Coding

cli-proxy-troubleshooting

排查 CLI Proxy API(codex-api-proxy)的配置、认证、模型注册和请求问题。适用场景包括:(1) AI 请求报错 unknown provider for model, (2) 模型列表中缺少预期模型, (3) codex-api-key/auth-dir 配置不生效, (4) CLI Proxy 启动后 AI 无法调用, (5) 认证成功但请求失败或超时。包含源码级排查方法:模型注册表架构、认证加载链路、 SanitizeCodexKeys 规则、常见错误的真实根因。

Archived SourceRecently Updated
Coding

visual-summary-analysis

Performs AI analysis on input video clips/image content and generates a smooth, natural scene description. | 视觉摘要智述技能,对传入的视频片段/图片内容进行AI分析,生成一段通顺自然的场景描述内容

Archived SourceRecently Updated
Coding

frontend-skill

全能高级前端研发工程师技能。擅长AI时代前沿技术栈(React最新 + shadcn/ui + Tailwind CSS v4 + TypeScript + Next.js),精通动效库与交互特效开发。采用Glue Code风格快速实现代码,强调高质量产品体验与高度友好的UI视觉规范。在组件调用、交互特效、全局Theme上保持高度规范:绝不重复造轮子,相同逻辑出现两次即封装为组件。具备安全意识,防范各类注入攻击。开发页面具有高度自适应能力,响应式设计贯穿始终。当用户无特殊技术栈要求时,默认采用主流前沿技术栈。

Archived SourceRecently Updated