ClawFeed: AI-Powered Multi-Source News Aggregator — Stop Scrolling, Start Knowing
Spending hours scrolling Twitter and RSS feeds, afraid of missing hot topics, only to be drowned in noise and growing more anxious. ClawFeed solves exactly this — using AI to automatically filter content from Twitter, RSS, HackerNews, Reddit, GitHub Trending, and more, generating structured summaries.
It supports four digest frequencies: 4-hourly briefs, daily highlights, weekly reviews, and monthly summaries. The Source Packs feature lets users bundle and share curated source sets with the community. Interesting items can trigger AI-powered deep analysis (Mark & Deep Dive) that goes beyond summaries into real insight. Every user's digest auto-generates RSS/JSON Feed subscriptions, enabling "human curation + AI processing" information redistribution.
Clean tech stack: Node.js backend + SQLite zero-config storage + SPA frontend with English/Chinese UI and dark mode. One-click install via ClawHub, runs as an OpenClaw or Zylos skill, or deploys standalone. Google OAuth enables multi-user management of sources and bookmarks.
Background
Spending hours daily scrolling Twitter and RSS feeds, terrified of missing hot topics, only to be drowned in noise — the more you scroll, the more anxious you get. ClawFeed's core philosophy is "Stop scrolling. Start knowing." — using AI to filter signal from noise across thousands of information sources.
Core Features
Multi-Frequency Digests
Four automatic summary granularities: 4-hourly briefs, daily highlights, weekly reviews, monthly summaries. Each frequency generated independently, subscribe as needed.
Rich Source Support
- **Twitter/X**: Follow specific users (@karpathy) or Twitter lists
- **RSS/Atom**: Any RSS feed
- **HackerNews**: HN front page
- **Reddit**: Specific subreddits (e.g., /r/MachineLearning)
- **GitHub Trending**: Filter by programming language
- **Website scraping**: Any web page
- **Custom API**: JSON endpoints
- **Digest Feed**: Subscribe to other ClawFeed users' digests for meta-curation
Source Packs
Bundle curated source sets into Packs for one-click community installation. Think npm packages for information curation — interested in AI frontiers? Install someone's AI Source Pack.
Mark & Deep Dive
Bookmark interesting items while browsing. Bookmarked items trigger AI deep analysis generating detailed reports beyond summaries — real depth, not just overviews.
Feed Output
Every user's digest auto-generates three subscription formats:
- HTML page (`/feed/:slug`)
- JSON Feed (`/feed/:slug.json`)
- RSS (`/feed/:slug.rss`)
Your AI-curated results become subscribable, enabling chain propagation of information.
Smart Curation
Configure filtering via `templates/curation-rules.md` and customize AI output format via `templates/digest-prompt.md`.
Technical Architecture
- **Backend**: Node.js, default port 8767
- **Database**: SQLite (zero-config, portable)
- **Auth**: Google OAuth 2.0 (multi-user; read-only without OAuth)
- **Frontend**: SPA with English/Chinese UI, Dark/Light theme toggle
- **API**: Full RESTful — digests, auth, bookmarks, sources, Source Packs, Feed output, changelog, roadmap
Deployment
ClawHub one-click install (`clawhub install clawfeed`), OpenClaw/Zylos skill, or standalone (`git clone → npm install → npm start`). Supports Caddy reverse proxy with path prefix.
Use Cases
- Personal information aggregation replacing anxious Twitter/HN/Reddit scrolling
- Team knowledge sharing via Source Packs and Feed output
- AI Agent integration as OpenClaw/Zylos skill
- Information redistribution through curated Feed subscriptions
License: MIT | Author: Kevin He | Live Demo: https://clawfeed.kevinhe.io