ClawFeed: AI-Powered News Digest with Structured Summaries from Twitter/RSS and Web Dashboard
ClawFeed is an open-source AI news aggregation tool that automatically curates content from Twitter, RSS, HackerNews, Reddit, GitHub Trending and more, using AI for smart filtering and noise reduction to generate structured multi-frequency digests (4-hourly/daily/weekly/monthly). Key highlights: - **Source Packs**: Users can package and share curated source bundles with the community — like npm for information sources - **Mark & Deep Dive**: Bookmark content to trigger AI-powered deep analysis, going beyond summaries into real insight - **Feed Output**: Every user's digest is subscribable via RSS/JSON Feed, enabling "human curation + AI processing" information redistribution - **Multi-tenant**: SQLite + Google OAuth supporting independent source management per user Clean tech stack: Node.js backend + SQLite + SPA frontend. Works as an OpenClaw/Zylos skill or standalone deployment. A feature-complete, hackable reference for developers building personal information aggregation systems.
ClawFeed Overview
ClawFeed is an open-source AI news aggregation tool by Kevin He. Core philosophy: "Stop scrolling. Start knowing." — using AI to filter signal from noise across thousands of information sources.
Core
Features #
Multi-Frequency Digests
Four time granularities: 4-hourly briefs, daily highlights, weekly reviews, monthly summaries. Each frequency is independently generated; users choose their subscription level. #
Rich Source
Support - **Twitter/X**: Follow specific users (@karpathy) or Twitter lists - **RSS/Atom**: Any RSS feed - **HackerNews**: HN front page - **Reddit**: Specific subreddits (e.g., /r/MachineLearning) - **GitHub Trending**: Filter by programming language - **Website scraping**: Any web page - **Custom API**: JSON endpoints - **Digest Feed**: Subscribe to other ClawFeed users' digests (meta-curation) #
Source Packs
Users can bundle curated source sets into Packs and publish them for one-click community installation. Think npm packages but for information curation — interested in AI frontiers? Just install someone's AI Source Pack. #
Mark
& Deep Dive Bookmark interesting items while browsing digests. Bookmarked items trigger AI-powered deep analysis, generating detailed reports beyond summaries. #
Feed Output
Every user's digest auto-generates three subscription formats: - HTML page (`/feed/:slug`) - JSON Feed (`/feed/:slug.json`) - RSS (`/feed/:slug.rss`) Your AI-curated results become subscribable, enabling chain propagation of information.
Technical
Architecture - **Backend**: Node.js, port 8767 - **Database**: SQLite (zero-config, portable) - **Auth**: Google OAuth 2.0 (multi-user; read-only without OAuth) - **Frontend**: SPA with Dark/Light theme toggle - **i18n**: English + Chinese UI - **AI configurable**: `templates/curation-rules.md` for filtering rules, `templates/digest-prompt.md` for output format
Deployment
#
As OpenClaw
Skill ``` clawhub install clawfeed ``` Auto-detected via SKILL.md. Agent generates digests via cron, serves dashboard, handles bookmark commands. #
As Zylos
Skill ``` cd ~/.zylos/skills/ && git clone https://github.com/kevinho/clawfeed.git ``` #
Standalone
## API Full RESTful
API: digest CRUD, authentication, bookmarks, source management, Source Packs, Feed output, changelog and roadmap queries. Write operations require auth; read operations (digest browsing, Feed subscriptions) are public.
## Use Cases - Personal information aggregation: Replace manual Twitter/HN/Reddit scrolling - Team knowledge sharing: Via Source Packs and Feed output - AI Agent integration:
As OpenClaw/Zylos skill for automated operation **License**: MIT | **Live Demo**: https://clawfeed.kevinhe.io