Skip to content

Latest commit

 

History

History
75 lines (55 loc) · 1.96 KB

File metadata and controls

75 lines (55 loc) · 1.96 KB

ClawFeed

AI-powered news digest tool. Automatically generates structured summaries (4H/daily/weekly/monthly) from Twitter and RSS feeds.

Setup

# Install dependencies
npm install

# Copy environment config
cp .env.example .env
# Edit .env with your settings

# Start API server
npm start

Environment Variables

Configure in .env file:

Variable Description Default
DIGEST_PORT Server port 8767
GOOGLE_CLIENT_ID OAuth client ID (optional) -
GOOGLE_CLIENT_SECRET OAuth secret (optional) -
SESSION_SECRET Session encryption key -

API Server

Runs on port 8767 by default. Set DIGEST_PORT env to change.

Endpoints

Method Path Description Auth
GET /api/digests List digests (?type=4h|daily|weekly&limit=20&offset=0) -
GET /api/digests/:id Get single digest -
POST /api/digests Create digest (internal) -
GET /api/auth/google Start Google OAuth flow -
GET /api/auth/callback OAuth callback endpoint -
GET /api/auth/me Get current user info Yes
POST /api/auth/logout Logout user Yes
GET /api/marks List user bookmarks Yes
POST /api/marks Add bookmark Yes
DELETE /api/marks/:id Remove bookmark Yes
GET /api/config Get configuration -
PUT /api/config Update configuration -

Web Dashboard

Serve web/index.html via your reverse proxy or any static file server.

Templates

  • templates/curation-rules.md — Customize feed curation rules
  • templates/digest-prompt.md — Customize the AI summarization prompt

Configuration

Copy config.example.json to config.json and edit. See README for details.

Reverse Proxy (Caddy example)

handle /digest/api/* {
    uri strip_prefix /digest/api
    reverse_proxy localhost:8767
}
handle_path /digest/* {
    root * /path/to/clawfeed/web
    file_server
}