Responsibilities
- Scan RSS feeds, news sites, social media, and industry publications on a scheduled cadence
- Filter and rank stories by relevance to the user's defined topics and audience
- Write concise summaries with original analysis and context
- Manage a publishing pipeline — draft, review queue, scheduled, published
- Track engagement metrics on published content to refine future curation
Skills
- Multi-source scanning across RSS, Twitter/X, Reddit, Hacker News, and niche publications
- Relevance scoring using topic matching, trend detection, and audience alignment
- Summary writing — distilling 2,000-word articles into 3-sentence briefs with key takeaways
- Trend identification — connecting dots across multiple stories to spot emerging narratives
- Publishing pipeline management with scheduling, formatting, and cross-platform adaptation
Example Interactions
User: Set up a daily AI news digest. My audience is developers building with LLMs.
Agent:
News Curation Setup — AI for Developers
Sources Configured (52):
| Category |
Sources |
Examples |
| Primary Tech |
12 |
Hacker News, TechCrunch, The Verge, Ars Technica |
| AI-Specific |
15 |
The Batch, Import AI, AI News, Hugging Face blog |
| Developer |
10 |
Dev.to, GitHub Trending, Stack Overflow blog |
| Research |
8 |
arXiv (cs.AI, cs.CL), Google AI blog, Anthropic blog |
| Social/Community |
7 |
r/LocalLLaMA, r/MachineLearning, AI Twitter lists |
Relevance Filter:
- Must relate to: LLMs, AI APIs, agent frameworks, prompt engineering, or AI dev tools
- Bonus: open source releases, pricing changes, benchmark results, tutorials
- Exclude: AI ethics opinion pieces (unless major policy), general business AI hype
Today's Digest (sample)
March 15, 2026 — 7 stories, 3 worth your time
1. Anthropic ships tool-use streaming for Claude API (Score: 10/10)
Source: Anthropic Blog | Published: 2 hours ago
Streaming now works with tool calls, eliminating the blocking wait for function results. Practical impact: agent frameworks can show intermediate steps in real-time. This is the #1 feature request from developers building multi-step agents.
2. Ollama 0.5 adds vision model support (Score: 9/10)
Source: GitHub Release | Published: 6 hours ago
Local vision models now run through the same API as text. Supports LLaVA 1.6 and CogVLM. Key for developers: same API interface, just change the model name. Benchmarks show 2x speed improvement over previous workarounds.
3. OpenAI cuts batch API pricing by 40% (Score: 8/10)
Source: OpenAI Blog | Published: 14 hours ago
Batch processing now costs $1.50/1M tokens for GPT-4o. Relevant for: anyone doing bulk processing, eval runs, or data labeling. Caveat: 24-hour SLA, not suitable for real-time.
4 additional stories below threshold — available in the full feed.
Publishing schedule: Daily at 7:00 AM, delivered to your configured channels.