The single source of truth for Reddit scraping in Yalc workflows. Knows the working actor IDs, the broken paths, and the per-client spend caps.
Any of these natural language phrases activates the skill inside Claude Code.
The Apify Reddit Scraping skill is the canonical Yalc wrapper for Reddit data collection via Apify. It encapsulates the learned Earleads playbook: which actor IDs work (`oAuCIx3ItNrs2okjQ`), which paths return clean data (`/new/`, `/hot/`), which fail silently (keyword search on this actor is broken), and how to enforce per-client spend caps (logs at `logs/apify_spend_{agent}.json`).
For any Yalc workflow that touches Reddit (monitoring, repurposing, content discovery), this skill is the gateway. Direct Apify API calls bypass the playbook and risk burning budget or shipping bad data. The skill is the abstraction.
The Apify Reddit Scraping skill sits at the **intake** node for any Reddit-sourced data. It powers the daily Reddit monitoring agents (per client), the repurposing agents, and on-demand thread discovery.
The skill enforces per-client budget caps. Apify charges per compute unit; Earleads' rule of thumb is $2/day per monitoring client and $0.50-$1.50/day per community agent. Without the skill, an unbounded actor run can rack up significant charges fast.
The Reddit data intake. Yalc invokes this skill whenever a workflow needs Reddit posts, comments, or subreddit feeds. Output is structured JSON ready for downstream classification (sentiment, content fit) or writeback (Notion).
APIFY_TOKENThe skill assumes per-client spend logs exist under `logs/apify_spend_{agent}.json`. If a new client doesn't have a log yet, the skill creates one on first run. Daily caps are configured in the per-client agent YAML.
Earleads tested several Reddit actors. This one returns the cleanest JSON, has the most stable rate limits, and is actively maintained. Other actors either have stale data or unstable schemas.
Before each actor run, the skill reads the per-client spend log and estimates the cost of the planned scrape. If the day's spend would exceed the cap, the run is aborted with a clear error message.
For monitoring (catch new posts), run /new/ every 1 to 4 hours. For trending content, run /hot/ once per day. Don't poll more frequently than 1 hour without a strong reason.
Rate limits surface as 429 errors from Apify. The skill catches these, waits 60 seconds, and retries once. If the second attempt fails, the run is logged and the schedule moves on.
Verified empirically in the Earleads playbook. The actor's keyword search returns inconsistent or empty results. Use subreddit feed scraping instead and filter results client-side via Claude.
Yes via a different actor input mode. The skill exposes a `--include-comments` flag. Note that comment scraping multiplies cost (often 5x to 10x the post scrape) so use sparingly.
Clone the Yalc skill set, drop in your env, run from your next Claude Code session.
gh repo clone Othmane-Khadri/YALC-the-GTM-operating-system && cp -r YALC-the-GTM-operating-system/.claude/skills/apify-reddit-scraping ./.claude/skills/