Reddit Scraper OLD avatar
Reddit Scraper OLD

Pricing

Pay per event

Go to Store
Reddit Scraper OLD

Reddit Scraper OLD

Developed by

Fatih Tahta

Fatih Tahta

Maintained by Community

NEW AND FASTER MODEL IS OUT - Most affordable, enterprise-grade Reddit scraper. Extracts full post data from any search query using residential proxies to bypass blocks. Delivers clean JSON for market research and analysis. Cost-effective at pay-as-you-go $1.80 per 1k results. Updated weekly.

5.0 (2)

Pricing

Pay per event

2

2

2

Last modified

2 days ago

The Reddit Scraper

Slug: fatihtahta/reddit-scraper Price: $1.80 per 1,000 saved posts (everything else included)

Scrape Reddit search results fast and clean. JSON only, no HTML parsing, no proxy headaches. Hand it some queries, pick sort & time filters, toggle NSFW, cap how many posts you want, and you’re done.

What you get • Full post payload: id, title, selftext, author, score, upvote_ratio, num_comments, subreddit, created_utc, url, kind: "post" • Sort + time filters: sort = new, top, hot, relevance, comments; timeframe = hour, day, week, month, year, all (where Reddit supports it) • NSFW switch: includeNsfw (false by default) • Result cap: maxPosts stops the run once the limit is reached • Proxy rotation: Residential IPs via Apify Proxy (auto) • Output ready for anything: JSON in an Apify Dataset (download as CSV, XLSX, Parquet, via API, etc.) • Built with TypeScript, Crawlee 3, got-scraping

How it works (under the hood) 1. For each query, call https://old.reddit.com/search.json with your sort/time/NSFW params. 2. Follow the after cursor up to maxPages. 3. Fetch each post’s .json?raw_json=1 endpoint once. 4. Push a normalized object per post to the Dataset.

Input schema (fields) • queries (string array, required): search phrases • sort (string, optional, default: new): new, top, hot, relevance, comments • timeframe (string, optional, default: day): hour, day, week, month, year, all (used with top, relevance, comments) • includeNsfw (boolean, optional, default: false): include over_18 results • maxPages (integer, optional, default: 1): number of result pages per query (≈100 posts/page) • maxPosts (integer, optional, default: 1000): hard limit on how many posts to save • maxConcurrency (integer, optional, default: 120): tune only if you fork/self-host • proxyConfiguration (object, optional): Apify Proxy config; defaults to residential rotation

Example input

{ "queries": ["web scraping", "data extraction"], "sort": "top", "timeframe": "week", "includeNsfw": false, "maxPages": 2, "maxPosts": 250, "proxyConfiguration": { "useApifyProxy": true, "apifyProxyGroups": ["RESIDENTIAL"] } }

Output (one item per post)

{ "kind": "post", "query": "web scraping", "id": "18xyz1", "title": "Best practices for large-scale scraping?", "body": "I’m interested in…", "author": "data_guru", "score": 152, "upvote_ratio": 0.93, "num_comments": 41, "subreddit": "datascience", "created_utc": "2025-08-24T14:32:19.000Z", "url": "https://www.reddit.com/r/datascience/comments/18xyz1/best_practices_for_large_scale_scraping/" }

Download via API: https://api.apify.com/v2/datasets/<DATASET_ID>/items?format=json

Formats: JSON, CSV, XLSX, Parquet, XML, etc.

Quick start on Apify 1. Click Try for free. 2. Fill the form with your queries and options. 3. Run it. Watch logs and dataset populate in real time. 4. Schedule it daily/weekly to keep data fresh. 5. Pull results via the Dataset API into Python, Zapier/Make, N8N, Tableau, Power BI, etc.

Pricing • $1.80 per 1,000 stored posts • All infra & proxy costs bundled in. You only pay for successful results. ⸻

Changelog • 2025-08-26 — v1.0.0: Initial public release

Support

Questions or custom needs? Email fatih@fatihtahta.com

Happy scraping!

Share Actor: