
Reddit Scraper Search | Fast & Reliable | $1.8 / 1K
Pricing
$1.80 / 1,000 results

Reddit Scraper Search | Fast & Reliable | $1.8 / 1K
Most affordable, enterprise-grade Reddit scraper. Extracts full post data from any search query using residential proxies to bypass blocks. Delivers clean JSON for market research and analysis. Cost-effective at pay-as-you-go $1.80 per 1k results. Updated weekly.
5.0 (2)
Pricing
$1.80 / 1,000 results
2
5
5
Last modified
2 days ago
The Reddit Scraper
Slug: fatihtahta/reddit-scraper Price: $1.80 per 1,000 saved posts (everything else included)
Scrape Reddit search results fast and clean. JSON only, no HTML parsing, no proxy headaches. Hand it some queries, pick sort & time filters, toggle NSFW, cap how many posts you want, and you’re done.
⸻
What you get • Full post payload: id, title, selftext, author, score, upvote_ratio, num_comments, subreddit, created_utc, url, kind: "post" • Sort + time filters: sort = new, top, hot, relevance, comments; timeframe = hour, day, week, month, year, all (where Reddit supports it) • NSFW switch: includeNsfw (false by default) • Result cap: maxPosts stops the run once the limit is reached • Proxy rotation: Residential IPs via Apify Proxy (auto) • Output ready for anything: JSON in an Apify Dataset (download as CSV, XLSX, Parquet, via API, etc.) • Built with TypeScript, Crawlee 3, got-scraping
⸻
How it works (under the hood) 1. For each query, call https://old.reddit.com/search.json with your sort/time/NSFW params. 2. Follow the after cursor up to maxPages. 3. Fetch each post’s .json?raw_json=1 endpoint once. 4. Push a normalized object per post to the Dataset.
⸻
Input schema (fields) • queries (string array, required): search phrases • sort (string, optional, default: new): new, top, hot, relevance, comments • timeframe (string, optional, default: day): hour, day, week, month, year, all (used with top, relevance, comments) • includeNsfw (boolean, optional, default: false): include over_18 results • maxPages (integer, optional, default: 1): number of result pages per query (≈100 posts/page) • maxPosts (integer, optional, default: 1000): hard limit on how many posts to save • maxConcurrency (integer, optional, default: 120): tune only if you fork/self-host • proxyConfiguration (object, optional): Apify Proxy config; defaults to residential rotation
Example input
{ "queries": ["web scraping", "data extraction"], "sort": "top", "timeframe": "week", "includeNsfw": false, "maxPages": 2, "maxPosts": 250, "proxyConfiguration": { "useApifyProxy": true, "apifyProxyGroups": ["RESIDENTIAL"] } }
Output (one item per post)
{ "kind": "post", "query": "web scraping", "id": "18xyz1", "title": "Best practices for large-scale scraping?", "body": "I’m interested in…", "author": "data_guru", "score": 152, "upvote_ratio": 0.93, "num_comments": 41, "subreddit": "datascience", "created_utc": "2025-08-24T14:32:19.000Z", "url": "https://www.reddit.com/r/datascience/comments/18xyz1/best_practices_for_large_scale_scraping/" }
Download via API: https://api.apify.com/v2/datasets/<DATASET_ID>/items?format=json
Formats: JSON, CSV, XLSX, Parquet, XML, etc.
⸻
Quick start on Apify 1. Click Try for free. 2. Fill the form with your queries and options. 3. Run it. Watch logs and dataset populate in real time. 4. Schedule it daily/weekly to keep data fresh. 5. Pull results via the Dataset API into Python, Zapier/Make, N8N, Tableau, Power BI, etc.
⸻
Pricing • $1.80 per 1,000 stored posts • All infra & proxy costs bundled in. You only pay for successful results. ⸻
Changelog • 2025-08-26 — v1.0.0: Initial public release
⸻
Support
Questions or custom needs? Email fatih@fatihtahta.com
Happy scraping!
Share Actor: