Reddit Scraper - Posts, Comments & Users
Pricing
from $3.00 / 1,000 result scrapeds
Reddit Scraper - Posts, Comments & Users
Extract posts, comments, communities & user profiles from any subreddit at scale. Fetches all comments including hidden/collapsed ones. Breaks Reddit's 1000-post limit with date windowing. No login needed, no browser. $0.003 per result. Supports search, sorting, NSFW filtering & date filtering.
Pricing
from $3.00 / 1,000 result scrapeds
Rating
5.0
(1)
Developer

Better Devs Scrape
Actor stats
0
Bookmarked
12
Total users
3
Monthly active users
2 days ago
Last modified
Categories
Share
π‘ Reddit Scraper β Extract Posts, Comments, Users & Subreddit Data
The most feature-rich Reddit scraper on Apify. Extract posts, comments, communities, and user profiles from any subreddit, search query, or URL β at just $0.003 per result. No Reddit account needed, no browser, pure HTTP.
A powerful Reddit API alternative for market research, sentiment analysis, brand monitoring, lead generation, and academic research. Paste URLs or type subreddit names, click Run, get clean structured JSON, CSV, or Excel.
π Why Choose This Reddit Scraper?
| Feature | This Scraper | Other Reddit Scrapers |
|---|---|---|
| π° Price per result | $0.003 | $0.004β$0.005+ |
| β‘ Speed | 50 items in under 60 seconds | Minutes with browser |
| π¬ All comments | β Fetches collapsed/hidden comments | β Only top ~500 |
| π Break 1000-post limit | β Date windowing bypasses cap | β Stuck at ~1000 |
| π― Post filters | β Score, flair, domain, author | β None |
| π€ User profiles | β Karma, account age, mod status | β Posts only |
| ποΈ Community data | β Members, rules, description | β Not available |
| π‘οΈ Resilience | β Auto-retry + fallback mirrors | β Fails on 429/403 |
| π§ Memory | 256 MB (no browser) | 1 GB+ with browser |
| π± Simple input | β Just type subreddit names | β Full URLs only |
| π Login required | β No | Often yes |
π What Is Reddit Scraping?
Reddit scraping is the process of automatically extracting data from Reddit β the world's largest forum with 1.7+ billion monthly visits. Instead of manually copying posts and comments, a Reddit scraper collects structured data (titles, text, scores, timestamps, user info) from subreddits, search results, and user profiles at scale.
Common reasons to scrape Reddit:
- Market research β Track what customers say about your product or competitors
- Sentiment analysis β Feed Reddit discussions into NLP pipelines for opinion mining
- Content discovery β Find trending topics, viral posts, and emerging discussions
- Lead generation β Identify users discussing problems your product solves
- Brand monitoring β Get alerts when your brand is mentioned across subreddits
- Academic research β Build datasets of online discussions for social science studies
- SEO & content ideas β Discover what questions real people ask in any niche
- AI training data β Collect diverse conversational data for language model fine-tuning
This scraper works as a Reddit API alternative β no OAuth tokens, no rate limit headaches, no developer application needed. Just paste URLs or type subreddit names and get data.
π What Data Can You Extract?
| Category | What You Get |
|---|---|
| π Posts | Title, body, score, comment count, media URLs, flair, awards, gallery images, video URLs |
| π¬ Comments | Full comment trees including nested replies, scores, author flair, depth level |
| ποΈ Communities | Subreddit metadata, member counts, active users, description, rules, type |
| π€ Users | Karma breakdown, account age, premium status, moderator status, verification |
| π Search | Search across all of Reddit or within specific subreddits |
βοΈ How to Scrape Reddit
1οΈβ£ Paste Reddit URLs (subreddits, posts, user profiles)2οΈβ£ Set your limits (max posts, comments, items)3οΈβ£ Click "Start" and get clean, structured data
The scraper automatically:
- π¬ Fetches all comments β Including collapsed/hidden comments beyond Reddit's initial ~500
- π‘οΈ Resilient β Automatic retries with fallback mirrors when Reddit is unavailable
- π Anti-blocking β Built-in session management to avoid rate limits
- π Breaks pagination limits β Date windowing gets posts beyond Reddit's ~1000 cap
π₯ Input Example
{"subreddits": ["technology", "programming"],"searches": ["artificial intelligence"],"minScore": 50,"flairFilter": "Discussion","maxItems": 100,"maxPostCount": 25,"maxComments": 50,"includeNSFW": false,"proxy": { "useApifyProxy": true, "apifyProxyGroups": ["RESIDENTIAL"] }}
π Input Parameters
What to Scrape
| Parameter | Type | Default | Description |
|---|---|---|---|
startUrls | array | β | Reddit URLs to scrape (posts, subreddits, users, leaderboard) |
subreddits | string[] | β | Subreddit names to scrape (without r/ prefix) |
searches | string[] | β | Search queries to run on Reddit |
searchCommunityName | string | β | Restrict search to a specific subreddit |
searchPosts | boolean | true | Include posts in search results |
searchComments | boolean | false | Include comments in search results |
searchCommunities | boolean | false | Include subreddits in search results |
searchUsers | boolean | false | Include users in search results |
Limits
| Parameter | Type | Default | Description |
|---|---|---|---|
maxItems | integer | 50 | Maximum total items saved to dataset |
maxPostCount | integer | 25 | Max posts per subreddit or search |
maxComments | integer | 10 | Max comments per post (0 to skip) |
maxCommunitiesCount | integer | 2 | Max community listing pages |
maxUserCount | integer | 2 | Max user listing pages |
postsPerSource | integer | 0 | Max posts per individual source (0 = no limit) |
Sorting & Filtering
| Parameter | Type | Default | Description |
|---|---|---|---|
sort | string | "hot" | Sort order: relevance, hot, top, new, rising, comments |
time | string | "all" | Time filter: all, hour, day, week, month, year |
commentSort | string | "confidence" | Comment sort: confidence (best), top, new, controversial, old, qa |
includeNSFW | boolean | false | Include NSFW content |
postDateLimit | string | β | Only scrape posts after this date (YYYY-MM-DD) |
Post Filters
| Parameter | Type | Default | Description |
|---|---|---|---|
minScore | integer | β | Only posts with at least N upvotes |
maxScore | integer | β | Only posts with at most N upvotes |
minComments | integer | β | Only posts with at least N comments |
maxCommentsFilter | integer | β | Only posts with at most N comments |
flairFilter | string | β | Only posts matching this flair |
domainFilter | string | β | Only posts from this domain |
authorFilter | string | β | Only posts by this author |
postsPerSource | integer | 0 | Max posts per individual source |
Skip Options
| Parameter | Type | Default | Description |
|---|---|---|---|
skipComments | boolean | false | Skip comment extraction (faster) |
skipUserPosts | boolean | false | Skip user's submitted posts |
skipCommunity | boolean | false | Skip subreddit metadata |
Advanced
| Parameter | Type | Default | Description |
|---|---|---|---|
enableDateWindowing | boolean | false | Break past Reddit's ~1000 post limit using date-range windows |
proxy | object | Residential | Proxy config (residential strongly recommended) |
debugMode | boolean | false | Enable verbose logging |
π€ Sample Output
Post
{"id": "t3_1rda27h","parsedId": "1rda27h","url": "https://www.reddit.com/r/AskReddit/comments/1rda27h/what_screams_i_am_deeply_insecure/","username": "curious_mind","title": "What screams \"I am deeply insecure\" but people do it thinking it makes them look cool?","communityName": "r/AskReddit","body": null,"numberOfComments": 8432,"upVotes": 24567,"upVoteRatio": 0.94,"isVideo": false,"over18": false,"createdAt": "2026-02-24T08:15:00.000Z","scrapedAt": "2026-02-24T15:58:42.000Z","flair": null,"link": null,"imageUrls": [],"videoUrl": null,"isGallery": false,"stickied": false,"locked": false,"archived": false,"spoiler": false,"awardsCount": 12,"dataType": "post"}
Comment
{"id": "t1_lm8x9y2","parsedId": "lm8x9y2","postId": "1rda27h","url": "https://www.reddit.com/r/AskReddit/comments/1rda27h/-/lm8x9y2/","parentId": "t3_1rda27h","username": "witty_replier","communityName": "r/AskReddit","body": "People who brag about how little sleep they get.","createdAt": "2026-02-24T09:30:00.000Z","scrapedAt": "2026-02-24T15:58:43.000Z","upVotes": 3421,"numberOfReplies": 87,"depth": 0,"dataType": "comment"}
Community
{"id": "2qh1i","name": "t5_2qh1i","title": "Ask Reddit...","displayName": "AskReddit","numberOfMembers": 48000000,"activeUserCount": 12500,"subredditType": "public","description": "r/AskReddit is the place to ask and answer thought-provoking questions.","over18": false,"createdAt": "2008-01-25T00:00:00.000Z","scrapedAt": "2026-02-24T15:58:32.000Z","url": "https://www.reddit.com/r/AskReddit/","dataType": "community"}
User
{"id": "1w72lch","userId": "t2_1w72lch","url": "https://www.reddit.com/user/spez/","username": "spez","totalKarma": 654321,"postKarma": 123456,"commentKarma": 530865,"isGold": true,"isMod": true,"hasVerifiedEmail": true,"createdAt": "2005-06-06T00:00:00.000Z","scrapedAt": "2026-02-24T15:49:43.000Z","dataType": "user"}
π° How Much Does It Cost to Scrape Reddit?
This Actor uses Pay-Per-Event pricing. You only pay for results actually saved.
| Event | Cost |
|---|---|
| π Per result (post, comment, community, or user) | $0.003 |
π΅ Cost Examples
| Scenario | Items | Estimated Cost |
|---|---|---|
| 1 subreddit, 25 posts, 10 comments each | ~276 | ~$0.83 |
| 10 subreddits, 50 posts each, no comments | ~510 | ~$1.53 |
| Search query, 100 posts, 50 comments each | ~5,100 | ~$15.30 |
| Single post with all comments (2000+) | ~2,001 | ~$6.00 |
π No login, no browser, no hidden costs. Pure HTTP scraping means low compute costs for you.
π Integrate With Your Stack
Python
from apify_client import ApifyClientclient = ApifyClient("YOUR_API_TOKEN")run = client.actor("betterdevsscrape/reddit-scraper").call(run_input={"startUrls": [{"url": "https://www.reddit.com/r/technology/"}],"maxItems": 100,"maxPostCount": 25,"maxComments": 20,"proxy": {"useApifyProxy": True, "apifyProxyGroups": ["RESIDENTIAL"]},})for item in client.dataset(run["defaultDatasetId"]).iterate_items():print(f"[{item['dataType']}] {item.get('title') or item.get('body', '')[:80]}")
JavaScript / Node.js
import { ApifyClient } from 'apify-client';const client = new ApifyClient({ token: 'YOUR_API_TOKEN' });const run = await client.actor('betterdevsscrape/reddit-scraper').call({startUrls: [{ url: 'https://www.reddit.com/r/technology/' }],maxItems: 100,maxPostCount: 25,maxComments: 20,proxy: { useApifyProxy: true, apifyProxyGroups: ['RESIDENTIAL'] },});const { items } = await client.dataset(run.defaultDatasetId).listItems();items.forEach(item => {console.log(`[${item.dataType}] ${item.title || item.body?.slice(0, 80)}`);});
cURL
curl "https://api.apify.com/v2/acts/betterdevsscrape~reddit-scraper/runs" \-X POST \-H "Content-Type: application/json" \-H "Authorization: Bearer YOUR_API_TOKEN" \-d '{"startUrls": [{"url": "https://www.reddit.com/r/technology/"}],"maxItems": 50,"proxy": {"useApifyProxy": true, "apifyProxyGroups": ["RESIDENTIAL"]}}'
π¦ Export Options
- JSON β Full structured data
- CSV β Spreadsheet-ready format
- Excel β Direct download
π Integrations
Connect via Zapier, Make (Integromat), or any platform that supports REST APIs.
π― Use Cases
π Market Research & Competitor Analysis
Monitor product mentions, competitor discussions, and industry trends across subreddits. Scrape r/technology, r/startups, or niche communities to understand what customers want. Use minScore filters to focus on high-engagement discussions only.
πΌ Sentiment Analysis & Opinion Mining
Collect thousands of posts and comments about brands, products, or topics and feed them into NLP pipelines. The structured JSON output (with scores, timestamps, and nested comment trees) is ready for pandas, Spark, or any data processing tool.
π Content Discovery & Trend Detection
Find trending content, popular posts, and viral discussions in any niche. Use sort: "rising" to catch content before it goes viral, or sort: "top" with time: "week" to see the best of the week.
π Academic & Social Science Research
Gather large datasets of Reddit discussions for social media studies, discourse analysis, or behavioral research. Date windowing lets you build historical archives going back years.
π― Lead Generation & Sales Intelligence
Find users actively discussing problems your product solves. Filter by flair, domain, or author to pinpoint high-intent conversations. Scrape user profiles to understand potential leads.
π’ Brand Monitoring & Reputation Management
Track mentions of your brand, products, or competitors across Reddit in real-time. Set up scheduled runs via Apify to get daily or weekly reports delivered to your inbox, Slack, or CRM.
π SEO Research & Content Ideas
Discover what questions real people ask in your industry. Reddit threads are a goldmine for blog topics, FAQ pages, and long-tail keyword research.
π€ AI Training Data & LLM Fine-Tuning
Collect diverse conversational data from Reddit's thousands of communities. Posts and comment trees provide natural dialogue structure ideal for training chatbots, classifiers, and language models.
π‘ Tips for Effective Scraping
π§ͺ Start Small
Begin with maxItems: 10 to test your configuration before scaling up.
π Always Use Residential Proxies
Reddit blocks datacenter IPs aggressively. The scraper defaults to Apify residential proxies.
π Use Search for Discovery
searches finds content across all of Reddit. startUrls is for specific pages you already know.
β‘ Skip Comments for Speed
Each comment counts as a result. Set maxComments: 0 or skipComments: true if you only need post data.
π Date Windowing for Archives
Enable enableDateWindowing with high maxPostCount to scrape older posts beyond Reddit's 1000-post pagination cap.
π Sort Matters
Use "hot" for current popular content, "top" + time: "week" for the best of the week, "new" for chronological.
π§ͺ Example Configurations
Basic: Scrape one subreddit
{"startUrls": [{ "url": "https://www.reddit.com/r/technology/" }],"maxItems": 50}
Search across all of Reddit
{"searches": ["artificial intelligence"],"maxItems": 100,"maxPostCount": 50,"sort": "relevance"}
Scrape a specific post with all comments
{"startUrls": [{ "url": "https://www.reddit.com/r/AskReddit/comments/abc123/some_post/" }],"maxComments": 5000,"maxItems": 10000}
Scrape user profiles
{"startUrls": [{ "url": "https://www.reddit.com/user/spez/" }],"maxItems": 50}
Filter by score and flair
{"subreddits": ["programming", "javascript"],"minScore": 100,"flairFilter": "Discussion","maxItems": 50,"skipComments": true}
Only YouTube links from a subreddit
{"subreddits": ["technology"],"domainFilter": "youtube.com","minScore": 50,"maxItems": 100}
Track a specific author
{"startUrls": [{ "url": "https://www.reddit.com/r/announcements/" }],"authorFilter": "spez","maxItems": 50}
Large-scale subreddit archive with date windowing
{"startUrls": [{ "url": "https://www.reddit.com/r/technology/" }],"maxPostCount": 5000,"maxItems": 50000,"enableDateWindowing": true,"skipComments": true}
β Frequently Asked Questions
Does this require a Reddit account or API key?
No. This scraper works without any Reddit account, OAuth token, or API key. It accesses only publicly visible content β no authentication needed.
Is this a Reddit API alternative?
Yes. Unlike the official Reddit API (which requires developer applications, OAuth, and has strict rate limits), this scraper extracts the same data with zero setup. Just provide URLs or subreddit names and run.
Why are residential proxies required?
Reddit actively blocks datacenter IP ranges. Residential proxies provide real IP addresses that Reddit doesn't block. The scraper defaults to Apify's residential proxy pool for maximum reliability.
Can I scrape private or quarantined subreddits?
No. Only public subreddits and posts are accessible. Quarantined subreddits require a logged-in session which this scraper does not use.
What happens when Reddit rate limits or blocks?
The scraper automatically retries failed requests with exponential backoff and falls back to mirror sites when Reddit is temporarily unavailable. Sessions are rotated to avoid IP bans.
How do I scrape more than 1000 posts from a subreddit?
Enable enableDateWindowing in the input. This splits the request into weekly date-range search queries that bypass Reddit's ~1000 post pagination limit. You can scrape entire subreddit archives this way.
How do I get all comments on a post, including hidden ones?
Just set maxComments to a high number. The scraper automatically detects collapsed/hidden comment threads (Reddit's "load more comments") and fetches them via the morechildren API. No extra configuration needed.
Can I export Reddit data as CSV or Excel?
Yes. Apify datasets support JSON, CSV, Excel, XML, and RSS exports. You can also connect via API, Zapier, Make (Integromat), or Google Sheets.
Can I schedule Reddit scraping on a recurring basis?
Yes. Use Apify's built-in scheduler to run the scraper hourly, daily, or weekly. Combine with webhooks or integrations to get results delivered to Slack, email, Google Sheets, or your own API.
How fast is it?
50 items in under 60 seconds. Pure HTTP scraping with no browser means low memory usage (~256 MB) and fast execution. Most runs complete in under 2 minutes.
What output format does the data come in?
Each item has a dataType field ("post", "comment", "community", or "user") so you can easily filter. All items include timestamps, IDs, and URLs for traceability.
βοΈ Legal
Web scraping publicly available data is generally legal. This Actor extracts only publicly visible content from Reddit.
Users are responsible for:
- β Complying with applicable data protection laws (GDPR, CCPA, etc.)
- β Respecting Reddit's terms of service
- β Using extracted data ethically and legally
If you're unsure whether your use case is legitimate, consult your lawyers. You can also read Apify's blog post on the legality of web scraping.
π¬ Support & Feedback
- π‘ Feature requests? We'd love to hear what you need
- β Questions? Check the FAQ above or reach out
- β Happy with the results? Leave us a review on the Apify Store!
π More Scrapers by BetterDevsScrape
| Scraper | Description |
|---|---|
| π Google Maps Scraper | Extract businesses, reviews, images, contacts & emails from Google Maps |
| π Contact Details Extractor | Extract emails, phone numbers & 25+ social media profiles from any website |
Built with β€οΈ by BetterDevsScrape | View on Apify Store