Reddit Scraper Pro - Posts, Comments & Subreddits - No API Key
Pricing
Pay per usage
Reddit Scraper Pro - Posts, Comments & Subreddits - No API Key
Scrape Reddit posts, comments, subreddits, and user profiles at scale. No Reddit API key needed. Search Reddit by keyword, extract full comment threads, monitor subreddits. AI-ready JSON output.
Pricing
Pay per usage
Rating
0.0
(0)
Developer

George Kioko
Actor stats
0
Bookmarked
2
Total users
1
Monthly active users
a day ago
Last modified
Categories
Share
Reddit Scraper Pro - Posts, Comments & Subreddits - No API Key Required
Scrape Reddit posts, comments, user profiles, and entire subreddits at scale. No Reddit API key needed. No OAuth, no rate limit hassles, no application approval. Works with subreddit URLs, post URLs, user profiles, and keyword search.
Perfect for AI/LLM pipelines, market research, brand monitoring, sentiment analysis, content strategy, and academic research. Output is clean JSON optimized for ChatGPT, Claude, Gemini, LangChain, LlamaIndex, and other AI tools.
Why use this Reddit Scraper?
| Feature | This Scraper | Reddit API (free) | Manual Browsing |
|---|---|---|---|
| API key needed? | No | Yes (OAuth2) | No |
| Approval process? | None | Application required | None |
| Rate limits? | Auto-managed | 100 req/min | N/A |
| Posts + Comments? | Yes | Yes | One at a time |
| User profiles? | Yes | Limited | One at a time |
| Batch subreddits? | Yes | Pagination headaches | No |
| Search across Reddit? | Yes | Basic | Manual |
| AI-ready output? | Yes | Raw JSON | No |
| Cost per 1,000 results | $2 | Free (with limits) | Hours of work |
What data do you get?
Reddit Post Data
- Title, author, subreddit - Full post metadata
- Full text (self posts) - Complete selftext for text-based posts
- Score, upvote ratio - Engagement metrics for analysis
- Comment count - Know the discussion size
- Flair and awards - Post categorization and recognition
- Media detection - Images, videos, galleries, embeds identified
- Permalink and URL - Direct links to post and external content
- Timestamps - Created date in ISO format
Reddit Comment Data
- Full comment text - Including HTML formatting
- Author and score - Who said it, how it was received
- Thread depth - Nested reply structure preserved
- Parent ID - Reconstruct conversation threads
- Submitter flag - Know when OP replies
Reddit User Profile Data
- Username and display name
- Total karma (post + comment + awardee + awarder breakdown)
- Account age
- Moderator and verified status
- Profile description and avatar
Key Features
- No API Key Required - Scrapes Reddit's public JSON endpoints directly. Zero setup. No OAuth flow. No approved application needed.
- Subreddit Scraping - Extract all posts from any subreddit with pagination. Sort by hot, new, top, rising, or controversial.
- Reddit Search - Search across all of Reddit or within a specific subreddit. Sort by relevance, hot, new, top, or most comments.
- Full Comment Threads - Extract comments with configurable depth. Get top-level comments only or dive deep into reply chains.
- User Profile Scraping - Get user profile info, karma breakdown, and their submitted posts.
- NSFW Filtering - Include or exclude NSFW content with a single toggle.
- Smart Rate Limiting - Built-in request throttling and retry logic. Handles 429s gracefully with exponential backoff.
- Proxy Support - Use Apify residential proxies for large batches to avoid IP-based rate limits.
- Pay Per Result - Only pay for data actually extracted. No flat monthly fee. $0.002 per result.
Input Configuration
| Field | Type | Required | Description |
|---|---|---|---|
urls | Array | Yes | Reddit URLs (subreddit, post, user) or plain text search queries |
sort | Enum | No | Subreddit sort: "hot", "new", "top", "rising", "controversial". Default: "hot" |
searchMode | Enum | No | Search sort: "relevance", "hot", "new", "top", "comments". Default: "relevance" |
searchTime | Enum | No | Time filter: "hour", "day", "week", "month", "year", "all". Default: "all" |
scrapeComments | Boolean | No | Also extract comments from each post. Default: false |
maxComments | Integer | No | Max top-level comments per post, 1-500. Default: 50 |
commentDepth | Integer | No | Reply nesting depth, 1-10. Default: 2 |
maxResults | Integer | No | Maximum total results, 1-5000. Default: 100 |
includeNSFW | Boolean | No | Include NSFW posts. Default: false |
proxyConfiguration | Object | No | Apify proxy settings. Recommended for large batches. |
Supported URL Formats
# Subredditshttps://www.reddit.com/r/technologyhttps://old.reddit.com/r/pythonhttps://www.reddit.com/r/MachineLearning/topr/datascience# Individual postshttps://www.reddit.com/r/technology/comments/abc123/some_post_title/https://old.reddit.com/r/python/comments/xyz789/discussion_thread/# User profileshttps://www.reddit.com/user/spezhttps://www.reddit.com/u/some_usernameu/username# Search queries (plain text)best AI tools 2026python web scraping tutorialmachine learning beginner
Output Examples
Post Output
{"type": "post","id": "1abc234","title": "GPT-5 just dropped and it's incredible","author": "ai_enthusiast","subreddit": "technology","text": "Full self-text content here...","url": "https://www.reddit.com/r/technology/comments/1abc234/gpt5_just_dropped/","permalink": "https://www.reddit.com/r/technology/comments/1abc234/gpt5_just_dropped/","score": 15432,"upvoteRatio": 0.94,"numComments": 2341,"isNSFW": false,"flair": "AI","awards": 23,"media": null,"createdUtc": "2026-03-01T15:30:00.000Z","scrapedAt": "2026-03-02T21:34:36.000Z"}
Comment Output
{"type": "comment","id": "k5m6n7o","postId": "1abc234","author": "tech_reviewer","subreddit": "technology","text": "I've been testing it for 3 hours and the reasoning capability is insane.","score": 892,"depth": 0,"isSubmitter": false,"createdUtc": "2026-03-01T16:15:00.000Z","scrapedAt": "2026-03-02T21:34:37.000Z"}
User Profile Output
{"type": "user","username": "spez","displayName": "u/spez","totalKarma": 1234567,"postKarma": 456789,"commentKarma": 678901,"isMod": true,"isVerified": true,"accountCreatedUtc": "2005-06-06T00:00:00.000Z","scrapedAt": "2026-03-02T21:34:38.000Z"}
Pricing
Pay Per Event pricing - only pay for successful results.
| Event | Price | Description |
|---|---|---|
| Actor start | $0.005 | One-time fee per run |
| Result scraped | $0.002 | Per post, comment, or user profile extracted |
$2 per 1,000 results. Compare to Reddit API development time (days of OAuth setup) or manual data collection (hours per subreddit).
Use Cases
AI & Machine Learning
- Training Data - Build datasets from specific subreddits for fine-tuning LLMs on domain-specific content
- RAG Pipelines - Feed Reddit discussions into vector databases for retrieval-augmented generation
- Sentiment Analysis - Analyze public opinion on products, brands, or topics at scale
- Trend Detection - Monitor emerging topics across subreddits using keyword search
Market Research & Brand Monitoring
- Product Feedback - Extract user reviews and opinions from product-related subreddits
- Competitor Analysis - Monitor what people say about competitors on Reddit
- Brand Mentions - Track brand sentiment across relevant subreddits
- Consumer Insights - Understand pain points and feature requests from community discussions
Content & Marketing
- Content Ideas - Find trending topics and popular discussions for content inspiration
- Community Research - Understand what resonates in specific communities before posting
- Influencer Discovery - Identify top contributors and thought leaders in niche subreddits
- SEO Research - Find questions people ask on Reddit for content gap analysis
Academic & Research
- Social Media Research - Study online community dynamics and discourse patterns
- Public Opinion Analysis - Analyze sentiment on political, social, or cultural topics
- Data Journalism - Extract structured data from Reddit for investigative reporting
- Linguistic Analysis - Study language patterns, slang evolution, and communication styles
API Usage
Use the Apify API to run this scraper programmatically:
curl "https://api.apify.com/v2/acts/george.the.developer~reddit-scraper-pro/runs" \-X POST \-H "Content-Type: application/json" \-H "Authorization: Bearer YOUR_API_TOKEN" \-d '{"urls": ["https://www.reddit.com/r/technology","best AI tools 2026"],"sort": "top","searchTime": "week","scrapeComments": true,"maxResults": 200}'
Python Example
from apify_client import ApifyClientclient = ApifyClient("YOUR_API_TOKEN")run = client.actor("george.the.developer/reddit-scraper-pro").call(run_input={"urls": ["https://www.reddit.com/r/technology", "AI news"],"sort": "top","searchTime": "week","scrapeComments": True,"maxResults": 100,})for item in client.dataset(run["defaultDatasetId"]).iterate_items():if item["type"] == "post":print(f"[{item['score']}] {item['title']}")elif item["type"] == "comment":print(f" > {item['text'][:100]}")
JavaScript / Node.js Example
import { ApifyClient } from 'apify-client';const client = new ApifyClient({ token: 'YOUR_API_TOKEN' });const run = await client.actor('george.the.developer/reddit-scraper-pro').call({urls: ['https://www.reddit.com/r/technology'],sort: 'top',searchTime: 'week',maxResults: 50,});const { items } = await client.dataset(run.defaultDatasetId).listItems();items.filter(i => i.type === 'post').forEach(post => {console.log(`[${post.score}] ${post.title}`);});
Integrations
Works with all Apify integrations:
- LangChain / LlamaIndex - Direct Apify integration for AI/RAG pipelines
- Zapier - Automate Reddit monitoring workflows
- Make (Integromat) - Build content research automations
- Google Sheets - Export Reddit data to spreadsheets
- Webhooks - Get notified when scraping completes
- REST API - Full programmatic access from any language
- Slack / Discord - Send Reddit updates to your channels
FAQ
How does this work without a Reddit API key?
Reddit serves JSON data at any URL by appending .json to the path. This scraper uses these public endpoints with smart rate limiting and proxy rotation. No OAuth tokens, no API application, no approval process needed.
What if I get rate limited? The scraper has built-in rate limiting (delays between requests) and automatic retry with exponential backoff for 429 responses. For large batches (1000+ results), enable Apify proxy to distribute requests across multiple IPs.
Can I search within a specific subreddit?
Yes! You can either use a subreddit search URL like https://www.reddit.com/r/python/search?q=web+scraping or combine a subreddit URL with the search functionality.
How many results can I get? Up to 5,000 results per run. Reddit's pagination allows roughly 1,000 posts per listing. For larger datasets, run multiple batches with different subreddits or time periods.
Does this work with quarantined or private subreddits? No. This scraper only accesses publicly available Reddit content. Quarantined, private, and banned subreddits are not accessible without authentication.
How deep can comment threads go? You can configure comment depth from 1 (top-level only) to 10 levels deep. The default is 2, which captures top-level comments and their direct replies.
Why is this better than the Reddit API? The Reddit API requires OAuth2 setup, application approval, and has strict rate limits (100 req/min free tier). This scraper handles all of that for you — just pass URLs and get data. Plus, you get structured output ready for AI tools.
Changelog
- v1.0 (March 2026) - Initial release. Subreddit scraping, post extraction, comment threading, user profiles, Reddit search, PPE pricing.
Support
Having issues? Open an issue on this actor's page. We typically respond within 24 hours.