Reddit Scraper Pro - Posts, Comments & Subreddits - No API Key avatar

Reddit Scraper Pro - Posts, Comments & Subreddits - No API Key

Pricing

Pay per usage

Go to Apify Store
Reddit Scraper Pro - Posts, Comments & Subreddits - No API Key

Reddit Scraper Pro - Posts, Comments & Subreddits - No API Key

Scrape Reddit posts, comments, subreddits, and user profiles at scale. No Reddit API key needed. Search Reddit by keyword, extract full comment threads, monitor subreddits. AI-ready JSON output.

Pricing

Pay per usage

Rating

0.0

(0)

Developer

George Kioko

George Kioko

Maintained by Community

Actor stats

0

Bookmarked

2

Total users

1

Monthly active users

a day ago

Last modified

Share

Reddit Scraper Pro - Posts, Comments & Subreddits - No API Key Required

Scrape Reddit posts, comments, user profiles, and entire subreddits at scale. No Reddit API key needed. No OAuth, no rate limit hassles, no application approval. Works with subreddit URLs, post URLs, user profiles, and keyword search.

Perfect for AI/LLM pipelines, market research, brand monitoring, sentiment analysis, content strategy, and academic research. Output is clean JSON optimized for ChatGPT, Claude, Gemini, LangChain, LlamaIndex, and other AI tools.

Why use this Reddit Scraper?

FeatureThis ScraperReddit API (free)Manual Browsing
API key needed?NoYes (OAuth2)No
Approval process?NoneApplication requiredNone
Rate limits?Auto-managed100 req/minN/A
Posts + Comments?YesYesOne at a time
User profiles?YesLimitedOne at a time
Batch subreddits?YesPagination headachesNo
Search across Reddit?YesBasicManual
AI-ready output?YesRaw JSONNo
Cost per 1,000 results$2Free (with limits)Hours of work

What data do you get?

Reddit Post Data

  • Title, author, subreddit - Full post metadata
  • Full text (self posts) - Complete selftext for text-based posts
  • Score, upvote ratio - Engagement metrics for analysis
  • Comment count - Know the discussion size
  • Flair and awards - Post categorization and recognition
  • Media detection - Images, videos, galleries, embeds identified
  • Permalink and URL - Direct links to post and external content
  • Timestamps - Created date in ISO format

Reddit Comment Data

  • Full comment text - Including HTML formatting
  • Author and score - Who said it, how it was received
  • Thread depth - Nested reply structure preserved
  • Parent ID - Reconstruct conversation threads
  • Submitter flag - Know when OP replies

Reddit User Profile Data

  • Username and display name
  • Total karma (post + comment + awardee + awarder breakdown)
  • Account age
  • Moderator and verified status
  • Profile description and avatar

Key Features

  • No API Key Required - Scrapes Reddit's public JSON endpoints directly. Zero setup. No OAuth flow. No approved application needed.
  • Subreddit Scraping - Extract all posts from any subreddit with pagination. Sort by hot, new, top, rising, or controversial.
  • Reddit Search - Search across all of Reddit or within a specific subreddit. Sort by relevance, hot, new, top, or most comments.
  • Full Comment Threads - Extract comments with configurable depth. Get top-level comments only or dive deep into reply chains.
  • User Profile Scraping - Get user profile info, karma breakdown, and their submitted posts.
  • NSFW Filtering - Include or exclude NSFW content with a single toggle.
  • Smart Rate Limiting - Built-in request throttling and retry logic. Handles 429s gracefully with exponential backoff.
  • Proxy Support - Use Apify residential proxies for large batches to avoid IP-based rate limits.
  • Pay Per Result - Only pay for data actually extracted. No flat monthly fee. $0.002 per result.

Input Configuration

FieldTypeRequiredDescription
urlsArrayYesReddit URLs (subreddit, post, user) or plain text search queries
sortEnumNoSubreddit sort: "hot", "new", "top", "rising", "controversial". Default: "hot"
searchModeEnumNoSearch sort: "relevance", "hot", "new", "top", "comments". Default: "relevance"
searchTimeEnumNoTime filter: "hour", "day", "week", "month", "year", "all". Default: "all"
scrapeCommentsBooleanNoAlso extract comments from each post. Default: false
maxCommentsIntegerNoMax top-level comments per post, 1-500. Default: 50
commentDepthIntegerNoReply nesting depth, 1-10. Default: 2
maxResultsIntegerNoMaximum total results, 1-5000. Default: 100
includeNSFWBooleanNoInclude NSFW posts. Default: false
proxyConfigurationObjectNoApify proxy settings. Recommended for large batches.

Supported URL Formats

# Subreddits
https://www.reddit.com/r/technology
https://old.reddit.com/r/python
https://www.reddit.com/r/MachineLearning/top
r/datascience
# Individual posts
https://www.reddit.com/r/technology/comments/abc123/some_post_title/
https://old.reddit.com/r/python/comments/xyz789/discussion_thread/
# User profiles
https://www.reddit.com/user/spez
https://www.reddit.com/u/some_username
u/username
# Search queries (plain text)
best AI tools 2026
python web scraping tutorial
machine learning beginner

Output Examples

Post Output

{
"type": "post",
"id": "1abc234",
"title": "GPT-5 just dropped and it's incredible",
"author": "ai_enthusiast",
"subreddit": "technology",
"text": "Full self-text content here...",
"url": "https://www.reddit.com/r/technology/comments/1abc234/gpt5_just_dropped/",
"permalink": "https://www.reddit.com/r/technology/comments/1abc234/gpt5_just_dropped/",
"score": 15432,
"upvoteRatio": 0.94,
"numComments": 2341,
"isNSFW": false,
"flair": "AI",
"awards": 23,
"media": null,
"createdUtc": "2026-03-01T15:30:00.000Z",
"scrapedAt": "2026-03-02T21:34:36.000Z"
}

Comment Output

{
"type": "comment",
"id": "k5m6n7o",
"postId": "1abc234",
"author": "tech_reviewer",
"subreddit": "technology",
"text": "I've been testing it for 3 hours and the reasoning capability is insane.",
"score": 892,
"depth": 0,
"isSubmitter": false,
"createdUtc": "2026-03-01T16:15:00.000Z",
"scrapedAt": "2026-03-02T21:34:37.000Z"
}

User Profile Output

{
"type": "user",
"username": "spez",
"displayName": "u/spez",
"totalKarma": 1234567,
"postKarma": 456789,
"commentKarma": 678901,
"isMod": true,
"isVerified": true,
"accountCreatedUtc": "2005-06-06T00:00:00.000Z",
"scrapedAt": "2026-03-02T21:34:38.000Z"
}

Pricing

Pay Per Event pricing - only pay for successful results.

EventPriceDescription
Actor start$0.005One-time fee per run
Result scraped$0.002Per post, comment, or user profile extracted

$2 per 1,000 results. Compare to Reddit API development time (days of OAuth setup) or manual data collection (hours per subreddit).

Use Cases

AI & Machine Learning

  • Training Data - Build datasets from specific subreddits for fine-tuning LLMs on domain-specific content
  • RAG Pipelines - Feed Reddit discussions into vector databases for retrieval-augmented generation
  • Sentiment Analysis - Analyze public opinion on products, brands, or topics at scale
  • Trend Detection - Monitor emerging topics across subreddits using keyword search

Market Research & Brand Monitoring

  • Product Feedback - Extract user reviews and opinions from product-related subreddits
  • Competitor Analysis - Monitor what people say about competitors on Reddit
  • Brand Mentions - Track brand sentiment across relevant subreddits
  • Consumer Insights - Understand pain points and feature requests from community discussions

Content & Marketing

  • Content Ideas - Find trending topics and popular discussions for content inspiration
  • Community Research - Understand what resonates in specific communities before posting
  • Influencer Discovery - Identify top contributors and thought leaders in niche subreddits
  • SEO Research - Find questions people ask on Reddit for content gap analysis

Academic & Research

  • Social Media Research - Study online community dynamics and discourse patterns
  • Public Opinion Analysis - Analyze sentiment on political, social, or cultural topics
  • Data Journalism - Extract structured data from Reddit for investigative reporting
  • Linguistic Analysis - Study language patterns, slang evolution, and communication styles

API Usage

Use the Apify API to run this scraper programmatically:

curl "https://api.apify.com/v2/acts/george.the.developer~reddit-scraper-pro/runs" \
-X POST \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_TOKEN" \
-d '{
"urls": [
"https://www.reddit.com/r/technology",
"best AI tools 2026"
],
"sort": "top",
"searchTime": "week",
"scrapeComments": true,
"maxResults": 200
}'

Python Example

from apify_client import ApifyClient
client = ApifyClient("YOUR_API_TOKEN")
run = client.actor("george.the.developer/reddit-scraper-pro").call(run_input={
"urls": ["https://www.reddit.com/r/technology", "AI news"],
"sort": "top",
"searchTime": "week",
"scrapeComments": True,
"maxResults": 100,
})
for item in client.dataset(run["defaultDatasetId"]).iterate_items():
if item["type"] == "post":
print(f"[{item['score']}] {item['title']}")
elif item["type"] == "comment":
print(f" > {item['text'][:100]}")

JavaScript / Node.js Example

import { ApifyClient } from 'apify-client';
const client = new ApifyClient({ token: 'YOUR_API_TOKEN' });
const run = await client.actor('george.the.developer/reddit-scraper-pro').call({
urls: ['https://www.reddit.com/r/technology'],
sort: 'top',
searchTime: 'week',
maxResults: 50,
});
const { items } = await client.dataset(run.defaultDatasetId).listItems();
items.filter(i => i.type === 'post').forEach(post => {
console.log(`[${post.score}] ${post.title}`);
});

Integrations

Works with all Apify integrations:

  • LangChain / LlamaIndex - Direct Apify integration for AI/RAG pipelines
  • Zapier - Automate Reddit monitoring workflows
  • Make (Integromat) - Build content research automations
  • Google Sheets - Export Reddit data to spreadsheets
  • Webhooks - Get notified when scraping completes
  • REST API - Full programmatic access from any language
  • Slack / Discord - Send Reddit updates to your channels

FAQ

How does this work without a Reddit API key? Reddit serves JSON data at any URL by appending .json to the path. This scraper uses these public endpoints with smart rate limiting and proxy rotation. No OAuth tokens, no API application, no approval process needed.

What if I get rate limited? The scraper has built-in rate limiting (delays between requests) and automatic retry with exponential backoff for 429 responses. For large batches (1000+ results), enable Apify proxy to distribute requests across multiple IPs.

Can I search within a specific subreddit? Yes! You can either use a subreddit search URL like https://www.reddit.com/r/python/search?q=web+scraping or combine a subreddit URL with the search functionality.

How many results can I get? Up to 5,000 results per run. Reddit's pagination allows roughly 1,000 posts per listing. For larger datasets, run multiple batches with different subreddits or time periods.

Does this work with quarantined or private subreddits? No. This scraper only accesses publicly available Reddit content. Quarantined, private, and banned subreddits are not accessible without authentication.

How deep can comment threads go? You can configure comment depth from 1 (top-level only) to 10 levels deep. The default is 2, which captures top-level comments and their direct replies.

Why is this better than the Reddit API? The Reddit API requires OAuth2 setup, application approval, and has strict rate limits (100 req/min free tier). This scraper handles all of that for you — just pass URLs and get data. Plus, you get structured output ready for AI tools.

Changelog

  • v1.0 (March 2026) - Initial release. Subreddit scraping, post extraction, comment threading, user profiles, Reddit search, PPE pricing.

Support

Having issues? Open an issue on this actor's page. We typically respond within 24 hours.