Reddit Subreddit Posts Scraper - No API Key
Pricing
Pay per event
Reddit Subreddit Posts Scraper - No API Key
Scrape any public subreddit. Posts, scores, comments, authors, awards, flairs, timestamps. No API key, no OAuth, no login. Free public Reddit JSON. For alt data, social listening, AI training datasets.
Pricing
Pay per event
Rating
0.0
(0)
Developer
Mohieldin Mohamed
Actor stats
0
Bookmarked
1
Total users
0
Monthly active users
5 days ago
Last modified
Categories
Share
Reddit Subreddit Posts Scraper
Scrape any public subreddit for posts, scores, comments, authors, and metadata. No API key. No OAuth. No login. 100% free public Reddit JSON endpoints.
This actor pulls structured post data from any public subreddit using Reddit's official free JSON endpoints — no authentication required. Perfect for alt data, social listening, sentiment analysis, market research, AI training datasets, and content trend tracking.
What does Reddit Subreddit Scraper do?
You give it a list of subreddit names (e.g. wallstreetbets, webdev, saas, localllama, aitools). It pulls posts ranked by hot, new, top, rising, or controversial — paginating automatically up to your max — and returns each post as a clean structured row with:
- Title, author, score, upvote ratio, comment count
- Created timestamp (ISO format)
- URL (the linked content) and permalink (the Reddit thread)
- Linked domain (useful for tracking which sites get traction)
- Flair, awards, gilded count
- Body text for self-posts
- Thumbnail, media type
- Flags: stickied, locked, NSFW, video, self-post
Try it: leave the defaults (r/wallstreetbets, hot, top 50), press Start, and watch the dataset fill with the current top WSB stock-picking and meme posts in seconds.
Why use Reddit Subreddit Scraper?
Reddit is one of the most-cited alternative data sources for hedge funds and quant traders. The 2025 Nasdaq State of Alt Data report listed Reddit sentiment as a top-5 institutional alt data source. Beyond finance, Reddit is the backbone for:
- Hedge funds + quant traders — track r/wallstreetbets sentiment shifts before they hit price
- Marketing teams — find which posts and content domains are trending in your niche
- Founders + indie hackers — track r/SaaS, r/IndieHackers, r/Entrepreneur for market signals
- AI/ML researchers — collect training data for sentiment, summarization, instruction-following models (Reddit was famously sold to OpenAI for $60M as training data)
- Content marketers — find trending topics to write about
- Social listeners — monitor brand mentions, product feedback, complaints
- Journalists — break stories that emerge from niche communities first
This actor is dramatically simpler and cheaper than commercial alternatives like Brandwatch ($800+/month), Sprout Social ($249/month), or PRAW + custom infrastructure. It's also faster than the official Reddit API (which requires OAuth, app registration, and rate limit headaches).
How to use
- Click Try for free (or Start)
- Paste subreddit names into Subreddits (without the
r/prefix) - Pick the sort order (hot / new / top / rising / controversial)
- Set max posts per subreddit (default 50, max 1000)
- Click Start
- Download as JSON, CSV, HTML, or Excel — or schedule daily runs
Input
- Subreddits — list of subreddit names (e.g.
["wallstreetbets", "webdev", "saas"]) - Sort order —
hot/new/top/rising/controversial(default:hot) - Time filter —
hour/day/week/month/year/all(only fortopandcontroversial, default:day) - Max posts per subreddit — cap (default 50, max 1000)
- Include text — pull
selftextfor self-posts (default: yes) - Min score — filter out posts below this score (default: 0)
- Extract domains — parse linked URLs to get the domain (default: yes)
- Proxy configuration — optional, recommended for high-volume runs
Output
{"subreddit": "wallstreetbets","title": "NVDA earnings preview - what to expect","author": "stocksavvy","score": 1234,"upvoteRatio": 0.92,"numComments": 387,"createdAt": "2026-04-15T14:30:00.000Z","url": "https://example.com/nvda-earnings-preview","permalink": "https://www.reddit.com/r/wallstreetbets/comments/abc123/nvda_earnings_preview/","linkDomain": "example.com","flair": "DD","isVideo": false,"isSelfPost": false,"isOver18": false,"selftext": null,"thumbnailUrl": "https://b.thumbs.redditmedia.com/...jpg","mediaType": "link","awardsCount": 5,"gilded": 2,"stickied": false,"locked": false,"edited": false,"domain": "example.com","id": "abc123","extractedAt": "2026-04-15T19:00:00.000Z"}
Data table
| Field | Type | Description |
|---|---|---|
subreddit | string | Subreddit name |
title | string | Post title |
author | string | Reddit username (or null for deleted accounts) |
score | number | Upvotes minus downvotes |
upvoteRatio | number | 0.0–1.0 ratio of upvotes to total votes |
numComments | number | Comment count |
createdAt | string | ISO timestamp |
url | string | Linked URL |
permalink | string | Direct link to the Reddit thread |
linkDomain | string | Domain of the linked URL (e.g. example.com) |
flair | string | Post flair (e.g. "DD", "Discussion", "Meme") |
awardsCount | number | Total Reddit awards received |
gilded | number | Number of gold awards |
stickied | boolean | Pinned to the top of the subreddit? |
locked | boolean | Comments locked? |
isVideo | boolean | Reddit-hosted video? |
isSelfPost | boolean | Text-only self-post? |
isOver18 | boolean | NSFW flag |
selftext | string | Body text (for self-posts) |
thumbnailUrl | string | Thumbnail image URL |
mediaType | string | Reddit's post_hint (link, image, video, etc.) |
id | string | Reddit post ID |
extractedAt | string | When this scrape happened |
Pricing
This actor uses Apify's pay-per-event pricing — extremely cheap for both small spot-checks and bulk historical pulls:
- Actor start: $0.01 per run
- Per post extracted: $0.005 per post
Example costs:
- Daily snapshot of r/wallstreetbets top 50 → $0.26/day = $7.85/month
- Hourly check across 10 niche subreddits (50 posts each) → ~$15/month
- Bulk historical pull of 1000 posts from one subreddit → $5.01
- Track 100 subreddits × 50 posts daily → ~$15/day = $450/month
Compare to Brandwatch ($800+/month minimum), Sprout Social ($249/month), or building your own PRAW pipeline (which requires OAuth, rate limit handling, and infrastructure).
Free Apify tier members get $5/month in platform credits, which covers ~1,000 posts per month.
Tips and advanced options
- Schedule hourly runs during market hours to track r/wallstreetbets / r/stocks sentiment changes in near-real-time
- Use
sort: top+timeFilter: hourto catch breaking trending content - Use
sort: new+ frequent runs to catch every post the moment it's submitted (great for keyword alerts) - Track multiple sub-niches at once — pass an array of related subreddits (e.g.
["aitools", "localllama", "ChatGPT", "ClaudeAI"]) for a complete AI tools ecosystem snapshot - Filter by minScore to only get posts that are gaining traction — eliminates the long tail noise
- Pipe into a quant model to test Reddit sentiment as an alpha factor
- Combine with the SEC EDGAR Filing Monitor to correlate r/wallstreetbets buzz with SEC filings
- Combine with the Hiring Signal Tracker to triangulate company growth signals across multiple data sources
FAQ and support
Do I need a Reddit API key or OAuth credentials? No. This actor uses Reddit's free public JSON endpoints (https://reddit.com/r/{subreddit}/{sort}.json), which require no authentication. We don't even ask for your Reddit username.
What about rate limits? Reddit's public JSON endpoints allow ~60 requests per minute per IP. For large jobs (1000+ posts across many subreddits), enable Apify Proxy in the input to rotate IPs and avoid rate limits.
Does this work for private or quarantined subreddits? No. Only public subreddits accessible without login are supported.
How does this compare to PRAW or the official Reddit API? PRAW is excellent for Python developers but requires OAuth setup and rate limit handling. The official Reddit API has stricter rate limits and requires app registration. This actor is the simplest path to clean Reddit data without any of that hassle.
Is this legal? Yes. Reddit's public JSON endpoints are explicitly designed for programmatic access. We respect rate limits and identify ourselves with a clear User-Agent header.
Found a bug? Open an issue on the Issues tab.