Reddit Scraper avatar
Reddit Scraper

Pricing

$19.99/month + usage

Go to Apify Store
Reddit Scraper

Reddit Scraper

Scrape Reddit posts, comments, communities, and users without login. Extract data from subreddits, search results, user profiles. Sort by hot/new/top, filter by date, include/exclude NSFW. Keyword search, residential proxies, fast and reliable.

Pricing

$19.99/month + usage

Rating

0.0

(0)

Developer

SilentFlow

SilentFlow

Maintained by Community

Actor stats

0

Bookmarked

2

Total users

1

Monthly active users

a day ago

Last modified

Categories

Share

by SilentFlow

Unlimited Reddit web scraper to crawl posts, comments, communities, and users without login. Extract all data from subreddits, search results, and user profiles.

Why use this scraper?

  • No login required: Scrape all public Reddit data without authentication
  • Comprehensive data: Posts, comments, communities, and users with full metadata
  • Flexible input: Scrape by URL or search by keyword
  • Advanced filtering: Sort by hot/new/top, filter by date, include/exclude NSFW
  • High reliability: Built-in retry logic and residential proxy support

Use cases

IndustryApplication
Market researchMonitor brand mentions and sentiment across subreddits
Content analysisAnalyze trending topics and community discussions
Academic researchStudy online communities and user behavior
Competitive intelligenceTrack competitor discussions and product feedback
Trend monitoringIdentify emerging trends and opinions

Input parameters

URL scraping

ParameterTypeDescription
startUrlsarrayReddit URL(s) to scrape (subreddits, posts, users, search pages)

Supported URL types:

  • Subreddits: https://www.reddit.com/r/programming/
  • Subreddit channels: https://www.reddit.com/r/programming/hot
  • Posts: https://www.reddit.com/r/learnprogramming/comments/abc123/...
  • Users: https://www.reddit.com/user/username
  • User comments: https://www.reddit.com/user/username/comments/
  • Search: https://www.reddit.com/search/?q=keyword
  • Popular: https://www.reddit.com/r/popular/
  • Leaderboards: https://www.reddit.com/subreddits/leaderboard/crypto/
ParameterTypeDescription
searchesarrayKeywords to search on Reddit
searchCommunityNamestringRestrict search to a specific community
searchPostsbooleanSearch for posts (default: true)
searchCommentsbooleanSearch for comments (default: false)
searchCommunitiesbooleanSearch for communities (default: false)
searchUsersbooleanSearch for users (default: false)

Sorting & filtering

ParameterTypeDefaultOptions
sortstringnewrelevance, hot, top, new, rising, comments
timestringallall, hour, day, week, month, year
includeNSFWbooleantrueInclude NSFW content
postDateLimitstring-Only posts after this date (YYYY-MM-DD)

Limits

ParameterTypeDefaultDescription
maxItemsinteger10Maximum total items to save
maxPostCountinteger10Maximum posts per page/subreddit
maxCommentsinteger10Maximum comments per post
maxCommunitiesCountinteger2Maximum communities to scrape
maxUserCountinteger2Maximum users to scrape

Skip options

ParameterTypeDefaultDescription
skipCommentsbooleanfalseSkip comment extraction
skipUserPostsbooleanfalseSkip user post extraction
skipCommunitybooleanfalseSkip community info extraction

Advanced

ParameterTypeDefaultDescription
scrollTimeoutinteger40Request timeout in seconds
debugModebooleanfalseEnable detailed logging
proxyobjectresidentialProxy configuration (useApifyProxy, apifyProxyGroups)

Output data

Post example

{
"id": "t3_abc123",
"parsedId": "abc123",
"url": "https://www.reddit.com/r/programming/comments/abc123/example_post/",
"username": "dev_user",
"userId": "t2_abc123",
"title": "Example Post Title",
"communityName": "r/programming",
"parsedCommunityName": "programming",
"body": "Post body text...",
"html": null,
"numberOfComments": 42,
"upVotes": 256,
"upVoteRatio": 0.95,
"isVideo": false,
"isAd": false,
"over18": false,
"flair": "Discussion",
"link": "https://example.com/article",
"thumbnailUrl": "https://b.thumbs.redditmedia.com/...",
"videoUrl": "",
"imageUrls": ["https://i.redd.it/abc123.jpg"],
"createdAt": "2024-06-01T12:00:00Z",
"scrapedAt": "2024-06-02T10:30:00Z",
"dataType": "post"
}

Comment example

{
"id": "t1_xyz789",
"parsedId": "xyz789",
"url": "https://www.reddit.com/r/programming/comments/abc123/example_post/xyz789/",
"parentId": "t3_abc123",
"postId": "abc123",
"username": "commenter",
"userId": "t2_xyz789",
"category": "programming",
"communityName": "r/programming",
"body": "Great post!",
"html": "<div class=\"md\"><p>Great post!</p></div>",
"createdAt": "2024-06-01T13:00:00Z",
"scrapedAt": "2024-06-02T10:30:00Z",
"upVotes": 15,
"numberOfreplies": 3,
"dataType": "comment"
}

Community example

{
"id": "2fwo",
"name": "t5_2fwo",
"title": "Programming",
"url": "https://www.reddit.com/r/programming/",
"description": "Computer programming",
"over18": false,
"numberOfMembers": 5800000,
"createdAt": "2006-01-25T00:00:00Z",
"scrapedAt": "2024-06-02T10:30:00Z",
"dataType": "community"
}

User example

{
"id": "abc123",
"url": "https://www.reddit.com/user/dev_user/",
"username": "dev_user",
"description": "Software engineer and open source enthusiast",
"postKarma": 15000,
"commentKarma": 42000,
"over18": false,
"createdAt": "2020-01-15T00:00:00Z",
"scrapedAt": "2024-06-02T10:30:00Z",
"dataType": "user"
}

Field descriptions

FieldDescription
userIdReddit internal user ID (author_fullname)
htmlRaw HTML body (null for posts, populated for comments)
flairPost flair/tag (null if none)
linkExternal link attached to the post (empty if self-post)
thumbnailUrlReddit-generated thumbnail URL
videoUrlReddit video URL (v.redd.it) if the post is a video
imageUrlsArray of image URLs (empty array if no images)
isAdWhether the post is a promoted/sponsored post
postIdParsed post ID the comment belongs to
categoryCommunity name (parsed) for the comment
numberOfrepliesDirect reply count for the comment

Data fields

CategoryFields
Identityid, parsedId, url, username, userId
Contenttitle, body, html, flair
CommunitycommunityName, parsedCommunityName, category
EngagementupVotes, upVoteRatio, numberOfComments, numberOfreplies
MediaimageUrls, videoUrl, thumbnailUrl, link
FlagsisVideo, isAd, over18
MetacreatedAt, scrapedAt, dataType

Examples

Scrape a subreddit

{
"startUrls": [{"url": "https://www.reddit.com/r/programming/"}],
"maxItems": 50,
"maxPostCount": 20,
"maxComments": 10,
"sort": "hot"
}

Search for a keyword

{
"searches": ["machine learning"],
"searchPosts": true,
"searchCommunities": true,
"sort": "top",
"time": "month",
"maxItems": 100
}

Scrape a specific post with comments

{
"startUrls": [{"url": "https://www.reddit.com/r/learnprogramming/comments/lp1hi4/is_webscraping_a_good_skill_to_learn/"}],
"maxComments": 50
}

Search within a community

{
"searches": ["python"],
"searchCommunityName": "programming",
"searchPosts": true,
"sort": "new",
"maxItems": 50
}

Integrations

Python

from apify_client import ApifyClient
client = ApifyClient("YOUR_API_TOKEN")
run = client.actor("silentflow/reddit-scraper").call(run_input={
"startUrls": [{"url": "https://www.reddit.com/r/programming/"}],
"maxItems": 50,
"maxPostCount": 20,
"maxComments": 10,
"sort": "hot"
})
for item in client.dataset(run["defaultDatasetId"]).iterate_items():
if item["dataType"] == "post":
print(f"[{item['upVotes']}] {item['title']}")
elif item["dataType"] == "comment":
print(f" > {item['body'][:80]}")

JavaScript

import { ApifyClient } from 'apify-client';
const client = new ApifyClient({ token: 'YOUR_API_TOKEN' });
const run = await client.actor('silentflow/reddit-scraper').call({
searches: ['web scraping'],
searchPosts: true,
sort: 'top',
time: 'week',
maxItems: 100
});
const { items } = await client.dataset(run.defaultDatasetId).listItems();
items.forEach(item => {
if (item.dataType === 'post') {
console.log(`[${item.upVotes}] ${item.title}`);
}
});

Performance & limits

MetricValue
Items per requestup to 100
Average speed~50 items/second
Max items per run10,000
Supported contentPosts, Comments, Communities, Users

Tips for best results

  1. Use specific subreddits: Target specific communities for focused data
  2. Set realistic limits: Start with maxItems: 10 to test before large scrapes
  3. Use date filters: Combine postDateLimit with sort "new" for recent content
  4. Residential proxy: Enabled by default for best reliability
  5. Skip what you don't need: Use skipComments to speed up subreddit scraping

FAQ

Q: Can I scrape private subreddits? A: No, this scraper only accesses publicly available data.

Q: Why are some posts missing? A: Reddit may filter certain posts. NSFW content is included by default but can be toggled.

Q: How often can I run the scraper? A: No limits on frequency. Use residential proxies for best results.

Q: What happens if Reddit blocks the scraper? A: The scraper automatically rotates proxies and retries. If all attempts fail, try again later.

Support

Need help? We're here for you:

  • Bug reports: Open an issue on the actor page
  • Questions: Message us via Apify console
  • Feature requests: Let us know what you need
  • Custom solutions: Contact us for enterprise integrations or high-volume needs

Check out our other scrapers: SilentFlow on Apify