Reddit Scraper | All-In-One | $1.8 / 1K avatar
Reddit Scraper | All-In-One | $1.8 / 1K

Pricing

$1.80 / 1,000 results

Go to Apify Store
Reddit Scraper | All-In-One | $1.8 / 1K

Reddit Scraper | All-In-One | $1.8 / 1K

Developed by

Fatih Tahta

Fatih Tahta

Maintained by Community

All-in-one Reddit Scraper. Scrape posts and full comment threads from any search, subreddit, user, or direct post URL. This enterprise-grade scraper is the fastest in the market and delivers clean and detailed JSON.

5.0 (2)

Pricing

$1.80 / 1,000 results

6

66

65

Issues response

3.4 hours

Last modified

3 days ago

Reddit Scraper Pro

Slug: fatihtahta/reddit-scraper Price: $1.80 per 1,000 saved items (posts or comments)

The all-in-one Reddit data solution. Go beyond simple search—scrape posts, full comment threads with a configurable limit, subreddits, and user pages with a single tool. Whether you provide search queries or a list of direct URLs, this scraper uses enterprise-grade residential proxies to bypass blocks and deliver clean, structured JSON, ready for any analysis.


What Can This Reddit Scraper Do?

  • Scrape Anything on Reddit: Provides two powerful modes:
  • Search Mode: Scrape search results for any query with advanced sorting and time filters.
  • URL Mode: Directly scrape one or more URLs, including subreddits, user pages, or individual posts.
  • Deep Comment Scraping (Optional): A simple switch (scrapeComments) allows you to extract not just the post, but also up to a specified number of comments from the discussion tree.
  • Include NSFW Content: A new option (includeNsfw) allows you to scrape content from posts tagged for adults (18+).
  • Enterprise-Grade Proxies: All requests are routed through a pool of residential proxies, ensuring high success rates and avoiding blocks.
  • Built-in Resiliency: Automatically retries failed requests with intelligent backoff, gracefully handling network errors and timeouts.
  • Clean, Structured JSON: Outputs two distinct item types (post and comment) with clear schemas, perfect for market research, social listening, brand monitoring, or academic analysis.
  • Fast and Efficient: Built with TypeScript and the latest Crawlee framework for high-performance, concurrent scraping.

How Do I Use the Reddit Scraper?

The scraper operates with a clear priority system:

  1. Check for URLs: If the urls input is provided, it will be used. Search queries will be ignored.
  2. Normalize URLs: Each URL is intelligently identified (as a post, subreddit, user page, or search URL) and converted to the correct API endpoint.
  3. Use Search Queries: If no URLs are provided, the scraper falls back to queries mode, building search requests based on your terms and filters.
  4. Crawl Listings: It crawls the resulting pages (search results, subreddit pages, etc.), following pagination cursors to find post links.
  5. Fetch Posts & Comments: For each post found, it fetches the full data payload.
  • If scrapeComments is true, it processes the entire comment tree, saving up to the maxComments limit for each post.
  • If scrapeComments is false, it only saves the post data and ignores comments, saving time and resources.
  1. Push to Dataset: All extracted posts and comments are pushed as clean JSON objects into the dataset.

What Input Does the Reddit Scraper Require?

  • queries (array of strings, optional): A list of search terms to look up on Reddit. This input is ignored if urls is provided.
  • urls (array of strings, optional): A list of specific Reddit URLs to scrape. This has priority over queries.
  • scrapeComments (boolean, default: false): If true, the scraper will extract comments from posts.
  • maxComments (integer, default: 100): The maximum number of comments to save for each post. Only applies when scrapeComments is true.
  • maxPosts (integer, default: 100): A hard limit on the number of posts to save for each individual search term or URL. Does not include comments.
  • includeNsfw (boolean, default: false): If true, includes NSFW (over 18) results.
  • sort (string, default: relevance): Sort order for search results (relevance, hot, top, new, comments).
  • timeframe (string, default: all): Time range for search results. This only applies when sorting by top, relevance, or comments.

Input and Output Examples

Example Input

This example will scrape a specific subreddit for up to 50 posts and a single post URL, limiting the number of comments saved to 100 for each.

{
"includeNsfw": false,
"queries": [
"Cheesecake",
"Swimming Pool"
],
"scrapeComments": true,
"sort": "hot",
"timeframe": "year",
"urls": [
"https://www.reddit.com/r/socialmedia/"
]
}

Output Example

The dataset will contain two types of items, distinguished by the kind field. You can download the data in various formats such as JSON, HTML, CSV, or Excel.

For the post record (kind: "post"):

{
"kind": "post",
"query": "[https://www.reddit.com/r/technology/](https://www.reddit.com/r/technology/)...",
"id": "1d95j4g",
"title": "The State of AI in 2025: A Comprehensive Report",
"body": "This report covers the latest advancements...",
"author": "tech_analyst",
"score": 2451,
"upvote_ratio": 0.95,
"num_comments": 873,
"subreddit": "technology",
"created_utc": "2025-08-05T18:00:00.000Z",
"url": "[https://www.reddit.com/r/technology/comments/1d95j4g/](https://www.reddit.com/r/technology/comments/1d95j4g/)"
}

For the comment record (kind: "comment"):

{
"kind": "comment",
"query": "[https://www.reddit.com/r/technology/](https://www.reddit.com/r/technology/)...",
"id": "k5z1x2y",
"postId": "t3_1d95j4g",
"parentId": "t3_1d95j4g",
"body": "Great analysis, but I think you're underestimating the impact of quantum computing on these timelines.",
"author": "future_thinker",
"score": 142,
"created_utc": "2025-08-05T19:15:22.000Z",
"url": "[https://www.reddit.com/r/technology/comments/1d95j4g/the_state_of_ai_in_2025_a_comprehensive_report/k5z1x2y/](https://www.reddit.com/r/technology/comments/1d95j4g/the_state_of_ai_in_2025_a_comprehensive_report/k5z1x2y/)"
}

How Much Will It Cost to Scrape Reddit?

The Actor is priced at $1.80 per 1,000 saved items (posts or comments).

All infrastructure and residential proxy costs are bundled in. You only pay for successful results. This transparent pricing means you can easily estimate the cost of a run: scraping 10,000 posts and 25,000 comments would cost approximately (35,000 / 1,000) * $1.80 = $63.00.

Tips for an Efficient Scrape

Save Costs: To keep your costs down, only enable scrapeComments when you absolutely need the discussion data. Scraping only posts is much faster and cheaper.

Targeted Scraping: Use the URL Mode for scraping specific subreddits or posts to avoid unnecessary searches and get exactly the data you need.

Our scrapers are ethical and do not extract any private user data. They only extract data that is publicly available on Reddit. We believe that our scrapers, when used for ethical purposes, are safe and legal. However, you should be aware that your results could contain personal data. You should not scrape personal data unless you have a legitimate reason to do so. If you're unsure, consult your lawyers. You can also read our blog post on the legality of web scraping.

Support

Questions or custom needs? Open an issue on the Issues tab in Apify Console, and it will be solved around the clock.

Happy Scrapings! Fatih