Reddit Bulk Scrape 10000 IDs V1 — Posts, Comments, Subs, Users avatar

Reddit Bulk Scrape 10000 IDs V1 — Posts, Comments, Subs, Users

Pricing

from $1.99 / 1,000 results

Go to Apify Store
Reddit Bulk Scrape 10000 IDs V1 — Posts, Comments, Subs, Users

Reddit Bulk Scrape 10000 IDs V1 — Posts, Comments, Subs, Users

Bulk-hydrate up to 10,000 Reddit posts, comments, subreddits, or users per run. Paste IDs, names, or URLs — get one full record per item. No Reddit account, OAuth, or proxy required. Mix formats freely; duplicates auto-removed. $1.99 per 1,000 results.

Pricing

from $1.99 / 1,000 results

Rating

0.0

(0)

Developer

Red Crawler

Red Crawler

Maintained by Community

Actor stats

0

Bookmarked

4

Total users

3

Monthly active users

a day ago

Last modified

Share

Reddit Bulk Scrape

Endpoints Auth Proxy Pricing Cap

Hydrate up to 10000 Reddit IDs, names, or URLs in a single run. Pick an endpoint — Posts, Comments, Subreddits, or Users — paste your list, and get one fully populated dataset record per item. No Reddit account, OAuth, or proxy required.

A single-input actor: choose the endpoint, paste the list, hit Start.


Endpoints at a glance

#EndpointInputCap per runBest for
1Bulk Postspost IDs / URLs10000refreshing stored post lists, hydrating IDs / URLs
2Bulk Commentscomment IDs / URLs10000comment-list hydration, archival pipelines
3Bulk Subredditssubreddit names / IDs / URLs10000community-list enrichment, niche directories
4Bulk Usersusernames / IDs / URLs10000CRM enrichment, account-quality scoring

Inputs accept the most-permissive format Reddit uses for each entity:

EntityAccepted formats
postfull URL · prefixed t3_1s4a4j6 · stripped ID 1s4a4j6 · short URL https://reddit.com/comments/1s4a4j6
commentfull URL · prefixed t1_lwbnv0t · stripped ID lwbnv0t
subredditname AskReddit · prefixed r/AskReddit · full ID t5_2qh1i · full URL https://reddit.com/r/AskReddit
userusername spez · prefixed u/spez · full ID t2_1w72 · profile URL https://reddit.com/user/spez

Separate inputs with commas or newlines — both work. Mix prefixed, stripped, names, and URLs freely; duplicates are removed automatically.


What you can fetch

1. Bulk Posts

Hydrate a list of Reddit posts.

Inputs

FieldNotes
bulk_posts_idsComma- or newline-separated post inputs. Up to 10000.

Accepted formats — full IDs (t3_1s4a4j6), stripped IDs (1s4a4j6), full URLs (https://www.reddit.com/r/Wordpress/comments/1s4a4j6/), and short URLs (https://reddit.com/comments/1s4a4j6). Mix freely.

Returns per post — Reddit ID, fullname, title, body / selftext, author, author fullname, subreddit (name + prefixed + ID), score, ups / downs, upvote ratio, comment count, crosspost count, created + edited timestamps, permalink, external URL, domain, post-type flags (is_self, is_video, over_18, spoiler, locked, stickied, pinned, archived), distinguished status, removal category, link & author flair, thumbnail, media (images / video / gallery), awards, polls, crosspost source.

Use it when — refreshing a stored dataset (yesterday's IDs → today's score / comment counts / edits / deletions), turning a list of links into structured records, hydrating IDs from your own search results.

Note: Posts are always SFW in this actor. NSFW (over-18) posts are not returned. There is no toggle. The SFW lock applies to posts only — comments, subreddits, and users are returned as-is regardless of any age-gating on the parent post or community.


2. Bulk Comments

Hydrate a list of Reddit comments.

Inputs

FieldNotes
bulk_comments_idsComma- or newline-separated comment inputs. Up to 10000.

Accepted formats — full IDs (t1_lwbnv0t), stripped IDs (lwbnv0t), and full URLs (https://www.reddit.com/r/Wordpress/comments/1s4a4j6/comment/lwbnv0t/). Mix freely.

Returns per comment — ID, fullname, parent post fullname, parent comment ID, author + author fullname, body (markdown + HTML), score / ups / downs / controversiality, created + edited timestamps, permalink, OP flag (is_submitter), depth, stickied / distinguished / locked / archived / saved / gilded flags, score-hidden flag, subreddit, awards.

Use it when — hydrating comment IDs from your own pipelines, comment archives, sentiment analysis on a known set of comments.


3. Bulk Subreddits

Hydrate a list of subreddits.

Inputs

FieldNotes
bulk_subreddits_idsComma- or newline-separated subreddit inputs. Up to 10000.

Accepted formats — subreddit names (AskReddit), prefixed names (r/AskReddit), full IDs (t5_2qh1i), and full URLs (https://reddit.com/r/AskReddit). Mix freely.

Returns per subreddit — ID, fullname, display name (raw + prefixed), title, subscriber count, active user count, public + full description, created timestamp, language, type (public / private / restricted), NSFW flag, URL, header / icon / banner images, primary + key colors, submit text, allowed submission types (videos / images / polls / galleries).

Use it when — subreddit comparison reports, community sizing, profile-page enrichment, building niche directories.


4. Bulk Users

Hydrate a list of Reddit users.

Inputs

FieldNotes
bulk_users_idsComma- or newline-separated user inputs. Up to 10000.

Accepted formats — usernames (spez), prefixed names (u/spez), full IDs (t2_1w72), and profile URLs (https://reddit.com/user/spez). Mix freely.

Returns per user — ID, name, total karma split into post / comment / award / awardee karma, account creation timestamp, employee / mod / Reddit-Gold / verified / verified-email flags, profile icon, snoovatar image, mini subreddit info, accept-followers flag, hide-from-robots flag.

Use it when — CRM / lead enrichment from a list of usernames, account-quality scoring, finding which accounts are still alive, batch profile lookups for influencer research.


How to run

  1. Pick an endpoint in the "What to fetch" dropdown — Bulk Posts, Bulk Comments, Bulk Subreddits, or Bulk Users.
  2. Open the matching section and paste your IDs / names / URLs (comma- or newline-separated). Up to 10000 entries per run — duplicates are removed automatically.
  3. Click Start.

Default endpoint is Bulk Posts with a small prefilled list so the actor runs out of the box.


Output

Results are pushed to the actor's default dataset, one record per item. View as a table or download as JSON / CSV / Excel / XML.

BehaviorDetail
Record granularityOne dataset row per input item that resolved.
endpoint tagEvery row carries the endpoint field for downstream routing.
Column orderMost useful columns (id, title / name, score / karma, created date) placed first.
Unresolved IDsSilently dropped — compare input count vs dataset row count to spot misses.
Mixed formats1s4a4j6, t3_1s4a4j6, full URL all resolve to the same item. Mix freely.

Status & error reference

Run status (Apify-side, shown on the run page)

Apify UI cueStatusApify messageMeaningWhat to do
green checkSUCCEEDED"Actor succeeded with N results in the dataset"Run finished. Some or zero results pushed.Open the dataset to view results.
red exclamationFAILED"The Actor process failed…"Validation error or upstream Reddit fault.Check the run log. You are NOT charged for failed runs.
red clockTIMED-OUT"The Actor timed out. You can resurrect it with a longer timeout to continue where you left off."Run exceeded its timeout.Re-run with a smaller list (≤10000 per run).
red square outlineABORTED"The Actor process was aborted. You can resurrect it to continue where you left off."You stopped the run manually.No charge for unpushed results.

Common in-run conditions (visible in run log)

ConditionCauseResult
Empty result setNone of the inputs resolved (all deleted / banned / typoed).Run SUCCEEDED, 0 records, no charge.
Some IDs droppedSubset deleted / banned / not found.Run SUCCEEDED, fewer rows than inputs.
NSFW posts skippedBulk Posts endpoint and some inputs were NSFW.Run SUCCEEDED, NSFW posts excluded.
Validation error: endpoint is requiredMissing endpoint.Run FAILED immediately, no charge.
Validation error: list too longMore than 10000 entries.Run FAILED immediately, no charge.

Common edge cases

  • Deleted / banned items — returned with whatever metadata Reddit still exposes (often a stub with removed_by_category).
  • Private subreddits — not accessible. Reddit gates them behind a logged-in account and they're skipped.
  • Quarantined content — not returned. Reddit hides quarantined posts/communities from anonymous calls.
  • IDs that don't resolve — silently dropped.
  • Mixed formats — accepted, no normalization needed on your side.
  • NSFW posts — never returned (this actor is SFW-only for posts). Comments / subreddits / users on NSFW communities are returned normally.

Why this actor is fast

  • Speed — large runs scale linearly and finish in minutes, not hours. No browser to boot, no Playwright / Selenium / Puppeteer overhead. Competing browser-based scrapers typically take 15–60 seconds per item.
  • Reliability — zero browser flakiness. No headless-Chromium crashes. No JS-render timeouts. No captcha pages. No surprise mid-run failures from a browser quirk.
  • Footprint — well under the 512 MB allocation, even on full 10000-item runs. Most browser-based scrapers need 1–4 GB.

Pricing

$1.99 per 1,000 results. Pay-per-result — you're only charged for records actually pushed to the dataset.

VolumeCost
100 records~$0.20
500 records~$1.00
1,000 records$1.99
1,500 records~$2.99
10,000 records (one full run)~$19.90
OutcomeCharged?
SUCCEEDED with resultsYes — $1.99 per 1,000 records pushed.
SUCCEEDED with zero recordsNo.
FAILED (validation / upstream)No.
ABORTEDOnly for records already pushed before you stopped.

See the actor's Pricing tab for the live rate — this README is the source of truth at publish time but the Pricing tab always reflects the current price.