Mercari Japan Scraper | $4 / 1k | Fast & Reliable
Pricing
$3.99 / 1,000 results
Mercari Japan Scraper | $4 / 1k | Fast & Reliable
Scrape product listings from Mercari Japan including prices, sellers, conditions history and timestamps. Perfect for resale research, sourcing, and price tracking. Clean, export-ready data for analytics at $4 / 1k listings.
Pricing
$3.99 / 1,000 results
Rating
5.0
(2)
Developer
Fatih Tahta
Actor stats
3
Bookmarked
50
Total users
10
Monthly active users
1.2 days
Issues response
10 hours ago
Last modified
Categories
Share
Mercari Japan Scraper
Slug: fatihtahta/mercari-japan-scraper
Overview
Mercari Japan Scraper collects structured listing data from Mercari Japan, including item identity, title, pricing, status, brand details, media links, seller-related metadata, and scrape context fields for traceability. The actor supports search pages, category pages, and individual listing URLs, then returns normalized JSON records suitable for reporting and analysis. Mercari Japan is one of Japan’s largest consumer marketplaces, making it a valuable source for resale pricing intelligence, assortment tracking, and demand signals. With scheduled and repeatable runs, teams can replace manual collection with a consistent data pipeline. This saves time while improving coverage and comparability across periods.
Why Use This Actor
- Market research / analytics teams: Track pricing ranges, sold-vs-active status, and category movement to monitor market trends and benchmark product segments.
- Product & content teams: Identify high-velocity item themes, optimize merchandising decisions, and prioritize content based on real marketplace activity.
- Developers / data engineering pipelines: Feed structured listing records into BI tools, warehouses, and ETL workflows without building and maintaining custom collectors.
- Lead generation / enrichment teams: Enrich internal catalogs and prospect records with public listing attributes such as brand, price band, and availability status.
- Monitoring / competitive tracking teams: Run recurring snapshots to detect assortment shifts, price changes, and status transitions over time.
Input Parameters
Provide any combination of URLs, queries, and filters to control what gets collected in each run.
| Parameter | Type | Description | Default |
|---|---|---|---|
startUrls | string[] | One or more Mercari Japan URLs to scrape directly. You can mix search pages, category pages, and individual listing pages in the same run. | – |
keyword | string | Optional keyword override applied across search seeds for this run. Useful when you want one consistent term. | – |
sort | string | Optional sorting for search results. Allowed values: score (Relevance), created_time (Newest), price (Price). | – |
order | string | Optional sort direction. Allowed values: desc (Descending), asc (Ascending). | – |
status | string[] | Optional listing status filter. Allowed values: on_sale, sold_out, trading. | – |
priceMin | integer | Optional minimum listing price (JPY). Use to exclude lower-priced items. | – |
priceMax | integer | Optional maximum listing price (JPY). Use to cap results to your target range. | – |
limit | integer | Maximum listings to save per query. Use lower values for quick validation and higher values for deeper coverage. | 50000 |
Example Input
{"startUrls": ["https://jp.mercari.com/en/search?keyword=apple%20watch"],"keyword": "apple watch","sort": "created_time","order": "desc","status": ["on_sale", "sold_out"],"priceMin": 1000,"priceMax": 50000,"limit": 1000}
Output
6.1 Output destination
The actor writes results to an Apify dataset as JSON records. And the dataset is designed for direct consumption by analytics tools, ETL pipelines, and downstream APIs without post-processing.
6.2 Record envelope (all items)
Every record includes:
- type (string, required)
- id (number, required)
- url (string, required)
Recommended idempotency key: type + ":" + id
Use this key for deduplication and upserts so repeated runs can safely merge updates for the same entity.
6.3 Examples
Example: listing (type = "listing")
{"type": "listing","id": 84015541919,"url": "https://jp.mercari.com/item/m84015541919","listing_id": "m84015541919","title": "Apple Watch 箱のみ","price": "300","currency": "JPY","brand_name": "Apple","brand": {"id": "3272","name": "Apple","sub_name": "Apple"},"listing_status": "ITEM_STATUS_ON_SALE","listing_type": "ITEM_TYPE_MERCARI","condition_id": "2","seller": {"seller_id": "670753842","buyer_id": ""},"shipping": {"payer_id": "2","method_id": "14"},"catalog": {"category_id": "3682","size_ids": []},"store": {"name": "","details": null},"media": {"thumbnail_urls": ["https://static.mercdn.net/thumb/item/webp/m84015541919_1.jpg?1772980274"],"photo_urls": ["https://static.mercdn.net/item/detail/webp/photos/m84015541919_1.jpg?1772980274"]},"promotions": [],"auction": {"auction_id": "","bid_deadline": "2026-03-10T11:38:24.069450856Z","total_bid_count": "0","highest_bid": "300","initial_price": "300"},"timestamps": {"created_unix": "1772980274","updated_unix": "1772980578"},"engagement": {"is_liked": false,"is_no_price": false},"scrape_context": {"seed": {"type": null,"value": null},"source": {"source_url": null,"item_url": "https://jp.mercari.com/item/m84015541919"},"scraped_time": 1773034705}}
Field reference
Listing fields (type = "listing")
- listing_id (string, required): Marketplace listing identifier.
- title (string, optional): Listing title.
- price (string, optional): Listed price value as provided.
- currency (string, optional): Currency code (for example,
JPY). - brand_name (string, optional): Top-level brand label.
- brand.id (string, optional): Brand identifier.
- brand.name (string, optional): Brand name.
- brand.sub_name (string, optional): Secondary brand label.
- listing_status (string, optional): Listing availability/status value.
- listing_type (string, optional): Listing type classification.
- condition_id (string, optional): Item condition code.
- seller.seller_id (string, optional): Seller identifier.
- seller.buyer_id (string, optional): Buyer identifier when available.
- shipping.payer_id (string, optional): Shipping payer code.
- shipping.method_id (string, optional): Shipping method code.
- catalog.category_id (string, optional): Category identifier.
- catalog.size_ids (array, optional): Size identifiers.
- store.name (string, optional): Store name when present.
- store.details (object/null, optional): Additional store details.
- media.thumbnail_urls (array[string], optional): Thumbnail image URLs.
- media.photo_urls (array[string], optional): Main photo URLs.
- promotions (array, optional): Promotion entries, if present.
- auction.auction_id (string, optional): Auction identifier.
- auction.bid_deadline (string, optional): Auction bid deadline timestamp.
- auction.total_bid_count (string, optional): Total bid count.
- auction.highest_bid (string, optional): Highest bid value.
- auction.initial_price (string, optional): Auction initial price.
- timestamps.created_unix (string, optional): Creation time in Unix format.
- timestamps.updated_unix (string, optional): Update time in Unix format.
- engagement.is_liked (boolean, optional): Whether the item is marked liked in context.
- engagement.is_no_price (boolean, optional): Whether no-price mode is indicated.
- scrape_context.seed.type (string/null, optional): Seed type used for discovery.
- scrape_context.seed.value (string/null, optional): Seed value used for discovery.
- scrape_context.source.source_url (string/null, optional): Source page URL.
- scrape_context.source.item_url (string, optional): Item URL reference.
- scrape_context.scraped_time (number, optional): Scrape timestamp.
Data guarantees & handling
- Best-effort extraction: fields may vary by region/session/availability/UI experiments.
- Optional fields: null-check in downstream code.
- Deduplication: recommend
type + ":" + id.
How to Run on Apify
- Open the Actor in Apify Console.
- Configure your search parameters (for example keyword, status, and price range).
- Set the maximum number of outputs to collect.
- Click Start and wait for the run to finish.
- Download results in JSON, CSV, Excel, or other supported formats.
Scheduling & Automation
Scheduling
Automated Data Collection
You can schedule recurring runs to keep your Mercari dataset continuously fresh. This is useful for trend monitoring, recurring reports, and automated downstream updates.
- Navigate to Schedules in Apify Console
- Create a new schedule (daily, weekly, or custom cron)
- Configure input parameters
- Enable notifications for run completion
- (Optional) Add webhooks for automated processing
Integration Options
- Webhooks: Trigger downstream actions when a run completes
- Zapier: Connect to 5,000+ apps without coding
- Make (Integromat): Build multi-step automation workflows
- Google Sheets: Export results to a spreadsheet
- Slack/Discord: Receive notifications and summaries
- Email: Send automated reports via email
Performance
Estimated execution times:
- Small runs (< 1,000 outputs): ~2–3 minutes
- Medium runs (1,000–5,000 outputs): ~5–15 minutes
- Large runs (5,000+ outputs): ~15–30 minutes
Execution time varies based on filters, result volume, and how much information is returned per record.
Compliance & Ethics
Responsible Data Collection
This actor collects publicly available marketplace listing information from jp.mercari.com for legitimate business purposes, including:
- e-commerce and resale research and market analysis
- price monitoring and catalog enrichment
- competitive tracking and reporting
Users are responsible for ensuring their use complies with applicable laws, regulations, and platform terms. This section is informational and not legal advice.
Best Practices
- Use collected data in accordance with applicable laws, regulations, and the target site’s terms
- Respect individual privacy and personal information
- Use data responsibly and avoid disruptive or excessive collection
- Do not use this actor for spamming, harassment, or other harmful purposes
- Follow relevant data protection requirements where applicable (e.g., GDPR, CCPA)
Support
For help, use the Issues tab on the actor page in Apify Console. Include the input used (redacted), run ID, expected vs actual behavior, and an optional small output sample so issues can be reproduced quickly.