Bazaraki Scraper avatar
Bazaraki Scraper

Pricing

Pay per usage

Go to Apify Store
Bazaraki Scraper

Bazaraki Scraper

An Apify actor for scraping classified listings from bazaraki.com with category selection and automatic Excel export. Category Selection: Choose from predefined categories (Real Estate, Vehicles, Jobs, Services, Electronics, etc.) Record Limit: Specify the maximum number of ads to scrape

Pricing

Pay per usage

Rating

0.0

(0)

Developer

I G

I G

Maintained by Community

Actor stats

0

Bookmarked

2

Total users

1

Monthly active users

a day ago

Last modified

Share

Bazaraki.com Scraper

An Apify actor for scraping classified listings from bazaraki.com with category selection and Excel export.

Important

For a big amount of results, notice to put 0 in the timeout, under : input->run options-> 0

Features

  • Category Selection: Choose from predefined categories (Real Estate, Vehicles, Jobs, Services, Electronics, etc.)
  • Record Limit: Specify the maximum number of ads to scrape
  • Excel Export: Automatically exports all scraped data to an Excel file
  • Comprehensive Data Extraction: Extracts all listing details including:
    • Title
    • Price
    • Description
    • Location
    • Category
    • Images (as semicolon-separated URLs)
    • Contact information (phone, email)
    • Attributes/metadata (flattened into columns)
    • Posted date
    • Scraped timestamp
  • Smart Pagination: Automatically follows pagination until the record limit is reached
  • Rate Limiting: Respects the website with delays between requests

Input Parameters

The actor accepts the following input parameters:

  • category (required): Select the category to scrape from bazaraki.com

    • Options: All Categories, Real Estate, Vehicles, Jobs, Services, Electronics, Home & Garden, Fashion, Sports & Hobbies, Pets, Other
    • Default: "all"
  • maxRecords (required): Maximum number of ads to scrape

    • Type: Integer
    • Range: 1-10000
    • Default: 100

Output

The actor outputs:

  1. Dataset: All scraped listings are saved to the Apify dataset
  2. Excel File: An Excel file (.xlsx) is automatically generated and saved to the Key-Value Store with the filename format: bazaraki_{category}_{date}.xlsx

Excel File Structure

Each row in the Excel file contains:

  • url: Full URL of the listing
  • title: Listing title
  • price: Price (numeric)
  • description: Full description
  • location: Location information
  • category: Category name
  • images: Semicolon-separated list of image URLs
  • phone: Contact phone number
  • email: Contact email address
  • postedDate: Date when the ad was posted
  • scrapedAt: Timestamp when the data was scraped
  • attr_*: Additional attributes as separate columns (flattened from metadata)

Local Development

  1. Install dependencies:
$npm install
  1. (Optional) Update categories in input schema:
$npm run update-categories

This fetches the latest categories from bazaraki.com and updates the input schema dropdown.

  1. Run the actor locally:
$npm start

Or with Apify CLI:

$apify run

Deployment to Apify

  1. Make sure you have the Apify CLI installed:
$npm install -g apify-cli
  1. Login to Apify:
$apify login
  1. Push the actor to Apify:
$apify push

Usage Example

Input JSON:

{
"category": "real-estate",
"maxRecords": 500
}

Using Apify Platform:

  1. Select the category from the dropdown (e.g., "Real Estate")
  2. Enter the maximum number of records (e.g., 500)
  3. Run the actor
  4. Download the Excel file from the Key-Value Store after completion

Notes

  • The scraper includes delays between requests to be respectful to the website
  • Make sure to comply with bazaraki.com's robots.txt and terms of service
  • The scraper uses Cheerio for HTML parsing, which is fast and efficient
  • Images are exported as semicolon-separated URLs in a single column
  • Contact information and attributes are extracted using multiple selector strategies to handle different page layouts
  • The scraper automatically stops when the maximum number of records is reached
  • For large scrapes (1000+ records): The default timeout is 2 hours (7200 seconds). If you need to scrape more records, you may need to increase the timeout in the Apify platform settings when running the actor

License

ISC