Bazaraki Scraper
Pricing
Pay per usage
Bazaraki Scraper
An Apify actor for scraping classified listings from bazaraki.com with category selection and automatic Excel export. Category Selection: Choose from predefined categories (Real Estate, Vehicles, Jobs, Services, Electronics, etc.) Record Limit: Specify the maximum number of ads to scrape
Pricing
Pay per usage
Rating
0.0
(0)
Developer

I G
Actor stats
0
Bookmarked
2
Total users
1
Monthly active users
a day ago
Last modified
Categories
Share
Bazaraki.com Scraper
An Apify actor for scraping classified listings from bazaraki.com with category selection and Excel export.
Important
For a big amount of results, notice to put 0 in the timeout, under : input->run options-> 0
Features
- Category Selection: Choose from predefined categories (Real Estate, Vehicles, Jobs, Services, Electronics, etc.)
- Record Limit: Specify the maximum number of ads to scrape
- Excel Export: Automatically exports all scraped data to an Excel file
- Comprehensive Data Extraction: Extracts all listing details including:
- Title
- Price
- Description
- Location
- Category
- Images (as semicolon-separated URLs)
- Contact information (phone, email)
- Attributes/metadata (flattened into columns)
- Posted date
- Scraped timestamp
- Smart Pagination: Automatically follows pagination until the record limit is reached
- Rate Limiting: Respects the website with delays between requests
Input Parameters
The actor accepts the following input parameters:
-
category (required): Select the category to scrape from bazaraki.com
- Options: All Categories, Real Estate, Vehicles, Jobs, Services, Electronics, Home & Garden, Fashion, Sports & Hobbies, Pets, Other
- Default: "all"
-
maxRecords (required): Maximum number of ads to scrape
- Type: Integer
- Range: 1-10000
- Default: 100
Output
The actor outputs:
- Dataset: All scraped listings are saved to the Apify dataset
- Excel File: An Excel file (.xlsx) is automatically generated and saved to the Key-Value Store with the filename format:
bazaraki_{category}_{date}.xlsx
Excel File Structure
Each row in the Excel file contains:
url: Full URL of the listingtitle: Listing titleprice: Price (numeric)description: Full descriptionlocation: Location informationcategory: Category nameimages: Semicolon-separated list of image URLsphone: Contact phone numberemail: Contact email addresspostedDate: Date when the ad was postedscrapedAt: Timestamp when the data was scrapedattr_*: Additional attributes as separate columns (flattened from metadata)
Local Development
- Install dependencies:
$npm install
- (Optional) Update categories in input schema:
$npm run update-categories
This fetches the latest categories from bazaraki.com and updates the input schema dropdown.
- Run the actor locally:
$npm start
Or with Apify CLI:
$apify run
Deployment to Apify
- Make sure you have the Apify CLI installed:
$npm install -g apify-cli
- Login to Apify:
$apify login
- Push the actor to Apify:
$apify push
Usage Example
Input JSON:
{"category": "real-estate","maxRecords": 500}
Using Apify Platform:
- Select the category from the dropdown (e.g., "Real Estate")
- Enter the maximum number of records (e.g., 500)
- Run the actor
- Download the Excel file from the Key-Value Store after completion
Notes
- The scraper includes delays between requests to be respectful to the website
- Make sure to comply with bazaraki.com's robots.txt and terms of service
- The scraper uses Cheerio for HTML parsing, which is fast and efficient
- Images are exported as semicolon-separated URLs in a single column
- Contact information and attributes are extracted using multiple selector strategies to handle different page layouts
- The scraper automatically stops when the maximum number of records is reached
- For large scrapes (1000+ records): The default timeout is 2 hours (7200 seconds). If you need to scrape more records, you may need to increase the timeout in the Apify platform settings when running the actor
License
ISC