AiJobs.net Scraper
Pricing
Pay per usage
AiJobs.net Scraper
Automatically scrape AI job listings from AiJobs.net. Extract job titles, companies, locations, salaries, and full descriptions. Perfect for recruitment research, job market analysis, and career tracking with reliable data extraction.
Pricing
Pay per usage
Rating
0.0
(0)
Developer
Shahid Irfan
Actor stats
0
Bookmarked
1
Total users
0
Monthly active users
4 days ago
Last modified
Categories
Share
Extract job listings from aijobs.net with flexible inputs for direct URLs, keyword discovery, and location targeting. Build high-quality datasets for hiring intelligence, compensation benchmarking, and AI talent market monitoring with richer job detail fields such as tasks, perks, skills, education, role labels, and geographic breakdowns.
Features
- URL-first scraping — Start from a specific listing page or a direct job URL.
- Keyword and location search — Find relevant jobs with search-friendly input fields.
- Pagination support — Collect data across multiple listing pages with a page limit.
- Rich job detail extraction — Capture salary, level, tasks, perks, skills, education, roles, and region hierarchy.
- Clean datasets — Records exclude empty and null values for analysis-ready output.
Use Cases
AI Talent Market Research
Track demand for AI, ML, and data roles across countries and cities. Compare job volume and skills trends over time.
Hiring Pipeline Discovery
Identify companies actively hiring for specific roles and technologies. Build targeted outreach lists for recruiting and business development.
Compensation Benchmarking
Collect salary ranges when available to compare market rates across job families and regions.
Skills Trend Monitoring
Measure how often key technologies and competencies appear in active job listings.
Job Intelligence Dashboards
Feed clean job data into BI tools for recurring reporting and competitive intelligence.
Input Parameters
| Parameter | Type | Required | Default | Description |
|---|---|---|---|---|
url | String | No | "https://aijobs.net/" | Start URL for a listing page or direct job page. |
keyword | String | No | "python" | Keyword used to discover relevant jobs. |
location | String | No | "" | Location filter for region-specific jobs. |
results_wanted | Integer | No | 20 | Maximum number of jobs to save. |
max_pages | Integer | No | 5 | Maximum number of listing pages to request. |
startUrl | String | No | — | Alias for url for compatibility. |
proxyConfiguration | Object | No | Apify Proxy (Residential) | Proxy configuration for reliable data collection. |
Output Data
Each item in the dataset contains:
| Field | Type | Description |
|---|---|---|
title | String | Job title |
company | String | Hiring company name |
company_profile_url | String | Company profile URL when publicly available |
company_slug | String | Company slug derived from the profile path |
company_id | String | Company identifier when available in the public profile path |
location | String | Primary job location |
locations | Array | All detected location values |
salary | String | Human-readable salary badge when shown |
level | String | Seniority badge |
employment_type | String | Employment type |
employment_types | Array | All employment-type badges shown on the job |
posted_ago | String | Relative posting age shown on the page |
skills | Array | Skills associated with the role |
tasks | Array | Task statements listed on the job detail page |
perks | Array | Perks and benefits listed on the job detail page |
education | Array | Education labels shown on the job detail page |
roles | Array | Role labels associated with the job |
regions | Array | Region labels associated with the job |
countries | Array | Country labels associated with the job |
states | Array | State or province labels associated with the job |
cities | Array | City labels associated with the job |
apply_url | String | Public apply path URL shown on the job page |
apply_path | String | Relative apply path shown on the site |
apply_id | String | Apply-path identifier when available |
identifier | String | Job identifier when provided |
description_text | String | Cleaned plain-text description |
url | String | Job detail URL |
source | String | Data source domain |
Usage Examples
Basic Run
{"url": "https://aijobs.net/","results_wanted": 20,"max_pages": 5}
Keyword + Location Search
{"keyword": "data engineer","location": "Germany","results_wanted": 30,"max_pages": 6}
Direct Listing URL
{"url": "https://aijobs.net/jobs/api/","results_wanted": 25,"max_pages": 4}
Single Job URL
{"url": "https://aijobs.net/job/frontier-ai-research-lead-georgetown-university-main-campus-walsh-school-of-foreign-service-500-first-st-nw-7th-floor-74243/"}
Sample Output
{"title": "Senior / Lead Data Engineer (24x7 Data & AI Factory)","company": "Devoteam","company_profile_url": "https://aijobs.net/company/devoteam-317/","company_slug": "devoteam","company_id": "317","location": "Kraków, Poland","locations": ["Kraków, Poland", "Kraków, Lesser Poland, PL", "Lesser Poland, PL", "Poland", "Europe"],"salary": "PLN 258K-370K (estimate)","level": "Senior-level","employment_type": "Full Time","posted_ago": "17h ago","skills": ["Airflow", "Apache Spark", "DBT", "Data Observability"],"tasks": ["Build AI models", "Build data pipelines", "Improve data observability"],"perks": ["Conference attendance", "Hybrid work model", "Private medical healthcare"],"education": ["Bachelor of Engineering", "Bachelor of Science"],"roles": ["Data Engineer", "Lead Data Engineer", "Senior Data Engineer"],"countries": ["Poland"],"states": ["Lesser Poland, PL"],"cities": ["Kraków, Lesser Poland, PL"],"apply_url": "https://aijobs.net/job/1zBHEnUHQOy3wjW/apply/","apply_id": "1zBHEnUHQOy3wjW","identifier": "89359","description_text": "Advise junior team members; Build AI models; Build data pipelines; Design AI models; Design data pipelines; Improve data observability; Maintain AI models; Maintain data observability; Maintain data pipelines; Prepare business insights solutions; Support dataops and ai projects; Transform data; Troubleshoot AI models; Troubleshoot data observability; Troubleshoot data pipelines;","url": "https://aijobs.net/job/senior-lead-data-engineer-24x7-data-ai-factory-krakow-poland-89359/","source": "aijobs.net"}
Tips for Best Results
Start with Small Runs
- Use
results_wanted: 20for fast validation. - Increase result volume after confirming output quality.
Use Strong Keywords
- Prefer specific role names like
ml engineer,computer vision, ordata scientist. - Combine with location for more targeted datasets.
Set Practical Pagination Limits
- Increase
max_pagesfor broader discovery. - Keep limits reasonable to maintain fast run times.
Use Proxies for Stability
- Residential proxy settings are recommended for reliable multi-page runs.
Integrations
- Google Sheets — Export job data for collaborative review.
- Airtable — Build searchable hiring intelligence tables.
- Slack — Send run notifications and alerts.
- Make — Automate enrichment and downstream workflows.
- Zapier — Trigger alerts and CRM updates.
- Webhooks — Push datasets to your own services.
Export Formats
- JSON — Best for APIs and engineering workflows
- CSV — Best for spreadsheet analysis
- Excel — Best for business reporting
- XML — Best for system integrations
Frequently Asked Questions
Can I run with only a keyword?
Yes. Provide keyword and optionally location, and the actor will discover matching listings.
Can I run with only a URL?
Yes. You can provide either a listing URL or a direct job URL.
Why are some fields missing in certain records?
Some job listings do not provide every field. The actor keeps only available values to avoid null-heavy output.
How many jobs can I collect?
You can scale based on results_wanted and max_pages, depending on available listings.
Does user input override defaults?
Yes. Runtime input values always take priority over prefill/default values.
Support
For issues or feature requests, use the Apify Console issue/reporting channels for this actor.
Resources
Legal Notice
This actor is intended for legitimate data collection and market research. Users are responsible for complying with website terms and applicable laws in their jurisdiction.