AiJobs.net Scraper avatar

AiJobs.net Scraper

Pricing

Pay per usage

Go to Apify Store
AiJobs.net Scraper

AiJobs.net Scraper

Automatically scrape AI job listings from AiJobs.net. Extract job titles, companies, locations, salaries, and full descriptions. Perfect for recruitment research, job market analysis, and career tracking with reliable data extraction.

Pricing

Pay per usage

Rating

0.0

(0)

Developer

Shahid Irfan

Shahid Irfan

Maintained by Community

Actor stats

0

Bookmarked

1

Total users

0

Monthly active users

4 days ago

Last modified

Share

Extract job listings from aijobs.net with flexible inputs for direct URLs, keyword discovery, and location targeting. Build high-quality datasets for hiring intelligence, compensation benchmarking, and AI talent market monitoring with richer job detail fields such as tasks, perks, skills, education, role labels, and geographic breakdowns.

Features

  • URL-first scraping — Start from a specific listing page or a direct job URL.
  • Keyword and location search — Find relevant jobs with search-friendly input fields.
  • Pagination support — Collect data across multiple listing pages with a page limit.
  • Rich job detail extraction — Capture salary, level, tasks, perks, skills, education, roles, and region hierarchy.
  • Clean datasets — Records exclude empty and null values for analysis-ready output.

Use Cases

AI Talent Market Research

Track demand for AI, ML, and data roles across countries and cities. Compare job volume and skills trends over time.

Hiring Pipeline Discovery

Identify companies actively hiring for specific roles and technologies. Build targeted outreach lists for recruiting and business development.

Compensation Benchmarking

Collect salary ranges when available to compare market rates across job families and regions.

Skills Trend Monitoring

Measure how often key technologies and competencies appear in active job listings.

Job Intelligence Dashboards

Feed clean job data into BI tools for recurring reporting and competitive intelligence.


Input Parameters

ParameterTypeRequiredDefaultDescription
urlStringNo"https://aijobs.net/"Start URL for a listing page or direct job page.
keywordStringNo"python"Keyword used to discover relevant jobs.
locationStringNo""Location filter for region-specific jobs.
results_wantedIntegerNo20Maximum number of jobs to save.
max_pagesIntegerNo5Maximum number of listing pages to request.
startUrlStringNoAlias for url for compatibility.
proxyConfigurationObjectNoApify Proxy (Residential)Proxy configuration for reliable data collection.

Output Data

Each item in the dataset contains:

FieldTypeDescription
titleStringJob title
companyStringHiring company name
company_profile_urlStringCompany profile URL when publicly available
company_slugStringCompany slug derived from the profile path
company_idStringCompany identifier when available in the public profile path
locationStringPrimary job location
locationsArrayAll detected location values
salaryStringHuman-readable salary badge when shown
levelStringSeniority badge
employment_typeStringEmployment type
employment_typesArrayAll employment-type badges shown on the job
posted_agoStringRelative posting age shown on the page
skillsArraySkills associated with the role
tasksArrayTask statements listed on the job detail page
perksArrayPerks and benefits listed on the job detail page
educationArrayEducation labels shown on the job detail page
rolesArrayRole labels associated with the job
regionsArrayRegion labels associated with the job
countriesArrayCountry labels associated with the job
statesArrayState or province labels associated with the job
citiesArrayCity labels associated with the job
apply_urlStringPublic apply path URL shown on the job page
apply_pathStringRelative apply path shown on the site
apply_idStringApply-path identifier when available
identifierStringJob identifier when provided
description_textStringCleaned plain-text description
urlStringJob detail URL
sourceStringData source domain

Usage Examples

Basic Run

{
"url": "https://aijobs.net/",
"results_wanted": 20,
"max_pages": 5
}
{
"keyword": "data engineer",
"location": "Germany",
"results_wanted": 30,
"max_pages": 6
}

Direct Listing URL

{
"url": "https://aijobs.net/jobs/api/",
"results_wanted": 25,
"max_pages": 4
}

Single Job URL

{
"url": "https://aijobs.net/job/frontier-ai-research-lead-georgetown-university-main-campus-walsh-school-of-foreign-service-500-first-st-nw-7th-floor-74243/"
}

Sample Output

{
"title": "Senior / Lead Data Engineer (24x7 Data & AI Factory)",
"company": "Devoteam",
"company_profile_url": "https://aijobs.net/company/devoteam-317/",
"company_slug": "devoteam",
"company_id": "317",
"location": "Kraków, Poland",
"locations": ["Kraków, Poland", "Kraków, Lesser Poland, PL", "Lesser Poland, PL", "Poland", "Europe"],
"salary": "PLN 258K-370K (estimate)",
"level": "Senior-level",
"employment_type": "Full Time",
"posted_ago": "17h ago",
"skills": ["Airflow", "Apache Spark", "DBT", "Data Observability"],
"tasks": ["Build AI models", "Build data pipelines", "Improve data observability"],
"perks": ["Conference attendance", "Hybrid work model", "Private medical healthcare"],
"education": ["Bachelor of Engineering", "Bachelor of Science"],
"roles": ["Data Engineer", "Lead Data Engineer", "Senior Data Engineer"],
"countries": ["Poland"],
"states": ["Lesser Poland, PL"],
"cities": ["Kraków, Lesser Poland, PL"],
"apply_url": "https://aijobs.net/job/1zBHEnUHQOy3wjW/apply/",
"apply_id": "1zBHEnUHQOy3wjW",
"identifier": "89359",
"description_text": "Advise junior team members; Build AI models; Build data pipelines; Design AI models; Design data pipelines; Improve data observability; Maintain AI models; Maintain data observability; Maintain data pipelines; Prepare business insights solutions; Support dataops and ai projects; Transform data; Troubleshoot AI models; Troubleshoot data observability; Troubleshoot data pipelines;",
"url": "https://aijobs.net/job/senior-lead-data-engineer-24x7-data-ai-factory-krakow-poland-89359/",
"source": "aijobs.net"
}

Tips for Best Results

Start with Small Runs

  • Use results_wanted: 20 for fast validation.
  • Increase result volume after confirming output quality.

Use Strong Keywords

  • Prefer specific role names like ml engineer, computer vision, or data scientist.
  • Combine with location for more targeted datasets.

Set Practical Pagination Limits

  • Increase max_pages for broader discovery.
  • Keep limits reasonable to maintain fast run times.

Use Proxies for Stability

  • Residential proxy settings are recommended for reliable multi-page runs.

Integrations

  • Google Sheets — Export job data for collaborative review.
  • Airtable — Build searchable hiring intelligence tables.
  • Slack — Send run notifications and alerts.
  • Make — Automate enrichment and downstream workflows.
  • Zapier — Trigger alerts and CRM updates.
  • Webhooks — Push datasets to your own services.

Export Formats

  • JSON — Best for APIs and engineering workflows
  • CSV — Best for spreadsheet analysis
  • Excel — Best for business reporting
  • XML — Best for system integrations

Frequently Asked Questions

Can I run with only a keyword?

Yes. Provide keyword and optionally location, and the actor will discover matching listings.

Can I run with only a URL?

Yes. You can provide either a listing URL or a direct job URL.

Why are some fields missing in certain records?

Some job listings do not provide every field. The actor keeps only available values to avoid null-heavy output.

How many jobs can I collect?

You can scale based on results_wanted and max_pages, depending on available listings.

Does user input override defaults?

Yes. Runtime input values always take priority over prefill/default values.


Support

For issues or feature requests, use the Apify Console issue/reporting channels for this actor.

Resources


This actor is intended for legitimate data collection and market research. Users are responsible for complying with website terms and applicable laws in their jurisdiction.