LinkedIn Jobs Scraper — Pay only for new or changed listings
Pricing
from $1.00 / 1,000 results
LinkedIn Jobs Scraper — Pay only for new or changed listings
Scrape linkedin.com jobs — title, company, location, salary, full description, apply URL, contact email — with state-keyed delta pricing. Recurring runs only charge for new or changed listings; already-seen jobs are skipped. Telegram/Discord/Slack/WhatsApp/webhook notifications.
Pricing
from $1.00 / 1,000 results
Rating
0.0
(0)
Developer
Black Falcon Data
Actor stats
1
Bookmarked
3
Total users
1
Monthly active users
2 days ago
Last modified
Categories
Share
What does LinkedIn Jobs Scraper do?
LinkedIn Jobs Scraper extracts structured job data from linkedin.com — including salary data, contact details (email, apply URL), company metadata, full descriptions, and location data. It supports location filters and controllable result limits, so you can run the same query consistently over time.
Key features
- 🔁 Incremental mode (delta pricing) — recurring runs only emit and charge for listings that are new or whose tracked content changed. Pair with Apify Schedules to run hourly or daily and pay 90%+ less than full re-scrapes.
- 📝 Detail enrichment — fetch full job descriptions, criteria (seniority level, employment type, industry, job function, workplace type), applicant counts, and posting benefits. Toggle on with
enrichDetails: true. Description is returned as plain text, HTML, and Markdown. - 🌍 Cross-region in one run — combine
geoIds[],regions[](ISO-2 codes), and 11 built-in presets (nordic,dach,benelux,uk-ireland,eu-27,gcc,mena,asean,anglosphere,latam,nordics-extended). Results are deduped onjobId. - 🎯 Powerful filters —
keywords,datePosted(including uniquelastHour),jobType,experienceLevel,workType(onsite/remote/hybrid),salaryMin/salaryMax,companies,excludeCompanies,excludeKeywords,removeAgency,easyApply,distance, andsortBy. - 🔗 Related-jobs discovery — set
discoverRelated: trueto expand seed jobs with LinkedIn's own related-jobs feed. Adds 10-30% more relevant matches on thin-market queries, deduped against the main result set. - ♻️ Repost detection — when a previously expired listing reappears with matching content, it is flagged with
isRepost: trueandrepostOfId. SetskipReposts: trueto filter them out of the dataset. - 📦 Compact mode —
compact: truereturns core fields only (AI-agent / MCP-friendly). Cuts payload by 80%+ when piping into LLMs. - ✂️ Description truncation — cap
descriptionMaxLengthto control output size and per-record cost when feeding pipelines that don't need full text. - 🔔 Notifications built-in — push results to Telegram, Discord, Slack, WhatsApp, or a custom webhook (n8n / Make / Zapier) after each run. Pair with
notifyOnlyChangesto alert only on new or updated jobs. - 🔓 No login required — uses LinkedIn's public guest-jobs endpoints. Datacenter proxies are sufficient; no account credentials required.
What data can you extract from linkedin.com?
Each result includes Core listing fields (scrapedAt, portalUrl, jobId, linkedinJobId, jobUrl, title, location, and country, and more), detail fields when enrichment is enabled (description, descriptionHtml, descriptionMarkdown, and postingBenefits), contact and apply information (applyUrl, applyType, easyApply, and extractedEmails), and company metadata (company, companyUrl, companyId, and companyLogo). In standard mode, all fields are always present — unavailable data points are returned as null, never omitted. In compact mode, only core fields are returned.
Input
The main inputs are an optional location filter and a result limit. Additional filters and options are available in the input schema.
Key parameters:
keywords— Job search keywords (e.g. "software engineer", "nurse"). Leave blank to browse all jobs in the selected location.location— Free-text location (e.g. "Copenhagen, Denmark", "United States"). Use geoIds for higher precision.geoIds— Numeric LinkedIn geoIds (e.g. "103644278" = United States). Each geoId becomes a separate query, deduped on jobId. (default:[])regions— Two-letter country codes (e.g. "US", "GB", "DE"). Resolved to LinkedIn country geoIds. Use geoIds[] for unsupported markets. (default:[])regionPresets— Pre-defined country grouping. Combined with regions[] if both are set.datePosted— Filter by posting recency. "lastHour" is unique to this scraper. (default:"anytime")jobType— Multi-select employment type filter. (default:[])experienceLevel— Multi-select seniority filter. (default:[])workType— Multi-select onsite/remote/hybrid filter. (default:[])salaryMin— Minimum annual salary (USD). Mapped to LinkedIn's nearest f_SB2 bucket. Post-filtered exactly.salaryMax— Maximum annual salary. Post-filtered (LinkedIn has no native max filter).salaryIncludeUnknown— When salaryMin/Max set, include jobs with no salary data. (default:true)- ...and 35 more parameters
Input examples
Basic search — Keyword-driven search with a result cap.
→ Full payload per result — all standard fields populated where the source provides them.
{"keywords": "software engineer","maxResults": 50}
Filtered search — Narrow results with advanced filters — only matching jobs are returned.
→ Same field set as basic search; fewer, more relevant rows.
{"keywords": "software engineer","jobType": ["fulltime"],"workType": ["onsite"],"experienceLevel": ["internship"],"maxResults": 100}
Incremental tracking — Only emit jobs that changed since the previous run with this stateKey.
→ First run builds the baseline state. Subsequent runs emit only records that are new or whose tracked content changed. Set emitUnchanged: true to include unchanged records as well.
{"keywords": "software engineer","maxResults": 200,"incrementalMode": true,"stateKey": "software-engineer-tracker"}
Compact filtered output — Combine filters with compact mode for a lightweight AI-agent or MCP data source.
→ Core fields only — ideal for piping into LLMs or downstream tools without token overhead.
{"keywords": "software engineer","jobType": ["fulltime"],"workType": ["onsite"],"maxResults": 50,"compact": true}
Output
Each run produces a dataset of structured job records. Results can be downloaded as JSON, CSV, or Excel from the Dataset tab in Apify Console.
Example job record
{"scrapedAt": "2026-04-27T19:02:37.769Z","portalUrl": "https://www.linkedin.com","source": "linkedin","jobId": "1705fc4ee704bf3584cf2654b20e8f95383167563ff7ecd0184b58d2c7d66236","linkedinJobId": "4406118990","jobUrl": "https://www.linkedin.com/jobs/view/software-engineer-new-grad-at-notion-4406118990","title": "Software Engineer, New Grad","company": "Notion","companyUrl": "https://www.linkedin.com/company/notionhq","companyId": "notionhq","location": "San Francisco, CA","country": "CA","postedAt": "2026-04-24T00:00:00.000Z","applyUrl": "https://www.linkedin.com/jobs/view/software-engineer-new-grad-at-notion-4406118990","applyType": "unknown","description": "About Us Notion helps you build beautiful tools for your life’s work. In today's world of endless apps and tabs, Notion provides one place for teams to get everything done, seamlessly connecting docs,...","descriptionHtml": "<strong>About Us<br><br></strong>Notion helps you build beautiful tools for your life’s work. In today's world of endless apps and tabs, Notion provides one place for teams to get everything done, sea...","descriptionMarkdown": "About Us Notion helps you build beautiful tools for your life’s work. In today's world of endless apps and tabs, Notion provides one place for teams to get everything done, seamlessly connecting docs,...","seniorityLevel": "Not Applicable","employmentType": "Full-time","industry": "Software Development","jobFunction": "Engineering and Information Technology","workplaceType": null,"applicantCount": 200,"easyApply": false}
Incremental fields
When incremental: true, each record also carries:
changeType— one ofNEW,UPDATED,UNCHANGED,REAPPEARED,EXPIRED. Default output coversNEW/UPDATED/REAPPEARED; setemitUnchanged: trueoremitExpired: trueto opt into the others.firstSeenAt,lastSeenAt— ISO-8601 timestamps tracking the listing across runs.isRepost,repostOfId,repostDetectedAt— populated when a new listing matches the tracked content of a previously expired one. SetskipReposts: trueto drop detected reposts from the output.
How to scrape linkedin.com
- Go to LinkedIn Jobs Scraper in Apify Console.
- Configure the input and optional location filter.
- Set
maxResultsto control how many results you need. - Click Start and wait for the run to finish.
- Export the dataset as JSON, CSV, or Excel.
Use cases
- Extract job data from linkedin.com for market research and competitive analysis.
- Track salary trends across regions and categories over time.
- Monitor new and changed listings on scheduled runs without processing the full dataset every time.
- Build outreach lists using contact details and apply URLs from listings.
- Research company hiring patterns, employer profiles, and industry distribution.
- Use structured location data for regional analysis, mapping, and geo-targeting.
- Feed structured data into AI agents, MCP tools, and automated pipelines using compact mode.
- Export clean, structured data to dashboards, spreadsheets, or data warehouses.
How much does it cost to scrape linkedin.com?
LinkedIn Jobs Scraper uses pay-per-event pricing. You pay a small fee when the run starts and then for each result that is actually produced.
- Run start: $0.0005 per run
- Per result: $0.001 per job record
Example costs:
- 10 results: $0.01
- 100 results: $0.1
- 500 results: $0.5
Example: recurring monitoring savings
These examples compare full re-scrapes with incremental runs at different churn rates. Churn is the share of listings that are new or whose tracked content changed since the previous run. Actual churn depends on your query breadth, source activity, and polling frequency — the scenarios below are examples, not predictions.
Example setup: 250 results per run, daily polling (30 runs/month). Event-pricing examples scale linearly with result count.
| Churn rate | Full re-scrape run cost | Incremental run cost | Savings vs full re-scrape | Monthly cost after baseline |
|---|---|---|---|---|
| 5% — stable niche query | $0.25 | $0.01 | $0.24 (95%) | $0.39 |
| 15% — moderate broad query | $0.25 | $0.04 | $0.21 (85%) | $1.14 |
| 30% — high-volume aggregator | $0.25 | $0.08 | $0.17 (70%) | $2.27 |
Full re-scrape monthly cost at daily polling: $7.51. First month with incremental costs $0.63 / $1.35 / $2.44 for the 5% / 15% / 30% scenarios because the first run builds baseline state at full cost before incremental savings apply.
FAQ
How many results can I get from linkedin.com?
The number of results depends on the search query and available listings on linkedin.com. Use the maxResults parameter to control how many results are returned per run.
Does LinkedIn Jobs Scraper support recurring monitoring?
Yes. Enable incremental mode to only receive new or changed listings on subsequent runs. This is ideal for scheduled monitoring where you want to track changes over time without re-processing the full dataset.
Can I integrate LinkedIn Jobs Scraper with other apps?
Yes. LinkedIn Jobs Scraper works with Apify's integrations to connect with tools like Zapier, Make, Google Sheets, Slack, and more. You can also use webhooks to trigger actions when a run completes.
Can I use LinkedIn Jobs Scraper with the Apify API?
Yes. You can start runs, manage inputs, and retrieve results programmatically through the Apify API. Client libraries are available for JavaScript, Python, and other languages.
Can I use LinkedIn Jobs Scraper through an MCP Server?
Yes. Apify provides an MCP Server that lets AI assistants and agents call this actor directly. Use compact mode and descriptionMaxLength to keep payloads manageable for LLM context windows.
Is it legal to scrape linkedin.com?
This actor extracts publicly available data from linkedin.com. Web scraping of public information is generally considered legal, but you should always review the target site's terms of service and ensure your use case complies with applicable laws and regulations, including GDPR where relevant.
Your feedback
If you have questions, need a feature, or found a bug, please open an issue on the actor's page in Apify Console. Your feedback helps us improve.
You might also like
- Actiris Brussels Job Scraper — Scrape all active job listings from actiris.brussels — official Brussels public employment service..
- Adzuna Job Scraper — Global Jobs with Salary & Coordinates — Scrape adzuna.com job listings across 19 country markets with structured salary data.
- APEC.fr Scraper - French Executive Jobs — Scrape apec.fr - French executive job listings with salary ranges, company, location, skills,.
- Arbeitsagentur Scraper - German Jobs — Scrape arbeitsagentur.de - Germany’s official employment portal with 1M+ listings. Contact data,.
- Arbetsformedlingen Job Scraper — Scrape arbetsformedlingen.se (Platsbanken) — Sweden's official employment portal. Returns 84.
- AutoScout24 Scraper — Scrape autoscout24.com - Europe's largest used car marketplace with 770K+ listings. Structured.
- Bayt.com Scraper - Jobs from the Middle East — Scrape bayt.com - the leading Middle East job board. Salary data, experience requirements.
- Bilbasen Scraper - Denmark’s Car Marketplace — Scrape bilbasen.dk - Denmark’s largest car marketplace. Full vehicle specifications, seller.
Getting started with Apify
New to Apify? Create a free account with $5 credit — no credit card required.
- Sign up — $5 platform credit included
- Open this actor and configure your input
- Click Start — export results as JSON, CSV, or Excel
Need more later? See Apify pricing.