Lagou Tech Jobs Scraper (拉勾网) avatar

Lagou Tech Jobs Scraper (拉勾网)

Under maintenance

Pricing

from $3.50 / 1,000 results

Go to Apify Store
Lagou Tech Jobs Scraper (拉勾网)

Lagou Tech Jobs Scraper (拉勾网)

Under maintenance

Extract thousands of tech job listings from Lagou.com (拉勾网), China's largest IT recruitment platform. Scrape salary ranges, tech stacks, company details, funding stages, and more from ByteDance, Alibaba, Tencent, Baidu, and 100,000+ Chinese tech companies. No browser needed — fast, cheap, scalable.

Pricing

from $3.50 / 1,000 results

Rating

0.0

(0)

Developer

Logiover

Logiover

Maintained by Community

Actor stats

0

Bookmarked

2

Total users

1

Monthly active users

3 days ago

Last modified

Share

Lagou Tech Jobs Scraper (拉勾网) — China's #1 Tech Job Data Extractor

Extract thousands of tech job listings from Lagou.com (拉勾网), China's largest and most popular technology-focused recruitment platform. Get structured data on salaries, tech stacks, company funding stages, and more from top Chinese tech companies including ByteDance (字节跳动), Alibaba (阿里巴巴), Tencent (腾讯), Baidu (百度), Meituan (美团), JD.com (京东), Xiaomi (小米), Huawei (华为), Pinduoduo (拼多多), NetEase (网易), Didi (滴滴), Kuaishou (快手) and thousands more.

Why This Scraper?

Lagou.com is the dominant job board for China's tech industry with 500,000+ active listings across all major Chinese cities. Unlike generic Chinese job sites (Zhaopin, 51Job, BOSS Zhipin), Lagou focuses exclusively on technology, software engineering, product management, data science, AI/ML, and digital roles — making the data highly targeted and valuable.

This scraper delivers structured, clean JSON data ready for analysis, research, dashboards, and integration — no manual copying, no browser automation, no headless browsers.

Key Features

  • Fast & Cheap — Direct HTTP extraction, no browser needed. Runs on minimal memory (128 MB). ~$3.00 per 1,000 jobs
  • Massive Scale — Collects thousands of unique jobs per run using intelligent combo expansion (keywords × cities × experience filters)
  • Smart Deduplication — Built-in position ID tracking ensures zero duplicate records across all search combinations
  • Parallel Processing — Configurable concurrency (up to 10 parallel workers) for maximum throughput
  • Experience Filter Expansion — Automatically multiplies searches by 6 experience levels, surfacing jobs hidden behind filters
  • 25+ Chinese Cities — Pre-mapped support for all major tech hubs with English or Chinese city name input
  • Salary Parsing — Automatically extracts numeric min/max salary from Chinese formats (15k-25k, 1万-2万)
  • Rich Company Data — Funding stage, employee count, industry classification, company logo for every listing
  • Proxy-Ready — Designed for residential proxy rotation; each request uses a fresh IP

Output Data Fields (27 Fields)

Every scraped job includes the following structured fields:

FieldTypeExampleDescription
jobTitlestringPython开发工程师Full position title
jobUrlstringhttps://www.lagou.com/wn/jobs/123.htmlDirect link to listing
companyNamestring微梦创科网络科技(中国)有限公司Official company name
companyShortNamestring微博Company brand name
companyUrlstringhttps://www.lagou.com/gongsi/5832.htmlCompany profile link
salarystring20k-40kRaw salary string
salaryMininteger20Parsed minimum monthly salary (thousands RMB)
salaryMaxinteger40Parsed maximum monthly salary (thousands RMB)
citystring北京Job city
districtstring海淀区City district
workYearstring1-3年Required experience
educationstring本科Minimum education
jobNaturestring全职Full-time / Part-time
jobTypestring开发|测试|运维类Job category
skillsstringPython, MySQL, Redis, DockerRequired tech stack
companySizestring2000人以上Company size range
industryFieldstring社交媒体Industry sector
financeStagestring上市公司Funding stage (IPO, Series A-D, etc.)
companyLogostringhttps://...jpgCompany logo URL
jobDescriptionstring(optional)Full job description text
publishTimestring2026-04-18 15:13:30Listing publish timestamp
positionIdinteger12196627Lagou internal job ID
companyIdinteger5832Lagou internal company ID
searchKeywordstringPythonKeyword that found this job
searchCitystring北京City filter used
pageNumberinteger1Result page number
scrapedAtstring2026-04-18T10:04:51ZScrape timestamp (ISO)

Use Cases

  • Salary Benchmarking — Compare Python vs Java vs Go developer salaries across Beijing, Shanghai, Shenzhen, Hangzhou, and other cities
  • Tech Stack Analysis — Identify the most in-demand programming languages, frameworks, and tools in China's tech market
  • Hiring Trend Monitoring — Track which companies are hiring aggressively, which sectors are growing
  • Company Research — Map the Chinese tech ecosystem by funding stage, size, industry, and hiring volume
  • Recruitment Intelligence — Build candidate sourcing pipelines with targeted job market data
  • Academic Research — Study China's technology labor market, regional tech hub development, wage inequality
  • Market Entry Analysis — Understand local tech talent landscape before entering the Chinese market
  • Competitive Intelligence — Monitor competitors' hiring patterns and tech stack choices
  • Investment Due Diligence — Assess company growth through hiring velocity and role types
  • Data Journalism — Create stories about China's tech industry trends backed by hard data

Supported Cities (25+)

All major Chinese tech hubs are supported with English or Chinese input:

EnglishChineseTech Hub Rank
beijing北京#1 — AI, fintech, ByteDance HQ
shanghai上海#2 — Finance-tech, gaming
shenzhen深圳#3 — Hardware, Tencent/Huawei HQ
hangzhou杭州#4 — E-commerce, Alibaba HQ
guangzhou广州#5 — Internet services
chengdu成都#6 — Gaming, growing startup hub
nanjing南京#7 — Enterprise software
wuhan武汉#8 — Optics, semiconductor
xian西安#9 — Aerospace, emerging tech
suzhou苏州#10 — Biotech, manufacturing-tech
changsha长沙tianjin, chongqing, xiamen
fuzhou福州hefei, dongguan, jinan
kunming昆明zhuhai, ningbo, wuxi
dalian大连qingdao, zhengzhou, and more

Input Configuration

{
"keywords": ["Python", "Java", "React", "Go", "AI"],
"cities": ["beijing", "shanghai", "shenzhen", "hangzhou", "chengdu"],
"expandFilters": true,
"maxJobsTotal": 0,
"maxConcurrency": 3,
"requestDelay": 1000,
"proxyConfiguration": {
"useApifyProxy": true,
"apifyProxyGroups": ["RESIDENTIAL"]
}
}

Input Parameters

ParameterTypeDefaultDescription
keywordsstring[]requiredSearch keywords (Python, Java, React, Go, AI, DevOps, etc.)
citiesstring[]["beijing"]Cities to search — English or Chinese names accepted
expandFiltersbooleantrueMultiply searches by 6 experience levels for 6x more unique results
maxJobsTotalinteger0Global job limit (0 = unlimited)
maxConcurrencyinteger3Parallel workers (1-10). Higher = faster, more proxy usage
requestDelayinteger1000Milliseconds between requests
proxyConfigurationobjectResidentialProxy settings. Residential proxies strongly recommended

How It Works

  1. Combo Generation — Builds all keyword × city × experience filter combinations
  2. Parallel Fetching — Processes multiple combos simultaneously with configurable concurrency
  3. Data Extraction — Parses Lagou's server-rendered JSON (__NEXT_DATA__) from each page
  4. Smart Retry — Failed pages get 4 retries with fresh proxy IPs automatically
  5. Deduplication — Position IDs track across all combos, ensuring zero duplicates
  6. Dataset Push — Clean structured JSON pushed to Apify Dataset in real-time

Scale & Performance

ConfigurationCombosEst. Unique JobsEst. TimeEst. Cost
2 keywords × 3 cities36500-8003-5 min~$0.15
5 keywords × 5 cities1502,000-3,50010-15 min~$0.50
10 keywords × 10 cities6005,000-10,00030-45 min~$2.00
20 keywords × 25 cities3,00015,000-30,0002-3 hours~$8.00

Costs include Apify platform + residential proxy usage. Actual unique counts depend on job market overlap between keywords.

Keyword Ideas for Maximum Coverage

Programming Languages: Python, Java, JavaScript, TypeScript, Go, Rust, C++, PHP, Swift, Kotlin, Scala, Ruby

Frameworks & Tools: React, Vue, Angular, Spring, Django, Flask, Node.js, Docker, Kubernetes, Spark, Hadoop, TensorFlow, PyTorch

Roles: 前端 (Frontend), 后端 (Backend), 全栈 (Fullstack), 算法 (Algorithm), 数据 (Data), 产品经理 (Product Manager), 测试 (QA), 运维 (DevOps), 架构师 (Architect)

Domains: AI, 机器学习 (Machine Learning), 大数据 (Big Data), 云计算 (Cloud), 区块链 (Blockchain), 物联网 (IoT), 自动驾驶 (Autonomous Driving)

Integration & Export

Output is available in all Apify-supported formats:

  • JSON / CSV / Excel — Direct download from Apify Console
  • API — RESTful access to dataset via Apify API
  • Webhook — Trigger downstream pipelines on completion
  • Google Sheets — Auto-sync with Google Sheets integration
  • Amazon S3 / Google Cloud — Push to cloud storage
  • Zapier / Make — Connect to 5,000+ apps
  • Custom — Use Apify SDK in Python, JavaScript, or any HTTP client

Frequently Asked Questions

Q: Why only ~45 jobs per keyword+city combination? Lagou's platform limits search results to 3 pages (45 jobs) per query. The scraper compensates by using experience filter expansion, which creates 6 separate searches per keyword+city — each returning different job sets. This is why expandFilters: true is recommended.

Q: Do I need residential proxies? Yes, strongly recommended. Lagou is a Chinese platform and may block datacenter IP ranges. Apify's residential proxy group provides reliable access.

Q: Can I use Chinese keywords? Absolutely. Both English (Python) and Chinese (Python开发工程师, 前端工程师, 算法) keywords work. Chinese keywords often surface different results.

Q: How often should I run this? Job listings on Lagou refresh frequently. For market monitoring, weekly runs are ideal. For comprehensive datasets, monthly runs with broad keyword coverage work well.

Q: Is this legal? This scraper accesses only publicly available job listing data displayed on Lagou.com. No login, authentication, or private data is accessed. Please review Lagou's terms of service and your local regulations regarding web scraping.

Changelog

  • v1.0 — Initial release. HTML + NEXT_DATA extraction, parallel processing, experience filter expansion, 25+ city support, salary parsing, smart deduplication.

Support

Found a bug or need a custom feature? Open an issue on this actor's page or contact the developer directly through Apify Console.


Built for researchers, recruiters, analysts, and developers who need structured Chinese tech job market data at scale.