thebluebook scraper avatar

thebluebook scraper

Pricing

from $1.00 / 1,000 results

Go to Apify Store
thebluebook scraper

thebluebook scraper

Our thebluebook.com scraper makes it simple to collect contractor profiles at scale. It automatically gathers URLs from all search pages and extracts complete details for every profile including company info, contacts, trades, certifications, and project history.

Pricing

from $1.00 / 1,000 results

Rating

0.0

(0)

Developer

youssef farhan

youssef farhan

Maintained by Community

Actor stats

1

Bookmarked

3

Total users

1

Monthly active users

4 days ago

Last modified

Share

Blue Book Construction Directory Scraper — Extract Contractor Profiles at Scale

Scrape contractor and subcontractor profiles from thebluebook.com — the construction industry's largest online directory — and get structured company data including contacts, trade categories, certifications, project history, and email addresses sourced from company websites. Built for lead generation, market research, and construction intelligence workflows.

What you get

Company identity

  • name — Company name
  • company_id — Unique Blue Book ID
  • profile_link — Direct URL to Blue Book profile
  • city_state — City and state from profile header
  • full_address — Full street address (scraped from contacts sub-page)
  • website — Company website URL
  • scrape_date — ISO date the record was collected

Contact data

  • phone — Main office phone
  • email — Best contact email scraped from company website (domain-match prioritized over free providers)
  • contact_name / contact_role / contact_phone — Primary key contact (promoted to top level for easy access)
  • contacts[] — Full list of key contacts with name, role, and direct phone

Business profile

  • trade[] — Trade categories (e.g. Electrical Contractors, Plumbing)
  • service_area[] — Counties and regions serviced
  • certifications[] — Diversity certifications: MBE, WBE, DBE, SBE, etc.
  • established — Year founded
  • company_size — Employee headcount range
  • annual_volume — Annual revenue range
  • listed_since — Year first listed on Blue Book

Project history

  • projects[] — Construction projects with name, type, location, status, date, general contractor name and role

Sample output

{
"profile_link": "https://www.thebluebook.com/iProView/1424972",
"company_id": "1424972",
"name": "ABC Electrical Contractors",
"phone": "(713) 555-0100",
"email": "info@abcelectrical.com",
"website": "https://www.abcelectrical.com",
"city_state": "Houston, TX",
"full_address": "1234 Main St, Houston, TX 77002",
"trade": ["Electrical Contractors", "Lighting Contractors"],
"certifications": ["MBE", "DBE"],
"service_area": ["Harris County", "Fort Bend County", "Montgomery County"],
"established": "1998",
"company_size": "10-24 Employees",
"annual_volume": "$1M - $5M",
"listed_since": "2005",
"scrape_date": "2026-04-10",
"contact_name": "John Smith",
"contact_role": "Owners, Principals & Senior Executives",
"contact_phone": "(713) 555-0101",
"contacts": [
{ "name": "John Smith", "role": "Owners, Principals & Senior Executives", "phone": "(713) 555-0101" }
],
"projects": [
{
"project_name": "Downtown Office Tower",
"project_location": "Houston, TX",
"project_type": "Commercial",
"project_status": "Completed",
"project_date": "Mar 2023",
"gc_role": "General Contractor",
"gc_name": "Turner Construction",
"project_url": "https://www.thebluebook.com/iProView/1424972/project/..."
}
]
}

Use cases

  • Construction material suppliers building targeted outreach lists by trade and region
  • Staffing and recruitment agencies sourcing contractor companies with headcount and revenue data
  • Market research firms benchmarking contractor density, certifications, and project volume by metro
  • CRM and sales teams enriching leads with verified phone, email, address, and project history
  • Government and compliance teams identifying MBE/WBE/DBE-certified contractors for procurement
  • Proptech and construction intelligence platforms aggregating subcontractor data at scale

Pricing

EventPrice
Per company profile scraped$0.001

Real example: Scrape 1000 contractor profiles ≈ $1

How it works

  • Input: Provide one or more thebluebook.com search result URLs (e.g. electricians in Texas) or direct profile URLs
  • Pagination: Detects total page count automatically and fetches all pages concurrently (up to 5 parallel search requests)
  • Profile extraction: For each company, fetches the main profile + /locations-contacts/ + /construction-projects/ sub-pages in parallel — 3 pages per company
  • Email discovery: Crawls the company's own website (homepage + contact/about pages) to find and rank email addresses — domain-matching emails returned first
  • Output: Structured JSON records delivered to Apify Dataset, downloadable as JSON, CSV, or Excel — or pushed to your webhook in real time
  • Resumable: State is saved after every page and every profile — interrupted runs pick up where they left off

Why this scraper

  • Email included — Most Blue Book scrapers return empty email fields. This one crawls each company's own website and returns the best email found, prioritizing domain-matched addresses over generic inboxes.
  • Three sub-pages per profile — Contacts page (full address + key people) and projects page fetched concurrently with the main profile, not skipped.
  • Two-tier proxy strategy — Residential proxies on search/pagination pages where bot detection is tightest; standard proxies on profile pages to minimize cost. No proxy setup needed on your end.
  • Pay-per-profile, not per run — Idle time, retries, and failed requests don't cost you anything. You pay only for successfully scraped profiles.

Input example

{
"startUrls": [
{ "url": "https://www.thebluebook.com/iSearch/results/tx/houston/electrical-contractors/sc/261/" },
{ "url": "https://www.thebluebook.com/iSearch/results/ca/los-angeles/general-contractors/sc/240/" }
],
"maxItems": 500
}

To scrape a single company, pass its profile URL directly:

{
"startUrls": [
{ "url": "https://www.thebluebook.com/iProView/1424972" }
]
}

FAQ

Does it handle pagination automatically? Yes. The actor detects total page count from the search results page and fetches all pages without any additional input from you.

What output formats are supported? JSON, CSV, Excel, XML, and JSONL — all available from the Apify Dataset UI or via API.

How fresh is the data? Data is scraped live on each run. Schedule the actor daily, weekly, or monthly via Apify's built-in scheduler to keep your dataset current.

How does it find email addresses? It fetches the company's website (homepage + /contact, /contact-us, /about variants) and extracts emails from mailto links and page text. Domain-matched emails (e.g. info@companysite.com) are returned before Gmail or Yahoo addresses.

Can I run it on a schedule or trigger it via webhook? Yes to both. Use Apify's scheduler for recurring runs, or trigger via webhook/API on any external event.

Can I scrape a specific trade or region only? Yes — filter by constructing the right thebluebook.com search URL for your target trade category and location, then pass it as a startUrl.

What if a company has no website or contacts listed? The email, full_address, and contacts fields will be empty strings or empty arrays. The core company record is always returned.

Use via API or MCP

Call this actor programmatically via the Apify API or as an MCP server for AI agents (Claude, ChatGPT, Cursor, and others):

https://mcp.apify.com/actors/fayoussef/thebluebook-scraper

AI agents can trigger runs, pass input, and retrieve structured output directly — no manual steps required.

Need a custom scraper?

Need a different site, additional fields, or a managed data pipeline? Visit automationbyexperts.com for custom builds, retainers, and data-as-a-service.