US Business Entity Filings — New LLC & Corp Registrations Daily
Pricing
Pay per usage
US Business Entity Filings — New LLC & Corp Registrations Daily
Scrape fresh LLC, Corp, and LP filings from US Secretary of State portals daily. NY, FL, CO, CT today — all 50 states coming. Filter by filing date to get new businesses the moment they register. For sales leads, KYC, compliance, and due diligence.
Pricing
Pay per usage
Rating
0.0
(0)
Developer
PaxIQ
Actor stats
0
Bookmarked
2
Total users
1
Monthly active users
3 days ago
Last modified
Categories
Share
Get fresh business filings (LLC, Corp, LP, LLP, etc.) from US Secretary of State portals the moment they're registered — ideal for sales prospecting, KYC, compliance monitoring, and competitive intelligence.
Why this actor? Every new business is a warm lead. Solar installers, insurance agents, accountants, lawyers, and SaaS companies all use new LLC registrations to find prospects before the competition does. Filter by filing date to pull only today's new registrations.
Coverage today: NY, FL, CO, CT (4 states, ~4,000+ new entities/day). All 50 states in progress.
Phase 1 States
| State | Source | Platform | Notes |
|---|---|---|---|
| NY | data.ny.gov | Socrata SODA | Daily filing data |
| CT | data.ct.gov | Socrata SODA | Business registry |
| CO | sos.state.co.us | HTML form scrape | Date-filter POST |
| FL | dos.fl.gov/sunbiz | JSON API | sunbiz.org |
Output Schema (14 fields)
| Field | Description |
|---|---|
entity_name | Business name |
entity_type | LLC, Corporation, LP, LLP, etc. |
state | 2-letter state code |
state_entity_id | State's internal ID/entity number |
filing_date | YYYY-MM-DD formation/filing date |
status | Active, Inactive, Dissolved, etc. |
registered_agent | Registered agent name |
principal_address | Principal business address |
agent_address | Registered agent address |
owners | Comma-separated officer/member names |
county | County if available |
source_platform | socrata / portal_scrape / fl_sunbiz |
source_url | Portal URL |
scraped_at | ISO UTC timestamp |
Apify Input
{"states": ["NY", "FL"],"startDate": "2025-01-01","endDate": "2025-01-31","maxResults": 10000,"socrataAppToken": "optional-token"}
| Field | Type | Default | Description |
|---|---|---|---|
states | array | all | 2-letter state codes to scrape |
startDate | string | yesterday | YYYY-MM-DD start date |
endDate | string | yesterday | YYYY-MM-DD end date |
maxResults | integer | 10000 | Max records per state |
socrataAppToken | string | — | Socrata app token (boosts rate limits for NY/CT) |
Local Development
# Install depspip install -r requirements.txt# Run all configured states (yesterday's filings)python src/main.py# Run specific statespython src/main.py --states NY CT --start 2025-01-01 --end 2025-01-31# Test individual scraperspython src/socrata_scraper.py --state NY --start 2025-01-01 --end 2025-01-02 --max 5python src/colorado_scraper.py --start 2025-01-01 --end 2025-01-02 --max 10python src/florida_scraper.py --start 2025-01-01 --end 2025-01-02 --max 10
File Structure
biz-filings/├── .actor/│ ├── actor.json Apify actor metadata + dataset schema│ └── input_schema.json Apify UI input form├── src/│ ├── main.py Async Apify entry point│ ├── router.py STATE_REGISTRY + scraper dispatch│ ├── normalize.py Raw → 14-field normalized schema│ ├── socrata_scraper.py NY + CT (Socrata SODA API)│ ├── colorado_scraper.py CO (HTML form POST + table parse)│ └── florida_scraper.py FL (sunbiz.org JSON API)├── Dockerfile apify/actor-python:3.11 base├── requirements.txt httpx, beautifulsoup4, apify└── README.md
Architecture
main.py└─ router.py (STATE_REGISTRY dispatch)├─ socrata_scraper.py → NY, CT├─ colorado_scraper.py → CO└─ florida_scraper.py → FL│▼ raw dictsnormalize.py → 14-field standard record│▼Apify dataset / output/XX_filings.json
Rate Limiting
- Between pages: 0.5s delay
- Between states: 2.0s delay
- Colorado detail pages: 0.3s delay per entity (fetch_details=True)
Adding New States (Phase 2)
- Add entry to
STATE_REGISTRYinrouter.py - Add a
fetch_filings()function in a new<state>_scraper.py - Add field aliases to
normalize.py'spick()chains - Test with
python src/<state>_scraper.py --start ... --end ...