
Contact Details Scraper
Pricing
Pay per event

Contact Details Scraper
Free email extractor and lead scraper to extract and download emails, phone numbers, Facebook, Twitter, LinkedIn, and Instagram profiles from any website. Extract contact information at scale from lists of URLs and download the data as Excel, CSV, JSON, HTML, and XML.
4.0 (26)
Pricing
Pay per event
413
Total users
30.7k
Monthly users
1.5k
Runs succeeded
>99%
Response time
3 days
Last modified
22 days ago
My url volume increased by 5
Closed
Recently I uploaded 256 urls for remail retrieval. The output was 1250 urls. I am aware of the data and have been working with it. Some was previously deleted but Im not sure about all. How do i prevent this from happening again?
Hi, thank you for reaching out! If you’d like the Actor to only process the exact URLs you've uploaded, without generating or following additional links, you can adjust a few parameters in your input. maxDepth
set to 0 will ensure the Actor only scrapes the URLs you provide without following any links. sameDomain
set to false prevents the Actor from crawling other pages within the same domain. maxRequestsPerStartUrl
set to 1 limits the Actor to scrape only one page per input URL. Is this the desired outcome?
nacreous_theremin
Thanks Andrea. Im not familiar with parameters. I cant actually find what you are talking about.
Hey, it's just your input settings. Try adjusting them like this in your JSON input (I'm not listing all of the URLs here, you can include all of them of course):
1{ 2 "considerChildFrames": true, 3 "maxDepth": 0, 4 "maxRequests": 9999999, 5 "maxRequestsPerStartUrl": 1, 6 "sameDomain": false, 7 "startUrls": [ 8 { 9 "url": "http://www.qualitybuilders.com.au", 10 "method": "GET" 11 }, 12 { 13 "url": "http://www.integratedss.com.au", 14 "method": "GET" 15 } 16 ], 17 "useBrowser": false, 18 "waitUntil": "domcontentloaded", 19 "proxyConfig": { 20 "useApifyProxy": true 21 } 22}
nacreous_theremin
Thanks - I just input the data. Works much better.