Email ✉️ & Phone ☎️ Extractor
7 days trial then $30.00/month - No credit card required now
Email ✉️ & Phone ☎️ Extractor
7 days trial then $30.00/month - No credit card required now
Extract emails, phone numbers, and other contact information like Twitter, LinkedIn, Instagram... from websites you provide. Best for lead generation and data enrichment. Export data in structured formats and dominate your outreach game. Capture your leads almost for free, fast, and without limits.
hey I have a list of websites for some of them I don't want to extract details can I ask to ignore some of the websites so it will skip them? you closed my issue but you didn't gave me the right answer - I want to write websites I want to exclude, not websites I want to include is it possible? thanks
There is a way to exclude some websites yes, using the pseudo urls with a negative regex. If you are not familiar with regex, then no there is no other way to do it, sorry.
You could still create two different task, one that can follow links for some websites, and the other Task with websites you don’t want any follow links.
Does it make sense?
Sorry if I closed too early the previous issue. You can reopen issues any time, it’s done for that.
Let me know if I can close this one once you are ok with my answers !
Have a good day, and enjoy the Olympics games :)
Thank you very much for the detailed explanation! Can you give me an example of a situation where I ask not to scan a certain website? Let's say for example that I don't want to scan the following domain - https://en.batumiexpert.com/ and the following domain - https://hayatestate.com/
Can you give me an example that I can use so that I can understand better? Thanks!
Sure. Using pseudo-Urls you can write something that works I believe.
Here is a small website helping you to write the proper regex https://regexr.com. In your case, you want to "match" anything but some websites (https://en.batumiexpert.com/ and https://hayatestate.com/) which leads to something like this :
^(?!.(batumiexpert|hayatestate/)).$
Note that in this regex, I only look after "batumiexpert" and "hayatestate" without the .com to make it simpler. But feel free to play around with regexr to try new things and add other websites. To add oter websites "|" is equivalent to OR in coding languages.
Hope it helps,
Let me know if you have any questions, otherwise I will close the issue :)
not working for me :( can you watch that please? https://www.loom.com/share/82617f65ac054235a816298d363ab7de
that's the sheet -
https://docs.google.com/spreadsheets/d/1ogKj1JNT7_5GiYvzKrWeZ-T5M6_LOBj2k4zZsS1UOpU/edit?gid=0#gid=0
I see you have mulitple questions. Let's finish this issue "is it possible to exclude some website for extract?" first, if you may. From what I see on your results, it seems to work because I do not see any url like "https://en.batumiexpert.com" or "https://hayatestate.com" in the results.
You may think it's because the crawler does not click on links. To make sure of that, you may want to increase to "total number of pages" to something like 1000 and see what it does.
Feel free to open another issue about the "null" problem you mention so that we can discuss this issue in a dedicated topic ! it will be much easier for you and me :)
I am closing this issue since it's been 25 days. Feel free to reopen it if necessary
Actor Metrics
190 monthly users
-
50 stars
>99% runs succeeded
3.3 hours response time
Created in Oct 2021
Modified 2 months ago