
seloger mass products scraper (by search URL) ⚡
3 days trial then $25.00/month - No credit card required now

seloger mass products scraper (by search URL) ⚡
3 days trial then $25.00/month - No credit card required now
🔥Très simple! Entrez le lien vers la page de recherche et obtenir les résultats! ⚡ Extraire rapidement les infos détaillées sur les propriétés ( titre, description, photos, évaluations énergétique prix, contacts, transport et plus encore) à faible coût, avec exportation en JSON, CSV, HTML, EXCEL...
Actor Metrics
16 Monthly users
5.0 / 5 (1)
3 bookmarks
99% runs succeeded
2 hours response time
Created in Jul 2024
Modified 3 days ago
Errors during crawling
Hello, it's seems a lot of request now failed and structure seems change. Can you check and tell us if we must unsubscribe or you plan quick update ?
Thanks
Hi! Thanks for the feedback. Structure didn't change. Could you please share the input URL? (It's datadome, their bot protection who might have updated their algorithm and I adjusted this actor). I'll increase the retries so failed request will have higher chances of going through
xo7
If this can help you, everything works in build 0.0.155, but in the latest (0.0.163) I don't retrieve the same content (some request failed & exported field doesn't work). If you want reproduce you can check with my input url : "https://www.seloger.com/list.htm?projects=2,5&types=2,12,11,1&natures=1,2,4&places=[{%22subDivisions%22:[%2275%22]}]&surface=NaN/45&mandatorycommodities=0&enterprise=0&qsVersion=1.0&m=search_refine-redirection-search_results"
xo7
by example before in description structure I expect : "description": { "description": "...", "priceUnit": "€", "condoProperties": 33, "condoAnnualCharges": 2107, "classifiedDescription": "...", "aboutCoOwnership": "..." },
now I have directly a truncated string (in description)
Well, you're absolutely right. For structure, I updated this to scrape items from the search pages without going deep down to details page to make it more efficient and lightweight hence much faster. I'll revert that in few minutes (the scraper will do deep into each individual item's page). Will let you know once done so you confirm it's OK, then I'll look into the failing requests
Updated: structure is now back to what you are describing above! Looking into failed requests..
xo7
ok thank you in fact the 0.0.155 version now fail with : 2024-12-10T23:47:39.540Z /usr/src/app/node_modules/ow/dist/index.js:36 2024-12-10T23:47:39.542Z (0, test_1.default)(value, labelOrPredicate, predicate); so I will test again with your new version with deep request
thanks
Thanks, I've enabled the log debugging and I can see it's datadome. Please remove the file you attached. (Do you confirm the structure is OK now?)
xo7
I confirm the structure is OK now
xo7
Thanks for you feedback (and fix), everything seems to be fine now.