Properties Scraper D1 Private
3 days trial then $60.00/month - No credit card required now
Properties Scraper D1 Private
3 days trial then $60.00/month - No credit card required now
Scrape crawl millions of sale/rent real estate properties. Our real estate scraper also lets you monitor specific listing for new updates/listing. You can provide multiple search result listings to scrape/monitor.
🏡 What is Real Estate Properties Scraper?
This properties Scraper will enable you scrape any sale/rent listing.
You can simply take your listing url from browser and enter it into this actor. This actor will crawl through all pages of particular listing and generate dataset for you.
Listing url is something you get when you perform the search on site.
🚪 What can this Scraper do?
📈 Extract market data listings
👀 This actor is not just scraper but also has monitoring capability. You can turn on monitoring mode and it will give you only newly added properties compared to your previous scrapes.
📩 This actor also helps yu to identify which properties are not listed anymore. Please refer to Identifying delisted properties
⬇️ Download real estate data in Excel, CSV, JSON, and other formats
📚 How do I start scraping with this scraper?
- Register for your free Apify account here
- You don't need to provide your credit card details for free acount. Just click on "Get Started" button on above link and complete the registration only.
- Free account comes with reasonable credits to try out this actor. This actor also comes with free trial of 3 days without any commitment/upfront charge.
- Run this actor and verify the scraped data. Apify has huge integration possibilities. You can download the data or push the data into any 3rd party platform directly.
🌳 What data can I extract using this tool?
📝 | 📝 |
---|---|
Listing Title | Full Address |
Listing URL | ReferenceNo |
reraDedLicenceNo | Completion Status |
Bathrooms | Bedrooms |
Agent Name | Agent Phone |
Listing Type | Property Type |
Latitude | Longitude |
Completion | Agency Email |
Text Description | Formatted HTML Description |
Amenities | Images |
Price | Size |
Furnishing | Listing Date |
⬇️ Input
For simple usecase, you just need to provide browser url search result page & that's all. You can leave other fields as they are to be sensible defaults.
Input example
1{ 2 "listUrls": [ 3 { 4 "url": "" 5 } 6 ], 7 "propertyUrls": [ 8 { 9 "url": "" 10 } 11 ], 12 "fullScrape": true, 13 "monitoringMode": false, 14 "includePriceHistory": false, 15 "enableDelistingTracker" : false 16}
You can either provide listUrls
to search properties from or provide propertyUrls
directly to crawl.
Understading monitoring mode :
-
fullScrape
: This option is by default turned on. When enabled it always force actor to scrape complete listing from all pagination pages regardless of monitoring is enabled or not. -
monitoringMode
: This option when turned on will only scrape newly added property listings compared to previously scraped properties by this actor. It's important to turn off fullScrape setting if you are using this mode. If you keep fullScrape on, it will re-scrape complete listing again. -
includePriceHistory
: This option when turned on will also scrape procie history of given property when available. This may affect the speed of scraping considerably. Please turn it on only if you need this data. -
enableDelistingTracker
: This option when turned on will start tracking date against each property under Apify Key Value store. This KV store can be queried later to find out which properties are delisted.
⬆️ Output
The scraped data is stored in the dataset of each run. The data can be viewed or downloaded in many popular formats, such as JSON, CSV, Excel, XML, RSS, and HTML.
Output example
The result for scraping a single property like this:
1{ 2 "title": "Super Deluxe Villa for sale in Helwan \\ corner on two streets, the second piece of the main street", 3 "url": "", 4 "images": [], 5 "price": 3000000, 6 "bedrooms": 5, 7 "bathrooms": 8, 8 "size": 8000, 9 "coordinates": { 10 "lat": 25.348499298096, 11 "lng": 55.416801452637 12 }, 13 "location": "Al Ghubaiba, Sharjah, UAE", 14 "type": "sale", 15 "description": "Super Deluxe Villa for sale in Helwan area in Sharjah <br> The land area is 8000 square feet, a corner on two streets, the second piece of the main street <br> The building area is 4600 square feet<br> The villa consists of:<br>5 master rooms, each room has a private bathroom with a dressing <br>4 on the top floor and one room on the ground floor <br>Two halls <br>big sitting room <br>A kitchen with two doors, internal and external. <br>A maid's room and an ironing room have an external door <br>The materials used in the villa are first class <br>Air conditioning duct general <br>Interlock from Arabic German does not change color <br><br>3 million dirhams required", 16 "isVerified": false, 17 "hasDLDHistory": false, 18 "reraPermitUrl": null, 19 "addedOn": "November 6, 2023", 20 "agent": "mohamed", 21 "agencyName": "Al Wasl Estste", 22 "agencyEmail": "alwaslrealestate112@gmail.com", 23 "amenities": [ 24 "Maids Room", 25 "Central A/C & Heating", 26 "Balcony", 27 "Private Garden", 28 "Pets Allowed", 29 "Double Glazed Windows", 30 "Laundry Room", 31 "Broadband Internet", 32 "Satellite / Cable TV", 33 "Maintenance Staff", 34 "Storage Areas", 35 "Waste Disposal" 36 ], 37 "propertyType": "Villa", 38 "purpose": "Sale", 39 "furnished": "Unfurnished", 40 "updatedAt": "1699994557", 41 "completionStatus": "Ready", 42 "reraDedLicenceNo": "511946", 43 "propertyReference": "9873-L0fN5K", 44 "id": "2023-11-5-super-deluxe-villa-for-sale-in-helwan-corn-12-569" 45}
❓Limitations
Since allows only 80000 properties per listing/search result, you might want to break down your listing urls into smaller area if it has more than 80K results. Good News is that even if multiple list urls contains overlapping results, they will get deduplicated within same run data.
🔎 Identifying delisted properties
This actor provides you monitoring mode configuration using which you can get only incremental updates about newly added properties. In case, you also want to identify which properties are delisted from platform, you can use any of the following techniques with the help of this actor.
-
Running Always in full scraper mode : Run this actor always in full scrape mode and cross check the new incoming batch of data with your existing database. If any property that exists in yoru database but not in newly scraped data batch, that means it's not listed anymore
-
Use Key Value Store generated by scraper : If your are monitoring very large batch of data and you don't want to scrape everything all the time, this method involves bit of technicality but achieves the goal efectively. Apify has storage feature called Key-value store. When you run this scrape, this scraper stores every single property in key value store along with timestamp in store. Inside this store, key is property id itself and value is timestamp like this
{ lastSeen : '2023-11-02T05:59:25.763Z'}
Whenever you run this scraper, it will update the timestamp against particular id if it finds property on the platform. e.g. if we have 2 proprties with id
prop1
andprop2
and we scraped them both on November 1, key value storage would look like this :1prop1 -> { lastSeen : '2023-11-01T05:59:25.763Z'} 2prop2 -> { lastSeen : '2023-11-01T05:59:25.763Z'}
Now if you run this scraper again on December 1 and prop1 is not on the platform anymore but prop2 is still there, key value storage would change like this :
1prop1 -> { lastSeen : '2023-11-01T05:59:25.763Z'} 2prop2 -> { lastSeen : '2023-12-01T05:59:25.763Z'}
That means if any property has
lastSeen
less than latest batch you loaded, that property is delisted now. You can directly iterate through whole Key value storage using Apify key value storage API to identify this. Please refer to this API documentation to do the same. Please remember store name generated by this scrape will be.Alternatively, you can iterate through your existing database active properties and use this API to identify listing status.
For this approach to work, it's important that you enable this feature via
enableDelistingTracker
(Enable Delisting tracker) input.
🙋♀️ For custom solutions
In case you need some custom solution, you can contact me : dhrumil@techvasu.com
Or learn more about me on github : https://github.com/dhrumil4u360
Actor Metrics
2 monthly users
-
1 star
97% runs succeeded
Created in Jan 2024
Modified 9 days ago