Real Estate Technology
Market News & Data
General Info
Real Estate Strategies
![](http://bpimg.biggerpockets.com/assets/forums/sponsors/hospitable-deef083b895516ce26951b0ca48cf8f170861d742d4a4cb6cf5d19396b5eaac6.png)
Landlording & Rental Properties
Real Estate Professionals
Financial, Tax, & Legal
![](http://bpimg.biggerpockets.com/assets/forums/sponsors/equity_trust-2bcce80d03411a9e99a3cbcf4201c034562e18a3fc6eecd3fd22ecd5350c3aa5.avif)
![](http://bpimg.biggerpockets.com/assets/forums/sponsors/equity_1031_exchange-96bbcda3f8ad2d724c0ac759709c7e295979badd52e428240d6eaad5c8eff385.avif)
Real Estate Classifieds
Reviews & Feedback
Updated over 4 years ago on . Most recent reply
![Bryce DeCora's profile image](https://bpimg.biggerpockets.com/no_overlay/uploads/social_user/user_avatar/719938/1621496051-avatar-bryced8.jpg?twic=v1/output=image/crop=365x365@185x201/cover=128x128&v=2)
Washington Data Scraping API - Anyone Else Doing This?
My original plan was to invest solely in Snohomish and King County Washington. I wanted my custom CRM (built with Podio) to have the ability to auto-populate property data found on the county assessor site. This drove me to create an API endpoint written in Python that would do this on demand. I've met others who build code like this and store the data in tables to be queried later, but then you're stuck with older data.
Is anyone else using property scraping solutions on a large or small scale? As my company grows, I think I will start using a company like ATTOM Data to provide property data so I don't have to continue to coordinate the build and maintenance of scraping tools. Fellow nerds let me know what you're up to!
![](https://assets0.biggerpockets.com/uploads/uploaded_images/normal_1587489378-image.png)
Most Popular Reply
Yep... I've got bs full of assessor/auditor/recorder data. To keep from having older data, you use "insert on duplicate key update" queries and continually loop over all the sources data..... i created a threaded version of rolling curl in php that keeps about 150 curl threads running async on targets which collects up to about 230 html pages per second.. which pulls down every property large counties like dekalb pretty fast...
Confused though.... What did you create the API endpoint for? and where? isn't your script scraping and writing to Podio API?
If ATTOM has you county that works... Plenty of counties are missing though. custom scrape is the only way to get everything for those....