for an academic research project I would like to have all available data on Crunchbase. Crunchbase is one of the largest collections of startups, their founders, their funding rounds, acquisitions, IPOs etc.
The data availability for each entry (organization, person...) varies. I attach some examples of API responses when querying an organization, a person, a fundinground.
All data need to be retrieved and put into separate tables (e.g., organization one table, persons one table, funding rounds one table, IPOs one table, acqusitions one table, products one table...) that are linked via keys. We need some own conceptual thoughts of the programmer here since data are very clustered in the API.
It is important to know that we would like to have all data but data are paginated in the responses and there are API limits. We have requested to remove the API limits for our case but no answer yet. So the developer should take into account the limits right now which are: 50 calls per minute, 2,500 calls per day and 25k calls per month. We need to integrate a possibility into the script to example run it from this to that entry when calling it so we can start it every day.
My preferred format will be an import to a mysql table but also csv files will be okay too. If we do mysql then we will need a dump of the queries to set up all the tables needed that can be used for the import.
The best case will be that you supply the script and setup queries for the mysql tables (if needed) on the one hand and a final dataset dump on the other hand so we only have to use the script at a later time if we want to update our dataset.
Please don't hesitate to contact me if you have any further questions.
25 freelancers estão ofertando em média $238 para este trabalho
Hi I work towards providing reliable, relevant and robust IT solutions at most competitive prices to my customers. I ensure 100% customer satisfaction so lets start Thanks
Hello. I can help you. I can do php script to call api and put it in to Mysql DB. I will setup to cron to be close to limit. I can do call script by proxy/or diffident ip to increase limit.