Hi,
I have built a tool which allows me to archive lots of content to a local server from e621, using the API. Getting this content within the API rate limits is easy as I can get the required metadata (tags, parent/children, description, hashes, filetypes etc) from just calling https://e621.net/posts.json
This has been really successful, however I now have over 80K pieces of data, and I know that a significant amount of this data will have changed (deletions from e621, tag changes, description changes, parents and children added/removed)
What would be the most efficient way to update this data whilst keeping in rate limits? Making over 80K API calls at 1 per second to keep within limits is obviously not realistic. It is not possible for me to retrieve all data in one big search (don't think so, anyway) as the data came from multiple different searches. I still have a record for what these searches are, so I could theoretically just run the full searches again and update data, but I wish to start adding individual records outside of searches too so this method will not work long term.
Thanks!