Posted on: 10/29/15 01:46PM
jedi1357 said:
Fine. Just so you know, we are not really a big site.
I don't know how many resources we have but I wouldn't be surprised if it all fits into a single shelf at the server farm.
Fits into a couple units. We only have 2-3 servers total.
Downloading one image at a time isn't an issue. If that was enough to kill the site, the site would be long dead. If you were only downloading one image at a time, it probably wasn't you.
The issue started all the way back before the 14th, so if you only started developing your tool around the 25th, it wasn't you.
When I said that someone was hitting the site hard, I meant
hard.
Awakened said:
I wanted to inquire what is generally considered fair use of the API and of downloading stuff via third party applications.
I am developing a browser extension for keeping favorites, which benefits from getting detailed info about posts for better search, and i also have an implementation for local caching for the images, to reduce server load and improve performance.
How many requests per minute/hour/something are ok, and is it generally ok to download images programmatically for purpose of caching and/or archiving?
API requests don't really hit the site very hard since we use Solr for all tag searching. I'd say that an API pull every 5 seconds is pretty reasonable if it's something that's going to be continuous.
Please do not continuously archive the site. Our Terms of Service prohibits automatically downloading the site in such a way. However, caching images that the user has or is likely to load should be fine as long as you keep your number of simultaneous downloads low.
Just so you know, though, almost all of the load our server gets from loading images is from thumbnails. While it may speed up how fast it loads for the user, caching the actual images doesn't really provide any noticeable benefit to the site. Just throwing that out there since you said "to reduce server load".