I currently use GitHub - ipfs/go-ds-s3: An s3 datastore implementation as the datastore for my IPFS. But there are too many requests (HeadObject requests) to the s3 bucket that results in significantly increasing in costs.
Any advice on this? Can we configure the frequency of this routine?
And in the worst case, we may need to no longer use s3 as datastore. So i am looking for an tutorial on how to migrate the data from s3 to local machine.
Sorry for mentioning you @hector, but do you have any idea on this? About how to decrease the number of requests to S3 or migrating data from s3 to local machines (flatfs datastore).
I think datastore.Has() is implemented via GetSize().
If I’m not mistaken however, usually the response to such requests should be cached. Increasing the sizes of the cache might be one way to reduce them:
(unfortunately only bloom filter size is configurable).
Otherwise, if the requests are very random for very random keys there is not much to do other than not using S3. If nodes are meant to provide content publicly, they need to check if they have it when requested.
The point of the screenshot is not the price, it is about the number of requests that happened in just a few hours of usage. This is a solid base to calculate the potential cost of the real-world example. Maybe @filebase can send real-world screenshot of their usage.