且构网

分享程序员开发的那些事...
且构网 - 分享程序员编程开发的那些事

针对大量微小文件优化S3下载

更新时间:2023-09-10 18:00:52

处理大量数据总是需要根据底层系统构建存储。

Dealing with a large number of data always needs architecting your storage with regards to underlying systems.

如果您需要高吞吐量,则需要对s3键进行分区,以便它可以容纳大量请求。分布式计算具有高性能的自身需求,这就是这样的需求。

If you need high throughputs you need to partition your s3 keys so that it can accommodate a high number of requests. Distributed computing comes with own needs to serve with high performance, this is one such need.

请求率注意事项:

https ://docs.aws.amazon.com/AmazonS3/latest/dev/request-rate-perf-considerations.html

转移加速:

https://docs.aws.amazon.com/AmazonS3/latest/dev/transfer-acceleration.html

如何提高吞吐量:

https://aws.amazon.com/premiumsupport/knowledge-center/s3-bucket-performance-improve/

希望它有所帮助。

EDIT1

我发现您正在尝试将文件下载到Ephemeral存储,您需要了解存储限制。这些不适用于批量处理。

I see that you are trying to download files to Ephemeral storage, you need to be aware of storage limits. Those are not meant for bulk processing.

https ://docs.aws.amazon.com/lambda/latest/dg/limits.html