[GH-ISSUE #1543] Run integration tests with larger files #811

Open
opened 2026-03-04 01:48:59 +03:00 by kerem · 4 comments
Owner

Originally created by @gaul on GitHub (Feb 2, 2021).
Original GitHub issue: https://github.com/s3fs-fuse/s3fs-fuse/issues/1543

s3fs sets BIG_FILE_LENGTH to 25 MB which is large enough to trigger multipart upload but lower than other limits like the single file upload size. This limit also makes sense when using S3Proxy's transient backend which store objects in-memory. When apache/jclouds#96 merges, s3fs can instead use the filesystem backend which stores objects on-disk. This will allow the integration tests to use larger objects, e.g., > 5 GB single-part objects. Related to #1542.

Originally created by @gaul on GitHub (Feb 2, 2021). Original GitHub issue: https://github.com/s3fs-fuse/s3fs-fuse/issues/1543 s3fs sets `BIG_FILE_LENGTH` to 25 MB which is large enough to trigger multipart upload but lower than other limits like the single file upload size. This limit also makes sense when using S3Proxy's transient backend which store objects in-memory. When apache/jclouds#96 merges, s3fs can instead use the filesystem backend which stores objects on-disk. This will allow the integration tests to use larger objects, e.g., > 5 GB single-part objects. Related to #1542.
Author
Owner

@gaul commented on GitHub (Feb 4, 2021):

With larger files reading from /dev/urandom becomes a bottleneck. We may want to create our own helper which makes pseudo-random data faster.

<!-- gh-comment-id:773272077 --> @gaul commented on GitHub (Feb 4, 2021): With larger files reading from `/dev/urandom` becomes a bottleneck. We may want to create our own helper which makes pseudo-random data faster.
Author
Owner

@gaul commented on GitHub (Feb 4, 2021):

To use the filesystem blobstore to test larger files we will need a new release of S3Proxy to include the jclouds fix.

<!-- gh-comment-id:773374947 --> @gaul commented on GitHub (Feb 4, 2021): To use the filesystem blobstore to test larger files we will need a new release of S3Proxy to include the jclouds fix.
Author
Owner

@gaul commented on GitHub (Feb 7, 2021):

Testing with larger files is proving to be a rich source of bugs. Some tests require files as large as 11 GB to trigger various conditions which takes up to 15 minutes to run on my desktop. We could add some flags to s3fs and S3Proxy to use smaller thresholds to trigger conditions like copy part range more easily.

<!-- gh-comment-id:774674153 --> @gaul commented on GitHub (Feb 7, 2021): Testing with larger files is proving to be a rich source of bugs. Some tests require files as large as 11 GB to trigger various conditions which takes up to 15 minutes to run on my desktop. We could add some flags to s3fs and S3Proxy to use smaller thresholds to trigger conditions like copy part range more easily.
Author
Owner

@gaul commented on GitHub (May 30, 2021):

Related to #1665, currently blocked on macOS Java xattr compatibility.

<!-- gh-comment-id:850919758 --> @gaul commented on GitHub (May 30, 2021): Related to #1665, currently blocked on macOS Java xattr compatibility.
Sign in to join this conversation.
No milestone
No project
No assignees
1 participant
Notifications
Due date
The due date is invalid or out of range. Please use the format "yyyy-mm-dd".

No due date set.

Dependencies

No dependencies set.

Reference
starred/s3fs-fuse#811
No description provided.