mirror of
https://github.com/s3fs-fuse/s3fs-fuse.git
synced 2026-04-25 13:26:00 +03:00
[GH-ISSUE #821] Modify s3fs default behaviour #477
Labels
No labels
bug
bug
dataloss
duplicate
enhancement
feature request
help wanted
invalid
need info
performance
pull-request
question
question
testing
No milestone
No project
No assignees
1 participant
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference
starred/s3fs-fuse#477
Loading…
Add table
Add a link
Reference in a new issue
No description provided.
Delete branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @aurorak on GitHub (Sep 13, 2018).
Original GitHub issue: https://github.com/s3fs-fuse/s3fs-fuse/issues/821
Additional Information
The following information is very important in order to help us to help you. Omission of the following details may delay your support request or receive no attention at all.
Version of s3fs being used (s3fs --version)
1.84
Version of fuse being used (pkg-config --modversion fuse)
2.9.4
System information (uname -r)
3.10.0-514.el7.x86_64
Distro (cat /etc/issue)
Red Hat Enterprise Linux Server release 7.3 (Maipo)
s3fs command line used (if applicable)
/etc/fstab entry (if applicable):
s3fs syslog messages (grep s3fs /var/log/syslog, or s3fs outputs)
Not posting the whole log but a summary from the log.
Details about issue
We are trying to get s3fs to work with our custom solution of AWS S3. Lets call it blobstore. We have about 20 nodes in our database and each has s3fs reading from our blobstore. During each microbatch of our job we are writing files to our blobstore which are then read on each of our database nodes. The files once read are rarely ever read again. For each microbatch we are seeing connections from s3fs (on each node). The connections they make and their response code are summarized above. As you can see we are getting 3 404's and 1 successful connection. Is there any config changes that we could do to avoid seeing those 404's?
@gaul commented on GitHub (Jan 24, 2019):
s3fs is trying to detect whether a file or a directory exists, including legacy directories. Does
-o notsup_compat_direliminate some of these unwanted requests?@acdha commented on GitHub (Mar 15, 2019):
For my project
notsup_compat_dirbreaks compatibility because we have keys following the/suffix pattern but no empty directory objects. If, like awscli, it did the key prefix check first this would work without needing to create empty metadata files.@kahing commented on GitHub (Apr 9, 2019):
Prefix checks are 10x more expensive than 404. Metadata operations are usually a significant part of your overall see expenses. Do the 404s bother you or are they causing real problems?
@ggtakec commented on GitHub (Apr 9, 2019):
By using the latest s3fs and using curl version 7.51.0 or later, you can maintain an SSL session even with 404 errors. This will minimize the cost.
Please try it.
@gaul commented on GitHub (Oct 10, 2020):
This works as intended.