mirror of
https://github.com/s3fs-fuse/s3fs-fuse.git
synced 2026-04-25 21:35:58 +03:00
[GH-ISSUE #303] du -skh <directory> takes too much time #158
Labels
No labels
bug
bug
dataloss
duplicate
enhancement
feature request
help wanted
invalid
need info
performance
pull-request
question
question
testing
No milestone
No project
No assignees
1 participant
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference
starred/s3fs-fuse#158
Loading…
Add table
Add a link
Reference in a new issue
No description provided.
Delete branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @dejlek on GitHub (Nov 26, 2015).
Original GitHub issue: https://github.com/s3fs-fuse/s3fs-fuse/issues/303
Today I have compiled and installed s3fs from the master branch. One of the things I noticed is that
du -skh <directory>takes too long (20+ min) for a directory which contains only few subdirectories and all subdirectories in total have ~200 files (not big ones).Is this an expected behaviour?
@gaul commented on GitHub (Nov 28, 2015):
This is inherent to the design of s3fs which strives for close POSIX compatibility. Every call to
getdentsrequires an S3 HEAD object call per file to fill thestatstruct. goofys has looser POSIX compatibility and calls togetdentsonly require one S3 list objects call instead of one per file. This yields vastly improved performance fordu.@ggtakec commented on GitHub (Dec 13, 2015):
Hi, @dejlek
s3fs posts wasteful HEAD request for compatibility with other s3 client.
We can also discard this compatibility, but has been held so far.
It might have an impact on the speed of this order s3fs.
In order to alleviate this, s3fs can have the ability to cache the stat information for each object.
Then you can use max_stat_cache_size, stat_cache_expire, and enable_noobj_cache options.
In order to reduce unnecessary requests, I recommend that you use the enable_noobj_cache option.
If you can, pleae try to set these options.
Thanks in advance for your help.
@gaul commented on GitHub (Jan 24, 2019):
@dejlek master includes a readdir optimization that should reduce your run-time and setting
-o multireq_maxgreater than 20 will also improve speed. However s3fs does not do any kind of read-ahead optimization which would greatly accerelate this operation. I have been experimenting with parallel unix tools for this situation but am not ready to share my progress.@ggtakec commented on GitHub (Mar 29, 2019):
We launch new version 1.86, which is tuned some performance issue.
Please try to use it or master branch code.
I will close this, but if the problem persists, please reopen or post a new issue.
Thanks in advance for your assistance.