Linux – Why there shouldn’t be too many files in one directory that serves just static web requests

filesystemslinuxweb-server

This might be a very general question but I really like to find some detailed answers or clues.

I am discussing this with a friend, trying to convince him to put more than 300,000 files from one single folder to more than one (like 1000 per sub-directory). Those files are images and to be served online web-viewing, like:

www.example.com/folder/1.png
.
.
.
www.example.com/folder/300000.png

I simply remember many years ago when I worked at a online video serving company like Youtube. We put the screenshots in one folder and then the server were always crashing. At that time a "rumor" saying people should not put many files in one folder, but we do not know the detailed reason.

So how many files should I put in one folder? If there is a limitation, why? Any recommended ways to design this?

My server information:

No LSB modules are available.
Distributor ID: Debian
Description:    Debian GNU/Linux 7.8 (wheezy)
Release:    7.8
Codename:   wheezy

Core Build version:

Linux linode 4.1.5-x86_64-linode61 #7 SMP Mon Aug 24 13:46:31 EDT 2015 x86_64 GNU/Linux

I guess this case applies to many different kind of server software.

Best Answer

This isn't really a very big deal with newer filesystems such as XFS and ext4, but on older or misconfigured filesystems it can be a serious problem.


With older Linux filesystems such as ext3, a directory is just an unordered list of files.

That it is unordered is important, because it means that the only way for the system to find a file in a directory is to search it from the beginning to the end.

If a directory contains 3,000 files, it will take an average of 1,500 comparisons to find a random file in the directory. But if the directory contains 300,000 files, it will take an average of 150,000 comparisons to find a random file in that directory.

In either case, if the directory entry is not already cached in RAM, it must be loaded from disk, which would add a significant amount of time to the file access, proportionate to the size of the directory. Obviously a small dentry can be loaded faster than a large one.

Thus, it is much faster when you use a more hierarchical directory structure to separate large numbers of files into unique directories.

XFS does not suffer from this problem, as it uses a hash table to lookup directory entries. Thus it can handle a directory with hundreds of thousands of files nearly as easily as a directory with one file. But it still has the penalty of needing to load the larger data structure from disk. If you have enough RAM in the system, this isn't really a practical problem, though.

Ext4 also uses a hashed directory index.