I am the author of Weed-FS. WeedFS is ideal for your requirement. Hadoop cannot process many small files, in addition to your reasons, each file must have an entry in the main. If the number of files is large, the hdfs node wizard cannot scale.
Weed-FS speeds up when compiling with the latest releases of Golang.
Weed-FS has recently made many new improvements. Now you can easily test and compare with the built-in download tool. This file loads all files recursively into a directory.
weed upload -dir=/some/directory
Now you can compare "du -k / some / directory" to see disk usage, and "ls -l / your / weed / volume / directory" to see disk usage Weed-FS.
And I suppose you will need replication with a data center, rack support, etc. Now they are in mode
source share