Lazyweb, what's a reasonable removable media filesystem to use for a ridiculously large number of smallish files, hundreds of GB total? FAT32 is just incredibly slow.
@dalias@aburka that many small files is going to be the worst performance ever. If you can consolidate in a tarball, a zip, a rat, even without compression, before copying, you'll see a huge performance increase. This was my bane when I had to do backups on a PB of data across many multi-TB volumes. Weeks. When they moved to "archive" and I consolidated them, hours to write the same data to WORM tapes. It's the open/read/close cycle. Open/close are fixed times, so more is slower.
@emag@aburka It's not the open/close taking up time. It's the actual data transfer. But even if it were, putting the data into archives would just move the cost to archiving and unarchiving it. All the costs here except the disk being stupidly slow are pretty much fundamental.
@dalias I asked in the local Unix User Group, one person said ext4, one said fuse fs with tar back-end and linked this https://github.com/google/fuse-archive and one person said ext2 with this reasoning, but the ext4 person disagreed (I hope your instance has autotranslate):
@dalias Zu der Frage wegen vielen kleinen Dateien und removable Media filesystem: mir würde hier als Erstes ext2 einfallen. Wahrscheinlich könnte man auf Journaling und Dateiintegrität verzichten. FAT32 ist unter Linux wirklich super langsam bis hart zu unbrauchbar, die Erfahrung habe ich selbst schon gemacht.