There are plenty of utilities (GUI, such as filelight and TUI, such as dua as well) for analyzing disk usage by space, but I would like to view my folders based on the count of files, as I’m making backups, and folders with lots of small files (e.g. node_modules) take very long to move around, so I guess that I’d be better of compressing those into a single file before archiving, as it’s already highly unlikely that I’ll need to access them anyway.
Thanks for any pointers in advance!


The
findcommand could be your friend for getting a full depth count. Something like:find /path/name/here/ | wc -lOr just:
find . | wc -lfor the current directory.
There’s also a command called
locate(often with another letter in front, but accessible by just that word) which maintains a database of filenames on a system that can be used, provided it’s installed and has built that database.Pro: Faster than churning the disk every time with
find. (Though disk cache can help alleviate some of this).Cons: Can get out of date with a changing filesystem. Harder to search for relative paths like
.locate -r '^/path/name/here/' | wc -l… would be roughly equivalent to the first
findexample.