Hello and welcome to ponder upon a rather rare(?) problem.
I have a 12 TB Western Digital Gold HDD formatted to ReFS, that contains over 14 million “SQL-esque” files that are in need of a defragmentation. The problem is, that because of the ReFS metadata fragmentation and the degree of fragmentation in general, every defragmentation software is very slow at determining where to place the files (even in their fastest defragmentation mode), when ever there is existing files in the nearby sectors they try to place the files to. There is plenty of space left (over 8.8TB), so the slow defragmentation rate is not because of low space, but because there is millions of files, ranging from 4KB to 2.2MB.
Few notes about the situation;
-
When I first used NTFS for this HDD, the disk usage was nearly constantly 100% and the NTFS filesystem was corrupted very fast within 1-4 days, according to check disk - even with a 64k Cluster size.
-
The StorJ app (essentially a decentralized cloud storage solution, which allows people to rent their HDD space to the people using this service as cloud storage) currently downloads/writes about 38gb worth of files in average daily, which amounts to over 100 000 files per day typically.
-
My current work around has been to consolidate all the files to the inner sectors of the disk, before trying to consolidate them on the faster sectors, which has had limited success. The problem is, that I can’t keep the StorJ server offline long enough to defragment the whole disk that way, without risking my server to be disqualified. And, that the metadata is too fragmented for the defragmentation software to place files fast enough among them.
-
Thus far I have tried O&O Defrag, Perfect Disk, Defragler (thus far seemingly the most responsive when moving files in between the metadata or other files), Ultimate Defrag (DEMO mode only free option), Ultradefrag, Disk SpeedUp (crashes, because it is only 32-bit) and Auslogics Disk Defrag 11 and 12 (both crash, only 32-bit versions available).
So, I have a few questions about this;
-
Is there any way to defragment the ReFS metadata files, like it was possible to do with NTFS metadata on some defragmenting software, or am I out of luck with that?
-
Is there any enterprise grade and/or open source defragmentation software that can handle tens of millions of files more efficiently?
-
What would be the best alternative to ReFS in multimillion file server situations, that could be used in Windows 10/11 environment, defragmented or not even need defragmentation and setup with or without some tinkering?
The end goal would be to find a solution, that provides as fast access times / latencies to files as possible, without the urgent need to upgrade to NVME.
Bonus questions for future reference:
-
Does NVME read and write latencies matter much in situations, where low latency to serve the files is essential or are they so fast that the impact of latency differences on NVME drives are not that important?
-
What would be the best bang for a buck NVME drive to handle tens of millions of small (4KB - 2.2MB) files and hundreds of thousands writes per day?
-
Is there any NVME with SATA connections that has at least nearly as fast access times / latencies as their M.2 counterparts?
Thank you for your help in advance