Context
I’m aware that the “slow re-scan issue” has been discussed many times, but something still don’t make sense with my understanding of system resource utilisation.
System Specifications
- Environment: Supercomputer running Windows 2016
- File System: NTFS
- Hardware:
- RAM: 1.5TB
- CPU: Xeon Gold processors
- Storage: M.2 array (20GB+/s read/write, millions of IOPS)
- Sync Folder: 30TB, containing 20 million files
Observation
The initial scan took ~5 hours, which is understandable. However, the re-scan still requires about an hour to complete, and this is where my confusion lies.
The Puzzle
During the re-scan process, the system appears to be largely idle:
- All CPU cores usage is minimal (no core exceeds 10%)
- Disk activity is negligible (almost zero read/write operations)
- RAM and GPU resources are freely available
In essence, the entire computer seems unused, yet the re-scan takes a considerable amount of time like just idling there, and suddenly finishes.
I am aware that rescan is a single thread process, but that thread is not using much of anything, in any CPU core.
Key Questions
- Which component of the system is working so hard secretly during the re-scan? What is the actual bottleneck?
- Can we better utilize the available computer resources (CPU, disk, etc.) or purchase any new hardware to make it faster?
I’m seeking insights into the underlying mechanics of this re-scan process and potential strategies for optimization.