Unfortunately there is still a widely held belief that file fragmentation is a thing of the past and not necessary. Some IT guys even mention they read somewhere that “NTFS does not need fragmentation” or somehow prevents fragmentation. Unfortunately the truth is, things haven’t changed at all and file fragmentation is the result of the necessity of organizing files in linear storage. Fragmentation occurs due to the “physics” of storing data on disk drives and (almost) nothing can avoid it.
Each file typically use multiple sectors on the disk. When we place files side by side and then need to extend a file, the result is fragmentation. Obviously the file cannot be extended unless the file next to it is moved away, or we split the file and extend it elsewhere; hence, we created a fragmented file. No magic on earth can change that fact; however, there are of course algorithms in place to avoid this from happening with some probability but not with certainty.
Another place where fragmentation is evident and causes a performance degradation is the NTFS Master File Table. Small files get written directly into it and can fill it up quickly. This leads to the file system having to allocate more space elsewhere, i.e. a fragment, to extend the MFT.
As time goes by, even on a disk or disk array that is mostly empty, files and folders get fragmented. Since backups are usually sent to cheaper and more voluminous mechanic drives, each time the disk encounters a fragmented file, the heads have to be moved to the new file position and this incurs a seek time. Usually the seek time is in the range of a few ms. In a heavily fragmented file system, these milliseconds quickly add up to seconds and seconds to minutes and finally hours of unnecessary processing.
In short, if you want faster backups, you must defragment your backup drives as well as your data drives where the data is stored. Some intelligent defragmentation software also offers MFT defragmentation and fragmentation avoidance by leaving some space behind files that are likely to grow. Dynamically expanding virtual machine disk files are one of the most likely category of files where excessive file fragmentation becomes evident, since virtual disks constantly expand over time.
But it’s not only virtual disk files and virtual disk backups that are affected. Typical file server data is also not immune to fragmentation.
Consider the defrag output of this disk array that holds a file server backup for about two years now:
Pre-Optimization Report: Volume Information: Volume size = 10.91 TB Cluster size = 4 KB Used space = 2.28 TB Free space = 8.63 TB Fragmentation: Total fragmented space = 79% Average fragments per file = 2.47 Movable files and folders = 2311698 Unmovable files and folders = 4 Files: Fragmented files = 1133584 Total file fragments = 3381277 Folders: Total folders = 24824 Fragmented folders = 6039 Total folder fragments = 41735 Free space: Free space count = 409309 Average free space size = 22.10 MB Largest free space size = 379.21 GB Master File Table (MFT): MFT size = 2.37 GB MFT record count = 2495231 MFT usage = 100% Total MFT fragments = 45
This file server data backup array shows massive levels of disk fragmentation after being used for less than two years for nightly file server data, which consist only of documents, not virtual disk data.
Note the MFT is full and fragmented over 45 times! The largest free space is just 378 GB, even though the 11TB disk array still has a total of over 8.6 TB free.
The number of fragmented files and folders are also very high.
When a backup is run, the source and destination need to be compared and scanned. With fragmented drives, there will be millions of unnecessary head movements. Hence, the accumulated seek time causes backups to run for hours more than actually necessary.
Do I need to defragment SSD drives? Yes and no. Intelligent disk defragmentation software exists that can reduce the wear of the flash cells inside the SSD drives. Also, the speeds quoted by SSD manufacturers also refer to consecutive reading and writing. When blocks are dispersed randomly, the SSD drive performance also suffers. The effect is not as dramatic as with mechanical drives, however. The wear and tear on SSD drives, on the other hand, is a major cause of premature disk failure. So even if your SSD performs okay despite being fragmented, it may cause earlier failure.
Backup Software and How to Speed It Up
Especially in the case of file server backups, database, and virtual machine backups, run the defragmentation tool of your choice at regular intervals. If you are dealing with fixed sized virtual disks and databases, the defragmentation only needs to be done once on the host. However, for file server data it’s best to defragment often. Do not forget to also defragment your backup drives. To speed up file server backups, our backup software offers a feature to process multiple files simultaneously. This is especially useful when dealing with many small files, such as document files, as there is a certain amount of overhead when processing individual files.
Most of the time it’s best, however, not to run backups in parallel. Hard drives and system look-ahead caching are optimized to offer the best throughput when reading consecutive sectors. Even on most SSD drives performance is better when data is in a single block rather than split into thousands of fragments. As a bonus, some defragmentation tools offer fragmentation avoidance, which not only improves performance, but also reduces the wear and tear on your hardware. Server backup software only works at highest possible speeds when the underlying file system is free from fragmentation. This applies to the source folders as well as the backup folders. Fragmentation is a natural occurrence in sequential storage and results from files being repeatedly created, especially in limited space. The effect is seen with file server backup software as well as with Hyper-V backup software and VMware backup software. The former is prone to MFT fragmentation and the latter to file fragmentation, specifically when expanding virtual disks are being used.
Backup Software OverviewThe Best Backup Software in 2024 Download BackupChain®
BackupChain is the all-in-one server backup software for:
Disk Image Backup
Drive Cloning and Disk Copy
File Server Backup
Virtual Machine Backup
Server Backup Solution
- Best Practices for Server Backups
- NAS Backup: Buffalo, Drobo, Synology
- How to use BackupChain for Cloud and Remote
- DriveMaker: Map FTP, SFTP, S3 Sites to a Drive Letter (Freeware)
- VM Backup
- V4 Articles
- Knowledge Base
- Archive 2024
- Archive 2022
- Archive 2021
- Archive 2020
- Archive 2019
- Archive 2017
- Archive 2016
- Archive 2015
- Archive 2014
- Archive 2013
- BackupChain (German)
- German Help Pages
- BackupChain (Greek)
- BackupChain (Spanish)
- BackupChain (French)
- BackupChain (Dutch)
- BackupChain (Italian)
- BackupChain is an all-in-one, reliable backup solution for Windows and Hyper-V that is more affordable than Veeam, Acronis, and Altaro.
Other Backup How-To Guides
- How to Backup and Restore Hyper-V Virtual Machine
- Carbonite vs. Acronis vs. BackupChain Cloud Backup Plans Review
- How to Delete All VSS Shadows and Orphaned Shadows
- How to Create A File-based Backup of Hyper-V
- Ubuntu VM Hyper-V Backup and What You Need to Know for Linux Support
- Convert VHD files to VHDX, VMDK, VDI, and Physical Disk
- Windows 10 (Final) Download Link Enterprise Trial / Evaluation
- NAS Backup Software for QNAP, Buffalo, Drobo, and Synology
- How to Mount Wasabi as a Network Drive with Drive Letter
- How to Map SSH / SFTP as a Network Drive in Windows
- Choosing the Right File System: NTFS, ReFS, and exFAT Compared
- How Set up a Windows Disk Imaging Backup Task
- FTP Server TCP/IP Port Exhaustion Prevention
- How to Plan Hyper-V Deployment for Optimal Virtualization
- RAID Backup Software for RAID Array Backup (0, 1, 5, 10)
- 6 Common Cloud Backup Storage Pitfalls You Should Know
- How to Fix: The semaphore timeout period has expired, Error 121
- Backup Software with Encryption for Windows 11, Windows Server 2022
- VHD vs VHDX File Format: A Quick but Complete Comparison
- 18 Hyper-V Requirements and Recommended Hardware