When addressing system performance issues, a key element that is often overlooked is Disk Fragmentation. Even on a brand new system with plenty of RAM and high-end processors, the performance of the hard disk may be a bottleneck causing system performance issues. It takes time to load large data files into memory - issues become particularly noticeable when dealing with movies, video clips, database files or .ISO image files which may easily be several gigabytes in size. On a freshly formatted disk, these files load fairly quickly. Over time, however you may start to notice performance degradation - caused by disk fragmentation.
We touched on disk fragmentation when we were
discussing the Page File a couple of months ago
, but we never really got into the nuts and bolts of it. To understand disk fragmentation though, you need to understand the basic structure of hard disks. When you format a hard disk, the formatting process divides the disk into sectors, each of which contains space for 512 bytes of data. The file system then combines groups of sectors into clusters. A cluster is the smallest unit of space available for holding a single file - or part of a file. On NTFS disks, the cluster sizes are determined based on the drive size as shown below (this information is also available in
Microsoft KB 314878
). When formatting disks it is possible to change the cluster size, however this may cause additional performance issues.
Drive Size (Logical Volume)
512MB or less
513MB - 1,024MB (1GB)
1,024 bytes (1kb)
1,025MB - 2,048MB (2GB)
2,048 bytes (2kb)
4,096 bytes (4kb)
Using the information above, if you were to take a 100MB video file, the file would be divided into roughly 25,000 pieces. If you save this 100MB file onto a freshly formatted disk, the information would be written in contiguous clusters. Since all of the clusters holding the data for this file are physically adjacent to each other, the mechanical components of the hard disk work very efficiently, pulling the data in one operation. In addition, the hard disk's cache and the Windows disk cache can anticipate data requests and fetch data from nearby clusters. This data can then be retrieved by an application from cached memory which is faster than retrieving the information from the disk itself.
Seems pretty straightforward, right? The problem is that the hard disks don't stay neatly organized for very long. Whenever you add data to an existing file, the file system has to allocate more clusters for storage. Typically, these clusters wind up being in a different physical location on the disk. As you delete files, you create gaps in the arrangement of the contiguously stored files. As you save new files (and this is especially true for large files), the file system uses up all of these bits of free space - resulting in the new files being scattered all over the disk in noncontiguous pieces. And thus we end up with fragmented disks and system performance issues because the disk heads have to spend time moving from cluster to cluster before they can read or write the data.
Enter Disk Defragmenter. This utility physically rearranges the files so that they are stored (as much as possible) in physically contiguous clusters. In addition to the consolidation of files and folders, the Defragmenter utility also consolidates free space - meaning that it is less likely for new files to be fragmented when you save them. For operating systems prior to Windows Vista, you had to manually run the utility or schedule automatic defragmentation via a scheduled task. On Windows Vista, Disk Defragmenter runs as a low-priority background task that is automatically run on a weekly basis without requiring user intervention. On Windows Server 2008, which uses the same Disk Defragmenter, the automatic defragmentation is not enabled by default. Also, the color-coded display that was part of earlier versions of the utility has been retired (believe it or not, more than a few people have asked about that!). Aside from the GUI version of the tool, you can also use a command-line version that enables some more granular control over the process. The utility name is DEFRAG.EXE and does require administrative privileges to run. The basic operation of the utility involves passing it a driver letter, for example:
would perform a defragmentation of the C: drive. You can also specify other options through the use of command-line switches:
: Defragments all volumes on the system. You can use this switch without needing to specify a drive letter or mount point
: Perform an analysis of the selected drive and provides a summary output (shown below):
: Performs a partial defragmentation by consolidating only file fragments that are less than 64MB in size. This is the default setting
: Performs a full defragmentation by consolidating all file fragments regardless of size
: Force defragmentation of the volume even if the amount of free space is lower than normally required. When running this, be aware that it can result in slow system performance while the defragmentation is occurring
: Displays verbose reports. When used in combination with the
switch, only the analysis report is displayed. When used alone, both the analysis and defragmentation reports are shown.
: Runs the defragmentation in the background and only if the system is idle
: Optimizes boot files and applications, but leaves the rest of the drive untouched
So now that we've covered what disk fragmentation is and how to address it, there are some caveats. You must have at least 15 percent free space on the disk volume before Disk Defragmenter can completely defragment the volume. If you have less free space, then a partial defragmentation will occur (unless you force the defragmentation with the
switch). Also, you cannot defragment a volume that has been marked by the OS as possibly containing errors. This is where you would need to use the CHKDSK.EXE utility to ensure that there are no underlying disk issues. Some other things to look out for:
Empty the Recycle Bin before defragmenting. Disk Defragmenter does not defragment the Recycle Bin
As we discussed in our Page File post, if you want to defragment the page file, you need to zero it out first and then defragment the disk
By default, fragments that are greater than 64MB in size are ignored by Disk Defragmenter. Fragments of this size (which already contain at least 16,000 contiguous clusters) have a negligible impact on performance
Disk Defragmenter will not defragment files that are in use. For best results, shut down all running programs, or log off and log back in as an administrative account before defragmenting the disk
And with that, it's time to wrap up this post. Until next time ...