So, I was curious about, of all things, file systems. I am aware this is very nerdy of me, but since I got the Mac I was curious if I had to worry about such things as defragmentation. Windows users know that you need to defrag your hard drive every so often in order to keep your computer running smoothly. In the world of Mac, it is said that you don’t need to defrag. Why is this?
So, I decided to look into the different types of file systems and look for an answer. Is the file system Windows uses responsible for the difficulties?
The Different File Systems
File Allocation Table (FAT). This is a file system developed by Microsoft for MS-DOS and was used up to Windows ME. A disc formatted using FAT consists of a boot sector, file allocation tables, and the data. The boot sector contains the code necessary for your computer to boot up. The file allocation tables are a mapping of where on the disc certain files and directories reside. Then you have your data itself. The problem with the FAT file system is that when a file is deleted or a new file is removed, that blank space on the drive can be written to by something else. FAT does not take care of the location of new files when this happens, and that leads to file fragments being written all over the disk. The file allocation tables do allow the data to be found, but the read/write head on the hard drive would need to gather your data from disparate parts of the disk. This is why you see a performance decrease. The FAT file system is particularly susceptible to fragmentation by its very design. More information about FAT can be found on Wikipedia.
NTFS is short for New Technology File System. It was, again, developed by Microsoft for its Windows NT line of operating systems. That means NTFS is used in any version of Windows based on the NT kernel, including Windows 2000, XP, Server 2003, Server 2008 and the venerable Windows Vista. The major difference with NTFS is that it is based around metadata. Metadata is “data about data”, according to Wikipedia. In other words, the metadata is like a little mini-database on the drive which stores all kinds of information about the files and directories on the hard drive. NTFS supports things like compression, file-level security and other things useful to enterprise and those attributes are stored in the metadata. The cool thing about this way of dealing with files is that it is extendable to support other features. In fact, Microsoft has released five different versions of NTFS, each successive update providing more features. For information on NTFS can be found on Wikipedia.
As for fragmentation, NTFS is much more improved than FAT and is more efficient in how it handles data location on the hard drive. But, NTFS is subject to fragmentation. There was a myth for awhile that NTFS was not subject to fragmentation, but again, it was a myth. The NTFS file system is a very flexible one. As new attributes or capacity is called for from NTFS, it makes room and stores that information in the Master File Table. If a certain amount of space was set aside for a small file, and then that file becomes very large, portions of that file will need to be stored in other areas of the drive as NTFS creates new data storage areas are created. Also, the NTFS file system used clusters, just like FAT. So, yes, NTFS might have a reduced capacity for fragmentation, but it is still susceptible.
Ext3 is the file system used by Linux. What makes ext3 most notable in contrast to any Windows file system is that it is a journaled file system. A journaled file system is one where any and all changes to any file are logged into a journal before actually being written to the drive. The journal is stored in a designated area of the drive. By its nature, a journaled file system is much less likely to be corrupted. The journal is a running record of ALL actions to be performed on a file. So, in the event of interruption (such as a power failure), the events in the journal can simply be “replayed” to recreate the consistency between the journal and the files on the drive.
The nature of ext3 makes fragmentation all but non-existent. In fact, Wikipedia says that the Linux System Administrator Guide states, “Modern Linux filesystem(s) keep fragmentation at a minimum by keeping all blocks in a file close together, even if they can’t be stored in consecutive sectors. Some filesystems, like ext3, effectively allocate the free block that is nearest to other blocks in a file. Therefore it is not necessary to worry about fragmentation in a Linux system.”
However, when you look into OS X’s file system, which is also journaled, you begin to see why Linux is subject to fragmentation as well…
Hierarchical File System (HFS) is the file system used by Mac OS X. It was developed by Apple itself. We have the original HFS file system (often called Mac OS Standard) and the more recent revision HFS Plus (referred to as Max OS Extended). HFS has been through many revisions. HFS is hardly used anymore. HFS Plus was introduced with Mac OS 8.1. Most notable to our discussion is that Apple introduced journaling to their file system with Mac OS 10.3, along with several other features key to the way OS X operates.
There seems to be two schools of thought when it comes to defragmentation in OS X. Some say that it is not necessary because it is using a journaled file system. Other say that it is unnecessary, just not in the same way as Windows. OS X has the built-in ability to take care of file fragmentation and it will do this on it’s own. However, what can occur is drive fragmentation – small bits of free space between files. Performance-wise, it is almost a non issue and you will rarely gain any performance gain by doing a traditional defrag of an OS X machine. Drive fragmentation really only becomes an issue if you begin to fill your drive up to near full capacity. This is because OS X will run out of room for its own system files.
So, in short, there is no need to defragment in OS X unless you are beginning to fill up your hard drive. As this proceeds, you may begin to experience random OS X “weirdness” due to the OS running out of space for it’s temp files. When this occurs (or preferably before), a defrag of your hard drive will get rid of any slack space between files on the drive and recover space for the use of OS X.
I believe this will also be the case with Linux.
Other reference content:
- Does Mac OS X need a disk dragmenter/optimizer?
- Macintosh OS X Routine Maintenance
- Why Does Linux Need Defragmenting?
So, In Short
If you are running Windows, you are subject to fragmentation. FAT32 systems are VERY prone to it. NTFS is less prone, but still enough so. Linux and Mac users are subject to fragmentation, but unlike Windows, it does not cause any performance drag on the machine. Also, Mac users really only need to worry about fragmentation as their drives near capacity. The issue with Linux and Mac is not file fragmentation (as with Windows), but drive fragmentation.
Hope that helps. And, as always, I welcome any people with knowledge of this area to comment. I did the best research I could on this, but it is certainly possible I got something wrong. And when you throw the Linux and Mac “defrag debates” into the mix, there are certain to be opinions on both sides.