![]() ![]() ![]() The optimization policy specifies the files that should be considered for Data Deduplication. To preserve access semantics, Data Deduplication uses a file system filter (Dedup.sys) to redirect reads to optimized content completely transparently to the user or application that makes the read request.Ī file is considered optimized (or deduplicated) by Data Deduplication if it has been chunked, and its unique chunks have been stored in the chunk store. Data Deduplication is supported on NTFS formatted volumes.Ī file system filter is a plugin that modifies the default behavior of the file system. The file system is the software and on-disk data structure that the operating system uses to store files on storage media. This is the part of the file that Data Deduplication optimizes. The file stream is the main content of the file. For instance, Date Created, Last Read Date, Author, etc. The chunk store is an organized series of container files in the System Volume Information folder that Data Deduplication uses to uniquely store chunks.Īn abbreviation for Data Deduplication that's commonly used in PowerShell, Windows Server APIs and components, and the Windows Server community.Įvery file contains metadata that describes interesting properties about the file that are not related to the main content of the file. The Unoptimization job, which is a special job that should only be run manually, undoes the optimization done by deduplication and disables Data Deduplication for that volume.Ī chunk is a section of a file that has been selected by the Data Deduplication chunking algorithm as likely to occur in other, similar files. Additionally, Data Deduplication keeps backup copies of popular chunks when they are referenced more than 100 times in an area called the hotspot. When possible, Data Deduplication can automatically use volume features (such as mirror or parity on a Storage Spaces volume) to reconstruct the corrupted data. The Integrity Scrubbing job identifies corruption in the chunk store due to disk failures or bad sectors. The Garbage Collection job reclaims disk space by removing unnecessary chunks that are no longer being referenced by files that have been recently modified or deleted. ![]() The optimization process that Data Deduplication uses is described in detail in How does Data Deduplication work?. The Optimization job deduplicates by chunking data on a volume per the volume policy settings, (optionally) compressing those chunks, and storing chunks uniquely in the chunk store. "Under-the-hood" tweaks for interop with DPM/DPM-like solutionsĭata Deduplication uses a post-processing strategy to optimize and maintain a volume's space efficiency.Virtualized backup applications, such as Microsoft Data Protection Manager (DPM) "Under-the-hood" tweaks for Hyper-V interop.Virtualized Desktop Infrastructure (VDI) servers The following Usage Types provide reasonable Data Deduplication configuration for common workloads: Usage Type Modifications to ranges of a deduplicated files get written unoptimized to the disk and are optimized by the Optimization job the next time it runs. The filter redirects the read operation to the appropriate chunks that constitute the stream for that file in the chunk store. When optimized files are read, the file system sends the files with a reparse point to the Data Deduplication file system filter (Dedup.sys). Replace the original file stream of now optimized files with a reparse point to the chunk store.Place chunks in the chunk store and optionally compress. ![]()
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |