Sie sind auf Seite 1von 11

Page |1

DATA DEDUPLICATION

Index Page number

Introduction 2-3

Phase 1 implementation 3-5

Phase 2 implantation 5-9

Phase 3 implementation 9-11


Page |2

INTRODUCTION

 In computing, data deduplication is a specialized data


compression technique for eliminating duplicate copies of repeating data.
Related and somewhat synonymous terms are intelligent (data)
compression and single-instance (data) storage.
 Data deduplication -- often called intelligent compression or single-instance
storage -- is a process that eliminates redundant copies of data and reduces
storage overhead.
 Data deduplication techniques ensure that only one unique instance of data
is retained on storage media, such as disk, flash or tape. Redundant
data blocks are replaced with a pointer to the unique data copy.
 This technique is used to improve storage utilization and can also be applied
to network data transfers to reduce the number of bytes that must be sent.
 In the deduplication process, unique chunks of data, or byte patterns, are
identified and stored during a process of analysis.
 As the analysis continues, other chunks are compared to the stored copy
and whenever a match occurs, the redundant chunk is replaced with a small
reference that points to the stored chunk. Given that the same byte pattern
may occur dozens, hundreds, or even thousands of times (the match
frequency is dependent on the chunk size), the amount of data that must be
stored or transferred can be greatly reduced.

BENEFITS

Storage-based data deduplication reduces the amount of storage needed for a


given set of files. It is most effective in applications where many copies of very
similar or even identical data are stored on a single disk—a surprisingly common
scenario..
Page |3

 For example, a typical email system might contain 100 instances of the same
1 megabyte (MB) file attachment. If the email platform is backed up or
archived, all 100 instances are saved, requiring 100 MB of storage space.
With data deduplication, only one instance of the attachment is stored; each
subsequent instance is referenced back to the one saved copy. In this
example, a 100 MB storage demand drops to 1 MB.

PHASE 1 IMPLEMENTATION.

It can be configured in the 3 phase


1) Configure data duplication
2) Implementation in selected storage drives.
3) Execute in PowerShell

Configure data duplication

The below listed steps to configure data duplication.


1. Go to server manger ----> click on tools--->and select add roles and features

2. Then click on server roles--->click on files and storage services---> click on


files and iscsi services.
Page |4

3. Select the data deduplication ---->add future----->click on next --->install


Page |5

4. Wait for until installation complete.

PHASE 2 IMPLEMENTATION

2) Implementation in selected storage drives.

In the above steps we configure the data deduplication in our servers and now
we should selected the drives which want to implement the above configuration

The below steps are


1. Go to server manger ----> files and storage services--->click on volume -->
Page |6

Select the drive which we need configure the data deduplication.

2) right click on the selected drive and select the Configure data deduplication.
Page |7

3) Select the general purpose file server.


After that we select number of days for duplication files older than depend on
our need for example one new folder created the data deduplication will not
occurred for 3 days after 3 days it will start comprising the similar copies files .
Page |8

4) now we need set schedule for that click on deduplication schedule.


Page |9

5.) after select days and time according to our project need click on ok and apply
P a g e | 10

PHASE 3 IMPLEMENTATION

Execute in windows PowerShell


1)Now we should open windows PowerShell as administrator
2) type the command start “ Start –dedupjob –volume E: - type optimization

3) wait for until it process to 100%


P a g e | 11

After 100% progress completed the drive which is selected for data deduplication
the disk size is reduced automatically.

THANKING YOU
Bharath E.

Das könnte Ihnen auch gefallen