summaryrefslogtreecommitdiffstats
path: root/Documentation/admin-guide/device-mapper/dm-clone.rst
blob: b43a34c1430aec4f5894ab9bef476270a9891cff (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
.. SPDX-License-Identifier: GPL-2.0-only

========
dm-clone
========

Introduction
============

dm-clone is a device mapper target which produces a one-to-one copy of an
existing, read-only source device into a writable destination device: It
presents a virtual block device which makes all data appear immediately, and
redirects reads and writes accordingly.

The main use case of dm-clone is to clone a potentially remote, high-latency,
read-only, archival-type block device into a writable, fast, primary-type device
for fast, low-latency I/O. The cloned device is visible/mountable immediately
and the copy of the source device to the destination device happens in the
background, in parallel with user I/O.

For example, one could restore an application backup from a read-only copy,
accessible through a network storage protocol (NBD, Fibre Channel, iSCSI, AoE,
etc.), into a local SSD or NVMe device, and start using the device immediately,
without waiting for the restore to complete.

When the cloning completes, the dm-clone table can be removed altogether and be
replaced, e.g., by a linear table, mapping directly to the destination device.

The dm-clone target reuses the metadata library used by the thin-provisioning
target.

Glossary
========

   Hydration
     The process of filling a region of the destination device with data from
     the same region of the source device, i.e., copying the region from the
     source to the destination device.

Once a region gets hydrated we redirect all I/O regarding it to the destination
device.

Design
======

Sub-devices
-----------

The target is constructed by passing three devices to it (along with other
parameters detailed later):

1. A source device - the read-only device that gets cloned and source of the
   hydration.

2. A destination device - the destination of the hydration, which will become a
   clone of the source device.

3. A small metadata device - it records which regions are already valid in the
   destination device, i.e., which regions have already been hydrated, or have
   been written to directly, via user I/O.

The size of the destination device must be at least equal to the size of the
source device.

Regions
-------

dm-clone divides the source and destination devices in fixed sized regions.
Regions are the unit of hydration, i.e., the minimum amount of data copied from
the source to the destination device.

The region size is configurable when you first create the dm-clone device. The
recommended region size is the same as the file system block size, which usually
is 4KB. The region size must be between 8 sectors (4KB) and 2097152 sectors
(1GB) and a power of two.

Reads and writes from/to hydrated regions are serviced from the destination
device.

A read to a not yet hydrated region is serviced directly from the source device.

A write to a not yet hydrated region will be delayed until the corresponding
region has been hydrated and the hydration of the region starts immediately.

Note that a write request with size equal to region size will skip copying of
the corresponding region from the source device and overwrite the region of the
destination device directly.

Discards
--------

dm-clone interprets a discard request to a range that hasn't been hydrated yet
as a hint to skip hydration of the regions covered by the request, i.e., it
skips copying the region's data from the source to the destination device, and
only updates its metadata.

If the destination device supports discards, then by default dm-clone will pass
down discard requests to it.

Background Hydration
--------------------

dm-clone copies continuously from the source to the destination device, until
all of the device has been copied.

Copying data from the source to the destination device uses bandwidth. The user
can set a throttle to prevent more than a certain amount of copying occurring at
any one time. Moreover, dm-clone takes into account user I/O traffic going to
the devices and pauses the background hydration when there is I/O in-flight.

A message `hydration_threshold <#regions>` can be used to set the maximum number
of regions being copied, the default being 1 region.

dm-clone employs dm-kcopyd for copying portions of the source device to the
destination device. By default, we issue copy requests of size equal to the
region size. A message `hydration_batch_size <#regions>` can be used to tune the
size of these copy requests. Increasing the hydration batch size results in
dm-clone trying to batch together contiguous regions, so we copy the data in
batches of this many regions.

When the hydration of the destination device finishes, a dm event will be sent
to user space.

Updating on-disk metadata
-------------------------

On-disk metadata is committed every time a FLUSH or FUA bio is written. If no
such requests are made then commits will occur every second. This means the
dm-clone device behaves like a physical disk that has a volatile write cache. If
power is lost you may lose some recent writes. The metadata should always be
consistent in spite of any crash.

Target Interface
================

Constructor
-----------

  ::

   clone <metadata dev> <destination dev> <source dev> <region size>
         [<#feature args> [<feature arg>]* [<#core args> [<core arg>]*]]

 ================ ==============================================================
 metadata dev     Fast device holding the persistent metadata
 destination dev  The destination device, where the source will be cloned
 source dev       Read only device containing the data that gets cloned
 region size      The size of a region in sectors

 #feature args    Number of feature arguments passed
 feature args     no_hydration or no_discard_passdown

 #core args       An even number of arguments corresponding to key/value pairs
                  passed to dm-clone
 core args        Key/value pairs passed to dm-clone, e.g. `hydration_threshold
                  256`
 ================ ==============================================================

Optional feature arguments are:

 ==================== =========================================================
 no_hydration         Create a dm-clone instance with background hydration
                      disabled
 no_discard_passdown  Disable passing down discards to the destination device
 ==================== =========================================================

Optional core arguments are:

 ================================ ==============================================
 hydration_threshold <#regions>   Maximum number of regions being copied from
                                  the source to the destination device at any
                                  one time, during background hydration.
 hydration_batch_size <#regions>  During background hydration, try to batch
                                  together contiguous regions, so we copy data
                                  from the source to the destination device in
                                  batches of this many regions.
 ================================ ==============================================

Status
------

  ::

   <metadata block size> <#used metadata blocks>/<#total metadata blocks>
   <region size> <#hydrated regions>/<#total regions> <#hydrating regions>
   <#feature args> <feature args>* <#core args> <core args>*
   <clone metadata mode>

 ======================= =======================================================
 metadata block size     Fixed block size for each metadata block in sectors
 #used metadata blocks   Number of metadata blocks used
 #total metadata blocks  Total number of metadata blocks
 region size             Configurable region size for the device in sectors
 #hydrated regions       Number of regions that have finished hydrating
 #total regions          Total number of regions to hydrate
 #hydrating regions      Number of regions currently hydrating
 #feature args           Number of feature arguments to follow
 feature args            Feature arguments, e.g. `no_hydration`
 #core args              Even number of core arguments to follow
 core args               Key/value pairs for tuning the core, e.g.
                         `hydration_threshold 256`
 clone metadata mode     ro if read-only, rw if read-write

                         In serious cases where even a read-only mode is deemed
                         unsafe no further I/O will be permitted and the status
                         will just contain the string 'Fail'. If the metadata
                         mode changes, a dm event will be sent to user space.
 ======================= =======================================================

Messages
--------

  `disable_hydration`
      Disable the background hydration of the destination device.

  `enable_hydration`
      Enable the background hydration of the destination device.

  `hydration_threshold <#regions>`
      Set background hydration threshold.

  `hydration_batch_size <#regions>`
      Set background hydration batch size.

Examples
========

Clone a device containing a file system
---------------------------------------

1. Create the dm-clone device.

   ::

    dmsetup create clone --table "0 1048576000 clone $metadata_dev $dest_dev \
      $source_dev 8 1 no_hydration"

2. Mount the device and trim the file system. dm-clone interprets the discards
   sent by the file system and it will not hydrate the unused space.

   ::

    mount /dev/mapper/clone /mnt/cloned-fs
    fstrim /mnt/cloned-fs

3. Enable background hydration of the destination device.

   ::

    dmsetup message clone 0 enable_hydration

4. When the hydration finishes, we can replace the dm-clone table with a linear
   table.

   ::

    dmsetup suspend clone
    dmsetup load clone --table "0 1048576000 linear $dest_dev 0"
    dmsetup resume clone

   The metadata device is no longer needed and can be safely discarded or reused
   for other purposes.

Known issues
============

1. We redirect reads, to not-yet-hydrated regions, to the source device. If
   reading the source device has high latency and the user repeatedly reads from
   the same regions, this behaviour could degrade performance. We should use
   these reads as hints to hydrate the relevant regions sooner. Currently, we
   rely on the page cache to cache these regions, so we hopefully don't end up
   reading them multiple times from the source device.

2. Release in-core resources, i.e., the bitmaps tracking which regions are
   hydrated, after the hydration has finished.

3. During background hydration, if we fail to read the source or write to the
   destination device, we print an error message, but the hydration process
   continues indefinitely, until it succeeds. We should stop the background
   hydration after a number of failures and emit a dm event for user space to
   notice.

Why not...?
===========

We explored the following alternatives before implementing dm-clone:

1. Use dm-cache with cache size equal to the source device and implement a new
   cloning policy:

   * The resulting cache device is not a one-to-one mirror of the source device
     and thus we cannot remove the cache device once cloning completes.

   * dm-cache writes to the source device, which violates our requirement that
     the source device must be treated as read-only.

   * Caching is semantically different from cloning.

2. Use dm-snapshot with a COW device equal to the source device:

   * dm-snapshot stores its metadata in the COW device, so the resulting device
     is not a one-to-one mirror of the source device.

   * No background copying mechanism.

   * dm-snapshot needs to commit its metadata whenever a pending exception
     completes, to ensure snapshot consistency. In the case of cloning, we don't
     need to be so strict and can rely on committing metadata every time a FLUSH
     or FUA bio is written, or periodically, like dm-thin and dm-cache do. This
     improves the performance significantly.

3. Use dm-mirror: The mirror target has a background copying/mirroring
   mechanism, but it writes to all mirrors, thus violating our requirement that
   the source device must be treated as read-only.

4. Use dm-thin's external snapshot functionality. This approach is the most
   promising among all alternatives, as the thinly-provisioned volume is a
   one-to-one mirror of the source device and handles reads and writes to
   un-provisioned/not-yet-cloned areas the same way as dm-clone does.

   Still:

   * There is no background copying mechanism, though one could be implemented.

   * Most importantly, we want to support arbitrary block devices as the
     destination of the cloning process and not restrict ourselves to
     thinly-provisioned volumes. Thin-provisioning has an inherent metadata
     overhead, for maintaining the thin volume mappings, which significantly
     degrades performance.

   Moreover, cloning a device shouldn't force the use of thin-provisioning. On
   the other hand, if we wish to use thin provisioning, we can just use a thin
   LV as dm-clone's destination device.