root / block-migration.c @ feature-archipelago
History | View | Annotate | Download (22.8 kB)
1 | c163b5ca | lirans@il.ibm.com | /*
|
---|---|---|---|
2 | c163b5ca | lirans@il.ibm.com | * QEMU live block migration
|
3 | c163b5ca | lirans@il.ibm.com | *
|
4 | c163b5ca | lirans@il.ibm.com | * Copyright IBM, Corp. 2009
|
5 | c163b5ca | lirans@il.ibm.com | *
|
6 | c163b5ca | lirans@il.ibm.com | * Authors:
|
7 | c163b5ca | lirans@il.ibm.com | * Liran Schour <lirans@il.ibm.com>
|
8 | c163b5ca | lirans@il.ibm.com | *
|
9 | c163b5ca | lirans@il.ibm.com | * This work is licensed under the terms of the GNU GPL, version 2. See
|
10 | c163b5ca | lirans@il.ibm.com | * the COPYING file in the top-level directory.
|
11 | c163b5ca | lirans@il.ibm.com | *
|
12 | 6b620ca3 | Paolo Bonzini | * Contributions after 2012-01-13 are licensed under the terms of the
|
13 | 6b620ca3 | Paolo Bonzini | * GNU GPL, version 2 or (at your option) any later version.
|
14 | c163b5ca | lirans@il.ibm.com | */
|
15 | c163b5ca | lirans@il.ibm.com | |
16 | c163b5ca | lirans@il.ibm.com | #include "qemu-common.h" |
17 | 737e150e | Paolo Bonzini | #include "block/block_int.h" |
18 | c163b5ca | lirans@il.ibm.com | #include "hw/hw.h" |
19 | 1de7afc9 | Paolo Bonzini | #include "qemu/queue.h" |
20 | 1de7afc9 | Paolo Bonzini | #include "qemu/timer.h" |
21 | caf71f86 | Paolo Bonzini | #include "migration/block.h" |
22 | caf71f86 | Paolo Bonzini | #include "migration/migration.h" |
23 | 9c17d615 | Paolo Bonzini | #include "sysemu/blockdev.h" |
24 | c163b5ca | lirans@il.ibm.com | #include <assert.h> |
25 | c163b5ca | lirans@il.ibm.com | |
26 | 50717e94 | Paolo Bonzini | #define BLOCK_SIZE (1 << 20) |
27 | 50717e94 | Paolo Bonzini | #define BDRV_SECTORS_PER_DIRTY_CHUNK (BLOCK_SIZE >> BDRV_SECTOR_BITS)
|
28 | c163b5ca | lirans@il.ibm.com | |
29 | c163b5ca | lirans@il.ibm.com | #define BLK_MIG_FLAG_DEVICE_BLOCK 0x01 |
30 | c163b5ca | lirans@il.ibm.com | #define BLK_MIG_FLAG_EOS 0x02 |
31 | 01e61e2d | Jan Kiszka | #define BLK_MIG_FLAG_PROGRESS 0x04 |
32 | 323004a3 | Peter Lieven | #define BLK_MIG_FLAG_ZERO_BLOCK 0x08 |
33 | c163b5ca | lirans@il.ibm.com | |
34 | c163b5ca | lirans@il.ibm.com | #define MAX_IS_ALLOCATED_SEARCH 65536 |
35 | c163b5ca | lirans@il.ibm.com | |
36 | c163b5ca | lirans@il.ibm.com | //#define DEBUG_BLK_MIGRATION
|
37 | c163b5ca | lirans@il.ibm.com | |
38 | c163b5ca | lirans@il.ibm.com | #ifdef DEBUG_BLK_MIGRATION
|
39 | d0f2c4c6 | malc | #define DPRINTF(fmt, ...) \
|
40 | c163b5ca | lirans@il.ibm.com | do { printf("blk_migration: " fmt, ## __VA_ARGS__); } while (0) |
41 | c163b5ca | lirans@il.ibm.com | #else
|
42 | d0f2c4c6 | malc | #define DPRINTF(fmt, ...) \
|
43 | c163b5ca | lirans@il.ibm.com | do { } while (0) |
44 | c163b5ca | lirans@il.ibm.com | #endif
|
45 | c163b5ca | lirans@il.ibm.com | |
46 | a55eb92c | Jan Kiszka | typedef struct BlkMigDevState { |
47 | 323920c4 | Paolo Bonzini | /* Written during setup phase. Can be read without a lock. */
|
48 | a55eb92c | Jan Kiszka | BlockDriverState *bs; |
49 | a55eb92c | Jan Kiszka | int shared_base;
|
50 | a55eb92c | Jan Kiszka | int64_t total_sectors; |
51 | 5e5328be | Jan Kiszka | QSIMPLEQ_ENTRY(BlkMigDevState) entry; |
52 | 323920c4 | Paolo Bonzini | |
53 | 323920c4 | Paolo Bonzini | /* Only used by migration thread. Does not need a lock. */
|
54 | 323920c4 | Paolo Bonzini | int bulk_completed;
|
55 | 323920c4 | Paolo Bonzini | int64_t cur_sector; |
56 | 323920c4 | Paolo Bonzini | int64_t cur_dirty; |
57 | 323920c4 | Paolo Bonzini | |
58 | 52e850de | Paolo Bonzini | /* Protected by block migration lock. */
|
59 | 33656af7 | Marcelo Tosatti | unsigned long *aio_bitmap; |
60 | 323920c4 | Paolo Bonzini | int64_t completed_sectors; |
61 | e4654d2d | Fam Zheng | BdrvDirtyBitmap *dirty_bitmap; |
62 | a55eb92c | Jan Kiszka | } BlkMigDevState; |
63 | a55eb92c | Jan Kiszka | |
64 | c163b5ca | lirans@il.ibm.com | typedef struct BlkMigBlock { |
65 | 323920c4 | Paolo Bonzini | /* Only used by migration thread. */
|
66 | c163b5ca | lirans@il.ibm.com | uint8_t *buf; |
67 | c163b5ca | lirans@il.ibm.com | BlkMigDevState *bmds; |
68 | c163b5ca | lirans@il.ibm.com | int64_t sector; |
69 | 33656af7 | Marcelo Tosatti | int nr_sectors;
|
70 | c163b5ca | lirans@il.ibm.com | struct iovec iov;
|
71 | c163b5ca | lirans@il.ibm.com | QEMUIOVector qiov; |
72 | c163b5ca | lirans@il.ibm.com | BlockDriverAIOCB *aiocb; |
73 | 323920c4 | Paolo Bonzini | |
74 | 52e850de | Paolo Bonzini | /* Protected by block migration lock. */
|
75 | c163b5ca | lirans@il.ibm.com | int ret;
|
76 | 5e5328be | Jan Kiszka | QSIMPLEQ_ENTRY(BlkMigBlock) entry; |
77 | c163b5ca | lirans@il.ibm.com | } BlkMigBlock; |
78 | c163b5ca | lirans@il.ibm.com | |
79 | c163b5ca | lirans@il.ibm.com | typedef struct BlkMigState { |
80 | 323920c4 | Paolo Bonzini | /* Written during setup phase. Can be read without a lock. */
|
81 | c163b5ca | lirans@il.ibm.com | int blk_enable;
|
82 | c163b5ca | lirans@il.ibm.com | int shared_base;
|
83 | 5e5328be | Jan Kiszka | QSIMPLEQ_HEAD(bmds_list, BlkMigDevState) bmds_list; |
84 | 323920c4 | Paolo Bonzini | int64_t total_sector_sum; |
85 | 323004a3 | Peter Lieven | bool zero_blocks;
|
86 | 323920c4 | Paolo Bonzini | |
87 | 52e850de | Paolo Bonzini | /* Protected by lock. */
|
88 | 5e5328be | Jan Kiszka | QSIMPLEQ_HEAD(blk_list, BlkMigBlock) blk_list; |
89 | c163b5ca | lirans@il.ibm.com | int submitted;
|
90 | c163b5ca | lirans@il.ibm.com | int read_done;
|
91 | 323920c4 | Paolo Bonzini | |
92 | 323920c4 | Paolo Bonzini | /* Only used by migration thread. Does not need a lock. */
|
93 | c163b5ca | lirans@il.ibm.com | int transferred;
|
94 | 01e61e2d | Jan Kiszka | int prev_progress;
|
95 | e970ec0b | Liran Schour | int bulk_completed;
|
96 | 52e850de | Paolo Bonzini | |
97 | 52e850de | Paolo Bonzini | /* Lock must be taken _inside_ the iothread lock. */
|
98 | 52e850de | Paolo Bonzini | QemuMutex lock; |
99 | c163b5ca | lirans@il.ibm.com | } BlkMigState; |
100 | c163b5ca | lirans@il.ibm.com | |
101 | d11ecd3d | Jan Kiszka | static BlkMigState block_mig_state;
|
102 | c163b5ca | lirans@il.ibm.com | |
103 | 52e850de | Paolo Bonzini | static void blk_mig_lock(void) |
104 | 52e850de | Paolo Bonzini | { |
105 | 52e850de | Paolo Bonzini | qemu_mutex_lock(&block_mig_state.lock); |
106 | 52e850de | Paolo Bonzini | } |
107 | 52e850de | Paolo Bonzini | |
108 | 52e850de | Paolo Bonzini | static void blk_mig_unlock(void) |
109 | 52e850de | Paolo Bonzini | { |
110 | 52e850de | Paolo Bonzini | qemu_mutex_unlock(&block_mig_state.lock); |
111 | 52e850de | Paolo Bonzini | } |
112 | 52e850de | Paolo Bonzini | |
113 | 32c835ba | Paolo Bonzini | /* Must run outside of the iothread lock during the bulk phase,
|
114 | 32c835ba | Paolo Bonzini | * or the VM will stall.
|
115 | 32c835ba | Paolo Bonzini | */
|
116 | 32c835ba | Paolo Bonzini | |
117 | 13f0b67f | Jan Kiszka | static void blk_send(QEMUFile *f, BlkMigBlock * blk) |
118 | 13f0b67f | Jan Kiszka | { |
119 | 13f0b67f | Jan Kiszka | int len;
|
120 | 323004a3 | Peter Lieven | uint64_t flags = BLK_MIG_FLAG_DEVICE_BLOCK; |
121 | 323004a3 | Peter Lieven | |
122 | 323004a3 | Peter Lieven | if (block_mig_state.zero_blocks &&
|
123 | 323004a3 | Peter Lieven | buffer_is_zero(blk->buf, BLOCK_SIZE)) { |
124 | 323004a3 | Peter Lieven | flags |= BLK_MIG_FLAG_ZERO_BLOCK; |
125 | 323004a3 | Peter Lieven | } |
126 | 13f0b67f | Jan Kiszka | |
127 | 13f0b67f | Jan Kiszka | /* sector number and flags */
|
128 | 13f0b67f | Jan Kiszka | qemu_put_be64(f, (blk->sector << BDRV_SECTOR_BITS) |
129 | 323004a3 | Peter Lieven | | flags); |
130 | 13f0b67f | Jan Kiszka | |
131 | 13f0b67f | Jan Kiszka | /* device name */
|
132 | 13f0b67f | Jan Kiszka | len = strlen(blk->bmds->bs->device_name); |
133 | 13f0b67f | Jan Kiszka | qemu_put_byte(f, len); |
134 | 13f0b67f | Jan Kiszka | qemu_put_buffer(f, (uint8_t *)blk->bmds->bs->device_name, len); |
135 | 13f0b67f | Jan Kiszka | |
136 | 323004a3 | Peter Lieven | /* if a block is zero we need to flush here since the network
|
137 | 323004a3 | Peter Lieven | * bandwidth is now a lot higher than the storage device bandwidth.
|
138 | 323004a3 | Peter Lieven | * thus if we queue zero blocks we slow down the migration */
|
139 | 323004a3 | Peter Lieven | if (flags & BLK_MIG_FLAG_ZERO_BLOCK) {
|
140 | 323004a3 | Peter Lieven | qemu_fflush(f); |
141 | 323004a3 | Peter Lieven | return;
|
142 | 323004a3 | Peter Lieven | } |
143 | 323004a3 | Peter Lieven | |
144 | 13f0b67f | Jan Kiszka | qemu_put_buffer(f, blk->buf, BLOCK_SIZE); |
145 | 13f0b67f | Jan Kiszka | } |
146 | 13f0b67f | Jan Kiszka | |
147 | 25f23643 | Jan Kiszka | int blk_mig_active(void) |
148 | 25f23643 | Jan Kiszka | { |
149 | 25f23643 | Jan Kiszka | return !QSIMPLEQ_EMPTY(&block_mig_state.bmds_list);
|
150 | 25f23643 | Jan Kiszka | } |
151 | 25f23643 | Jan Kiszka | |
152 | 25f23643 | Jan Kiszka | uint64_t blk_mig_bytes_transferred(void)
|
153 | 25f23643 | Jan Kiszka | { |
154 | 25f23643 | Jan Kiszka | BlkMigDevState *bmds; |
155 | 25f23643 | Jan Kiszka | uint64_t sum = 0;
|
156 | 25f23643 | Jan Kiszka | |
157 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
158 | 25f23643 | Jan Kiszka | QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) { |
159 | 25f23643 | Jan Kiszka | sum += bmds->completed_sectors; |
160 | 25f23643 | Jan Kiszka | } |
161 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
162 | 25f23643 | Jan Kiszka | return sum << BDRV_SECTOR_BITS;
|
163 | 25f23643 | Jan Kiszka | } |
164 | 25f23643 | Jan Kiszka | |
165 | 25f23643 | Jan Kiszka | uint64_t blk_mig_bytes_remaining(void)
|
166 | 25f23643 | Jan Kiszka | { |
167 | 25f23643 | Jan Kiszka | return blk_mig_bytes_total() - blk_mig_bytes_transferred();
|
168 | 25f23643 | Jan Kiszka | } |
169 | 25f23643 | Jan Kiszka | |
170 | 25f23643 | Jan Kiszka | uint64_t blk_mig_bytes_total(void)
|
171 | 25f23643 | Jan Kiszka | { |
172 | 25f23643 | Jan Kiszka | BlkMigDevState *bmds; |
173 | 25f23643 | Jan Kiszka | uint64_t sum = 0;
|
174 | 25f23643 | Jan Kiszka | |
175 | 25f23643 | Jan Kiszka | QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) { |
176 | 25f23643 | Jan Kiszka | sum += bmds->total_sectors; |
177 | 25f23643 | Jan Kiszka | } |
178 | 25f23643 | Jan Kiszka | return sum << BDRV_SECTOR_BITS;
|
179 | 25f23643 | Jan Kiszka | } |
180 | 25f23643 | Jan Kiszka | |
181 | 52e850de | Paolo Bonzini | |
182 | 52e850de | Paolo Bonzini | /* Called with migration lock held. */
|
183 | 52e850de | Paolo Bonzini | |
184 | 33656af7 | Marcelo Tosatti | static int bmds_aio_inflight(BlkMigDevState *bmds, int64_t sector) |
185 | 33656af7 | Marcelo Tosatti | { |
186 | 33656af7 | Marcelo Tosatti | int64_t chunk = sector / (int64_t)BDRV_SECTORS_PER_DIRTY_CHUNK; |
187 | 33656af7 | Marcelo Tosatti | |
188 | 62155e2b | Marcelo Tosatti | if ((sector << BDRV_SECTOR_BITS) < bdrv_getlength(bmds->bs)) {
|
189 | 33656af7 | Marcelo Tosatti | return !!(bmds->aio_bitmap[chunk / (sizeof(unsigned long) * 8)] & |
190 | 33656af7 | Marcelo Tosatti | (1UL << (chunk % (sizeof(unsigned long) * 8)))); |
191 | 33656af7 | Marcelo Tosatti | } else {
|
192 | 33656af7 | Marcelo Tosatti | return 0; |
193 | 33656af7 | Marcelo Tosatti | } |
194 | 33656af7 | Marcelo Tosatti | } |
195 | 33656af7 | Marcelo Tosatti | |
196 | 52e850de | Paolo Bonzini | /* Called with migration lock held. */
|
197 | 52e850de | Paolo Bonzini | |
198 | 33656af7 | Marcelo Tosatti | static void bmds_set_aio_inflight(BlkMigDevState *bmds, int64_t sector_num, |
199 | 33656af7 | Marcelo Tosatti | int nb_sectors, int set) |
200 | 33656af7 | Marcelo Tosatti | { |
201 | 33656af7 | Marcelo Tosatti | int64_t start, end; |
202 | 33656af7 | Marcelo Tosatti | unsigned long val, idx, bit; |
203 | 33656af7 | Marcelo Tosatti | |
204 | 33656af7 | Marcelo Tosatti | start = sector_num / BDRV_SECTORS_PER_DIRTY_CHUNK; |
205 | 33656af7 | Marcelo Tosatti | end = (sector_num + nb_sectors - 1) / BDRV_SECTORS_PER_DIRTY_CHUNK;
|
206 | 33656af7 | Marcelo Tosatti | |
207 | 33656af7 | Marcelo Tosatti | for (; start <= end; start++) {
|
208 | 33656af7 | Marcelo Tosatti | idx = start / (sizeof(unsigned long) * 8); |
209 | 33656af7 | Marcelo Tosatti | bit = start % (sizeof(unsigned long) * 8); |
210 | 33656af7 | Marcelo Tosatti | val = bmds->aio_bitmap[idx]; |
211 | 33656af7 | Marcelo Tosatti | if (set) {
|
212 | 62155e2b | Marcelo Tosatti | val |= 1UL << bit;
|
213 | 33656af7 | Marcelo Tosatti | } else {
|
214 | 62155e2b | Marcelo Tosatti | val &= ~(1UL << bit);
|
215 | 33656af7 | Marcelo Tosatti | } |
216 | 33656af7 | Marcelo Tosatti | bmds->aio_bitmap[idx] = val; |
217 | 33656af7 | Marcelo Tosatti | } |
218 | 33656af7 | Marcelo Tosatti | } |
219 | 33656af7 | Marcelo Tosatti | |
220 | 33656af7 | Marcelo Tosatti | static void alloc_aio_bitmap(BlkMigDevState *bmds) |
221 | 33656af7 | Marcelo Tosatti | { |
222 | 33656af7 | Marcelo Tosatti | BlockDriverState *bs = bmds->bs; |
223 | 33656af7 | Marcelo Tosatti | int64_t bitmap_size; |
224 | 33656af7 | Marcelo Tosatti | |
225 | 33656af7 | Marcelo Tosatti | bitmap_size = (bdrv_getlength(bs) >> BDRV_SECTOR_BITS) + |
226 | 33656af7 | Marcelo Tosatti | BDRV_SECTORS_PER_DIRTY_CHUNK * 8 - 1; |
227 | 33656af7 | Marcelo Tosatti | bitmap_size /= BDRV_SECTORS_PER_DIRTY_CHUNK * 8;
|
228 | 33656af7 | Marcelo Tosatti | |
229 | 7267c094 | Anthony Liguori | bmds->aio_bitmap = g_malloc0(bitmap_size); |
230 | 33656af7 | Marcelo Tosatti | } |
231 | 33656af7 | Marcelo Tosatti | |
232 | 52e850de | Paolo Bonzini | /* Never hold migration lock when yielding to the main loop! */
|
233 | 52e850de | Paolo Bonzini | |
234 | c163b5ca | lirans@il.ibm.com | static void blk_mig_read_cb(void *opaque, int ret) |
235 | c163b5ca | lirans@il.ibm.com | { |
236 | c163b5ca | lirans@il.ibm.com | BlkMigBlock *blk = opaque; |
237 | a55eb92c | Jan Kiszka | |
238 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
239 | c163b5ca | lirans@il.ibm.com | blk->ret = ret; |
240 | a55eb92c | Jan Kiszka | |
241 | 5e5328be | Jan Kiszka | QSIMPLEQ_INSERT_TAIL(&block_mig_state.blk_list, blk, entry); |
242 | 33656af7 | Marcelo Tosatti | bmds_set_aio_inflight(blk->bmds, blk->sector, blk->nr_sectors, 0);
|
243 | a55eb92c | Jan Kiszka | |
244 | d11ecd3d | Jan Kiszka | block_mig_state.submitted--; |
245 | d11ecd3d | Jan Kiszka | block_mig_state.read_done++; |
246 | d11ecd3d | Jan Kiszka | assert(block_mig_state.submitted >= 0);
|
247 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
248 | c163b5ca | lirans@il.ibm.com | } |
249 | c163b5ca | lirans@il.ibm.com | |
250 | 32c835ba | Paolo Bonzini | /* Called with no lock taken. */
|
251 | 32c835ba | Paolo Bonzini | |
252 | 539de124 | Luiz Capitulino | static int mig_save_device_bulk(QEMUFile *f, BlkMigDevState *bmds) |
253 | a55eb92c | Jan Kiszka | { |
254 | 57cce12d | Jan Kiszka | int64_t total_sectors = bmds->total_sectors; |
255 | 57cce12d | Jan Kiszka | int64_t cur_sector = bmds->cur_sector; |
256 | 57cce12d | Jan Kiszka | BlockDriverState *bs = bmds->bs; |
257 | c163b5ca | lirans@il.ibm.com | BlkMigBlock *blk; |
258 | 13f0b67f | Jan Kiszka | int nr_sectors;
|
259 | a55eb92c | Jan Kiszka | |
260 | 57cce12d | Jan Kiszka | if (bmds->shared_base) {
|
261 | 32c835ba | Paolo Bonzini | qemu_mutex_lock_iothread(); |
262 | b1d10856 | Jan Kiszka | while (cur_sector < total_sectors &&
|
263 | 57cce12d | Jan Kiszka | !bdrv_is_allocated(bs, cur_sector, MAX_IS_ALLOCATED_SEARCH, |
264 | 57cce12d | Jan Kiszka | &nr_sectors)) { |
265 | c163b5ca | lirans@il.ibm.com | cur_sector += nr_sectors; |
266 | c163b5ca | lirans@il.ibm.com | } |
267 | 32c835ba | Paolo Bonzini | qemu_mutex_unlock_iothread(); |
268 | c163b5ca | lirans@il.ibm.com | } |
269 | a55eb92c | Jan Kiszka | |
270 | a55eb92c | Jan Kiszka | if (cur_sector >= total_sectors) {
|
271 | 82801d8f | Jan Kiszka | bmds->cur_sector = bmds->completed_sectors = total_sectors; |
272 | c163b5ca | lirans@il.ibm.com | return 1; |
273 | c163b5ca | lirans@il.ibm.com | } |
274 | a55eb92c | Jan Kiszka | |
275 | 82801d8f | Jan Kiszka | bmds->completed_sectors = cur_sector; |
276 | a55eb92c | Jan Kiszka | |
277 | 57cce12d | Jan Kiszka | cur_sector &= ~((int64_t)BDRV_SECTORS_PER_DIRTY_CHUNK - 1);
|
278 | 57cce12d | Jan Kiszka | |
279 | 6ea44308 | Jan Kiszka | /* we are going to transfer a full block even if it is not allocated */
|
280 | 6ea44308 | Jan Kiszka | nr_sectors = BDRV_SECTORS_PER_DIRTY_CHUNK; |
281 | c163b5ca | lirans@il.ibm.com | |
282 | 6ea44308 | Jan Kiszka | if (total_sectors - cur_sector < BDRV_SECTORS_PER_DIRTY_CHUNK) {
|
283 | 57cce12d | Jan Kiszka | nr_sectors = total_sectors - cur_sector; |
284 | c163b5ca | lirans@il.ibm.com | } |
285 | a55eb92c | Jan Kiszka | |
286 | 7267c094 | Anthony Liguori | blk = g_malloc(sizeof(BlkMigBlock));
|
287 | 7267c094 | Anthony Liguori | blk->buf = g_malloc(BLOCK_SIZE); |
288 | 13f0b67f | Jan Kiszka | blk->bmds = bmds; |
289 | 13f0b67f | Jan Kiszka | blk->sector = cur_sector; |
290 | 33656af7 | Marcelo Tosatti | blk->nr_sectors = nr_sectors; |
291 | a55eb92c | Jan Kiszka | |
292 | e970ec0b | Liran Schour | blk->iov.iov_base = blk->buf; |
293 | e970ec0b | Liran Schour | blk->iov.iov_len = nr_sectors * BDRV_SECTOR_SIZE; |
294 | e970ec0b | Liran Schour | qemu_iovec_init_external(&blk->qiov, &blk->iov, 1);
|
295 | a55eb92c | Jan Kiszka | |
296 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
297 | 13197e3c | Paolo Bonzini | block_mig_state.submitted++; |
298 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
299 | 13197e3c | Paolo Bonzini | |
300 | 32c835ba | Paolo Bonzini | qemu_mutex_lock_iothread(); |
301 | e970ec0b | Liran Schour | blk->aiocb = bdrv_aio_readv(bs, cur_sector, &blk->qiov, |
302 | e970ec0b | Liran Schour | nr_sectors, blk_mig_read_cb, blk); |
303 | d76cac7d | Liran Schour | |
304 | 13f0b67f | Jan Kiszka | bdrv_reset_dirty(bs, cur_sector, nr_sectors); |
305 | 32c835ba | Paolo Bonzini | qemu_mutex_unlock_iothread(); |
306 | a55eb92c | Jan Kiszka | |
307 | 32c835ba | Paolo Bonzini | bmds->cur_sector = cur_sector + nr_sectors; |
308 | 13f0b67f | Jan Kiszka | return (bmds->cur_sector >= total_sectors);
|
309 | c163b5ca | lirans@il.ibm.com | } |
310 | c163b5ca | lirans@il.ibm.com | |
311 | 32c835ba | Paolo Bonzini | /* Called with iothread lock taken. */
|
312 | 32c835ba | Paolo Bonzini | |
313 | e4654d2d | Fam Zheng | static void set_dirty_tracking(void) |
314 | c163b5ca | lirans@il.ibm.com | { |
315 | c163b5ca | lirans@il.ibm.com | BlkMigDevState *bmds; |
316 | 5e5328be | Jan Kiszka | |
317 | 5e5328be | Jan Kiszka | QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) { |
318 | e4654d2d | Fam Zheng | bmds->dirty_bitmap = bdrv_create_dirty_bitmap(bmds->bs, BLOCK_SIZE); |
319 | e4654d2d | Fam Zheng | } |
320 | e4654d2d | Fam Zheng | } |
321 | e4654d2d | Fam Zheng | |
322 | e4654d2d | Fam Zheng | static void unset_dirty_tracking(void) |
323 | e4654d2d | Fam Zheng | { |
324 | e4654d2d | Fam Zheng | BlkMigDevState *bmds; |
325 | e4654d2d | Fam Zheng | |
326 | e4654d2d | Fam Zheng | QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) { |
327 | e4654d2d | Fam Zheng | bdrv_release_dirty_bitmap(bmds->bs, bmds->dirty_bitmap); |
328 | c163b5ca | lirans@il.ibm.com | } |
329 | c163b5ca | lirans@il.ibm.com | } |
330 | c163b5ca | lirans@il.ibm.com | |
331 | b66460e4 | Stefan Hajnoczi | static void init_blk_migration_it(void *opaque, BlockDriverState *bs) |
332 | c163b5ca | lirans@il.ibm.com | { |
333 | 5e5328be | Jan Kiszka | BlkMigDevState *bmds; |
334 | 792773b2 | Jan Kiszka | int64_t sectors; |
335 | a55eb92c | Jan Kiszka | |
336 | d246673d | Markus Armbruster | if (!bdrv_is_read_only(bs)) {
|
337 | b66460e4 | Stefan Hajnoczi | sectors = bdrv_getlength(bs) >> BDRV_SECTOR_BITS; |
338 | 31f54f24 | Shahar Havivi | if (sectors <= 0) { |
339 | b66460e4 | Stefan Hajnoczi | return;
|
340 | b66460e4 | Stefan Hajnoczi | } |
341 | b66460e4 | Stefan Hajnoczi | |
342 | 7267c094 | Anthony Liguori | bmds = g_malloc0(sizeof(BlkMigDevState));
|
343 | b66460e4 | Stefan Hajnoczi | bmds->bs = bs; |
344 | b66460e4 | Stefan Hajnoczi | bmds->bulk_completed = 0;
|
345 | b66460e4 | Stefan Hajnoczi | bmds->total_sectors = sectors; |
346 | b66460e4 | Stefan Hajnoczi | bmds->completed_sectors = 0;
|
347 | b66460e4 | Stefan Hajnoczi | bmds->shared_base = block_mig_state.shared_base; |
348 | 33656af7 | Marcelo Tosatti | alloc_aio_bitmap(bmds); |
349 | 8591675f | Marcelo Tosatti | bdrv_set_in_use(bs, 1);
|
350 | 8442cfd0 | Fam Zheng | bdrv_ref(bs); |
351 | b66460e4 | Stefan Hajnoczi | |
352 | b66460e4 | Stefan Hajnoczi | block_mig_state.total_sector_sum += sectors; |
353 | b66460e4 | Stefan Hajnoczi | |
354 | b66460e4 | Stefan Hajnoczi | if (bmds->shared_base) {
|
355 | 539de124 | Luiz Capitulino | DPRINTF("Start migration for %s with shared base image\n",
|
356 | 539de124 | Luiz Capitulino | bs->device_name); |
357 | b66460e4 | Stefan Hajnoczi | } else {
|
358 | 539de124 | Luiz Capitulino | DPRINTF("Start full migration for %s\n", bs->device_name);
|
359 | b66460e4 | Stefan Hajnoczi | } |
360 | b66460e4 | Stefan Hajnoczi | |
361 | b66460e4 | Stefan Hajnoczi | QSIMPLEQ_INSERT_TAIL(&block_mig_state.bmds_list, bmds, entry); |
362 | b66460e4 | Stefan Hajnoczi | } |
363 | b66460e4 | Stefan Hajnoczi | } |
364 | b66460e4 | Stefan Hajnoczi | |
365 | 539de124 | Luiz Capitulino | static void init_blk_migration(QEMUFile *f) |
366 | b66460e4 | Stefan Hajnoczi | { |
367 | 69d63a97 | Jan Kiszka | block_mig_state.submitted = 0;
|
368 | 69d63a97 | Jan Kiszka | block_mig_state.read_done = 0;
|
369 | 69d63a97 | Jan Kiszka | block_mig_state.transferred = 0;
|
370 | 82801d8f | Jan Kiszka | block_mig_state.total_sector_sum = 0;
|
371 | 01e61e2d | Jan Kiszka | block_mig_state.prev_progress = -1;
|
372 | e970ec0b | Liran Schour | block_mig_state.bulk_completed = 0;
|
373 | 323004a3 | Peter Lieven | block_mig_state.zero_blocks = migrate_zero_blocks(); |
374 | 69d63a97 | Jan Kiszka | |
375 | 539de124 | Luiz Capitulino | bdrv_iterate(init_blk_migration_it, NULL);
|
376 | c163b5ca | lirans@il.ibm.com | } |
377 | c163b5ca | lirans@il.ibm.com | |
378 | 32c835ba | Paolo Bonzini | /* Called with no lock taken. */
|
379 | 32c835ba | Paolo Bonzini | |
380 | 539de124 | Luiz Capitulino | static int blk_mig_save_bulked_block(QEMUFile *f) |
381 | c163b5ca | lirans@il.ibm.com | { |
382 | 82801d8f | Jan Kiszka | int64_t completed_sector_sum = 0;
|
383 | c163b5ca | lirans@il.ibm.com | BlkMigDevState *bmds; |
384 | 01e61e2d | Jan Kiszka | int progress;
|
385 | 82801d8f | Jan Kiszka | int ret = 0; |
386 | c163b5ca | lirans@il.ibm.com | |
387 | 5e5328be | Jan Kiszka | QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) { |
388 | a55eb92c | Jan Kiszka | if (bmds->bulk_completed == 0) { |
389 | 539de124 | Luiz Capitulino | if (mig_save_device_bulk(f, bmds) == 1) { |
390 | 57cce12d | Jan Kiszka | /* completed bulk section for this device */
|
391 | 57cce12d | Jan Kiszka | bmds->bulk_completed = 1;
|
392 | c163b5ca | lirans@il.ibm.com | } |
393 | 82801d8f | Jan Kiszka | completed_sector_sum += bmds->completed_sectors; |
394 | 82801d8f | Jan Kiszka | ret = 1;
|
395 | 82801d8f | Jan Kiszka | break;
|
396 | 82801d8f | Jan Kiszka | } else {
|
397 | 82801d8f | Jan Kiszka | completed_sector_sum += bmds->completed_sectors; |
398 | c163b5ca | lirans@il.ibm.com | } |
399 | c163b5ca | lirans@il.ibm.com | } |
400 | a55eb92c | Jan Kiszka | |
401 | 8b6b2afc | Pierre Riteau | if (block_mig_state.total_sector_sum != 0) { |
402 | 8b6b2afc | Pierre Riteau | progress = completed_sector_sum * 100 /
|
403 | 8b6b2afc | Pierre Riteau | block_mig_state.total_sector_sum; |
404 | 8b6b2afc | Pierre Riteau | } else {
|
405 | 8b6b2afc | Pierre Riteau | progress = 100;
|
406 | 8b6b2afc | Pierre Riteau | } |
407 | 01e61e2d | Jan Kiszka | if (progress != block_mig_state.prev_progress) {
|
408 | 01e61e2d | Jan Kiszka | block_mig_state.prev_progress = progress; |
409 | 01e61e2d | Jan Kiszka | qemu_put_be64(f, (progress << BDRV_SECTOR_BITS) |
410 | 01e61e2d | Jan Kiszka | | BLK_MIG_FLAG_PROGRESS); |
411 | 539de124 | Luiz Capitulino | DPRINTF("Completed %d %%\r", progress);
|
412 | 82801d8f | Jan Kiszka | } |
413 | 82801d8f | Jan Kiszka | |
414 | 82801d8f | Jan Kiszka | return ret;
|
415 | c163b5ca | lirans@il.ibm.com | } |
416 | c163b5ca | lirans@il.ibm.com | |
417 | d76cac7d | Liran Schour | static void blk_mig_reset_dirty_cursor(void) |
418 | c163b5ca | lirans@il.ibm.com | { |
419 | c163b5ca | lirans@il.ibm.com | BlkMigDevState *bmds; |
420 | d76cac7d | Liran Schour | |
421 | d76cac7d | Liran Schour | QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) { |
422 | d76cac7d | Liran Schour | bmds->cur_dirty = 0;
|
423 | d76cac7d | Liran Schour | } |
424 | d76cac7d | Liran Schour | } |
425 | d76cac7d | Liran Schour | |
426 | 32c835ba | Paolo Bonzini | /* Called with iothread lock taken. */
|
427 | 32c835ba | Paolo Bonzini | |
428 | 539de124 | Luiz Capitulino | static int mig_save_device_dirty(QEMUFile *f, BlkMigDevState *bmds, |
429 | 539de124 | Luiz Capitulino | int is_async)
|
430 | d76cac7d | Liran Schour | { |
431 | d76cac7d | Liran Schour | BlkMigBlock *blk; |
432 | d76cac7d | Liran Schour | int64_t total_sectors = bmds->total_sectors; |
433 | c163b5ca | lirans@il.ibm.com | int64_t sector; |
434 | d76cac7d | Liran Schour | int nr_sectors;
|
435 | dcd1d224 | Juan Quintela | int ret = -EIO;
|
436 | a55eb92c | Jan Kiszka | |
437 | d76cac7d | Liran Schour | for (sector = bmds->cur_dirty; sector < bmds->total_sectors;) {
|
438 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
439 | 62155e2b | Marcelo Tosatti | if (bmds_aio_inflight(bmds, sector)) {
|
440 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
441 | 922453bc | Stefan Hajnoczi | bdrv_drain_all(); |
442 | 52e850de | Paolo Bonzini | } else {
|
443 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
444 | 62155e2b | Marcelo Tosatti | } |
445 | e4654d2d | Fam Zheng | if (bdrv_get_dirty(bmds->bs, bmds->dirty_bitmap, sector)) {
|
446 | 575a58d7 | Jan Kiszka | |
447 | d76cac7d | Liran Schour | if (total_sectors - sector < BDRV_SECTORS_PER_DIRTY_CHUNK) {
|
448 | d76cac7d | Liran Schour | nr_sectors = total_sectors - sector; |
449 | d76cac7d | Liran Schour | } else {
|
450 | d76cac7d | Liran Schour | nr_sectors = BDRV_SECTORS_PER_DIRTY_CHUNK; |
451 | d76cac7d | Liran Schour | } |
452 | 7267c094 | Anthony Liguori | blk = g_malloc(sizeof(BlkMigBlock));
|
453 | 7267c094 | Anthony Liguori | blk->buf = g_malloc(BLOCK_SIZE); |
454 | d76cac7d | Liran Schour | blk->bmds = bmds; |
455 | d76cac7d | Liran Schour | blk->sector = sector; |
456 | 33656af7 | Marcelo Tosatti | blk->nr_sectors = nr_sectors; |
457 | d76cac7d | Liran Schour | |
458 | 889ae39c | Liran Schour | if (is_async) {
|
459 | d76cac7d | Liran Schour | blk->iov.iov_base = blk->buf; |
460 | d76cac7d | Liran Schour | blk->iov.iov_len = nr_sectors * BDRV_SECTOR_SIZE; |
461 | d76cac7d | Liran Schour | qemu_iovec_init_external(&blk->qiov, &blk->iov, 1);
|
462 | d76cac7d | Liran Schour | |
463 | d76cac7d | Liran Schour | blk->aiocb = bdrv_aio_readv(bmds->bs, sector, &blk->qiov, |
464 | d76cac7d | Liran Schour | nr_sectors, blk_mig_read_cb, blk); |
465 | 52e850de | Paolo Bonzini | |
466 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
467 | d76cac7d | Liran Schour | block_mig_state.submitted++; |
468 | 33656af7 | Marcelo Tosatti | bmds_set_aio_inflight(bmds, sector, nr_sectors, 1);
|
469 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
470 | d76cac7d | Liran Schour | } else {
|
471 | dcd1d224 | Juan Quintela | ret = bdrv_read(bmds->bs, sector, blk->buf, nr_sectors); |
472 | dcd1d224 | Juan Quintela | if (ret < 0) { |
473 | d76cac7d | Liran Schour | goto error;
|
474 | c163b5ca | lirans@il.ibm.com | } |
475 | d76cac7d | Liran Schour | blk_send(f, blk); |
476 | a55eb92c | Jan Kiszka | |
477 | 7267c094 | Anthony Liguori | g_free(blk->buf); |
478 | 7267c094 | Anthony Liguori | g_free(blk); |
479 | a55eb92c | Jan Kiszka | } |
480 | d76cac7d | Liran Schour | |
481 | d76cac7d | Liran Schour | bdrv_reset_dirty(bmds->bs, sector, nr_sectors); |
482 | d76cac7d | Liran Schour | break;
|
483 | c163b5ca | lirans@il.ibm.com | } |
484 | d76cac7d | Liran Schour | sector += BDRV_SECTORS_PER_DIRTY_CHUNK; |
485 | d76cac7d | Liran Schour | bmds->cur_dirty = sector; |
486 | c163b5ca | lirans@il.ibm.com | } |
487 | 575a58d7 | Jan Kiszka | |
488 | d76cac7d | Liran Schour | return (bmds->cur_dirty >= bmds->total_sectors);
|
489 | d76cac7d | Liran Schour | |
490 | 889ae39c | Liran Schour | error:
|
491 | 539de124 | Luiz Capitulino | DPRINTF("Error reading sector %" PRId64 "\n", sector); |
492 | 7267c094 | Anthony Liguori | g_free(blk->buf); |
493 | 7267c094 | Anthony Liguori | g_free(blk); |
494 | 43be3a25 | Juan Quintela | return ret;
|
495 | d76cac7d | Liran Schour | } |
496 | d76cac7d | Liran Schour | |
497 | 32c835ba | Paolo Bonzini | /* Called with iothread lock taken.
|
498 | 32c835ba | Paolo Bonzini | *
|
499 | 32c835ba | Paolo Bonzini | * return value:
|
500 | ceb2bd09 | Juan Quintela | * 0: too much data for max_downtime
|
501 | ceb2bd09 | Juan Quintela | * 1: few enough data for max_downtime
|
502 | ceb2bd09 | Juan Quintela | */
|
503 | 539de124 | Luiz Capitulino | static int blk_mig_save_dirty_block(QEMUFile *f, int is_async) |
504 | d76cac7d | Liran Schour | { |
505 | d76cac7d | Liran Schour | BlkMigDevState *bmds; |
506 | ceb2bd09 | Juan Quintela | int ret = 1; |
507 | d76cac7d | Liran Schour | |
508 | d76cac7d | Liran Schour | QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) { |
509 | ceb2bd09 | Juan Quintela | ret = mig_save_device_dirty(f, bmds, is_async); |
510 | 43be3a25 | Juan Quintela | if (ret <= 0) { |
511 | d76cac7d | Liran Schour | break;
|
512 | d76cac7d | Liran Schour | } |
513 | d76cac7d | Liran Schour | } |
514 | d76cac7d | Liran Schour | |
515 | d76cac7d | Liran Schour | return ret;
|
516 | c163b5ca | lirans@il.ibm.com | } |
517 | c163b5ca | lirans@il.ibm.com | |
518 | 32c835ba | Paolo Bonzini | /* Called with no locks taken. */
|
519 | 32c835ba | Paolo Bonzini | |
520 | 59feec42 | Juan Quintela | static int flush_blks(QEMUFile *f) |
521 | c163b5ca | lirans@il.ibm.com | { |
522 | 5e5328be | Jan Kiszka | BlkMigBlock *blk; |
523 | 59feec42 | Juan Quintela | int ret = 0; |
524 | a55eb92c | Jan Kiszka | |
525 | d0f2c4c6 | malc | DPRINTF("%s Enter submitted %d read_done %d transferred %d\n",
|
526 | d11ecd3d | Jan Kiszka | __FUNCTION__, block_mig_state.submitted, block_mig_state.read_done, |
527 | d11ecd3d | Jan Kiszka | block_mig_state.transferred); |
528 | a55eb92c | Jan Kiszka | |
529 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
530 | 5e5328be | Jan Kiszka | while ((blk = QSIMPLEQ_FIRST(&block_mig_state.blk_list)) != NULL) { |
531 | 5e5328be | Jan Kiszka | if (qemu_file_rate_limit(f)) {
|
532 | 5e5328be | Jan Kiszka | break;
|
533 | 5e5328be | Jan Kiszka | } |
534 | 4b640365 | Jan Kiszka | if (blk->ret < 0) { |
535 | 59feec42 | Juan Quintela | ret = blk->ret; |
536 | 4b640365 | Jan Kiszka | break;
|
537 | 4b640365 | Jan Kiszka | } |
538 | a55eb92c | Jan Kiszka | |
539 | 5e5328be | Jan Kiszka | QSIMPLEQ_REMOVE_HEAD(&block_mig_state.blk_list, entry); |
540 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
541 | 13197e3c | Paolo Bonzini | blk_send(f, blk); |
542 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
543 | 13197e3c | Paolo Bonzini | |
544 | 7267c094 | Anthony Liguori | g_free(blk->buf); |
545 | 7267c094 | Anthony Liguori | g_free(blk); |
546 | a55eb92c | Jan Kiszka | |
547 | d11ecd3d | Jan Kiszka | block_mig_state.read_done--; |
548 | d11ecd3d | Jan Kiszka | block_mig_state.transferred++; |
549 | d11ecd3d | Jan Kiszka | assert(block_mig_state.read_done >= 0);
|
550 | c163b5ca | lirans@il.ibm.com | } |
551 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
552 | c163b5ca | lirans@il.ibm.com | |
553 | d0f2c4c6 | malc | DPRINTF("%s Exit submitted %d read_done %d transferred %d\n", __FUNCTION__,
|
554 | d11ecd3d | Jan Kiszka | block_mig_state.submitted, block_mig_state.read_done, |
555 | d11ecd3d | Jan Kiszka | block_mig_state.transferred); |
556 | 59feec42 | Juan Quintela | return ret;
|
557 | c163b5ca | lirans@il.ibm.com | } |
558 | c163b5ca | lirans@il.ibm.com | |
559 | 32c835ba | Paolo Bonzini | /* Called with iothread lock taken. */
|
560 | 32c835ba | Paolo Bonzini | |
561 | 889ae39c | Liran Schour | static int64_t get_remaining_dirty(void) |
562 | 889ae39c | Liran Schour | { |
563 | 889ae39c | Liran Schour | BlkMigDevState *bmds; |
564 | 889ae39c | Liran Schour | int64_t dirty = 0;
|
565 | 889ae39c | Liran Schour | |
566 | 889ae39c | Liran Schour | QSIMPLEQ_FOREACH(bmds, &block_mig_state.bmds_list, entry) { |
567 | e4654d2d | Fam Zheng | dirty += bdrv_get_dirty_count(bmds->bs, bmds->dirty_bitmap); |
568 | 889ae39c | Liran Schour | } |
569 | 889ae39c | Liran Schour | |
570 | acc906c6 | Paolo Bonzini | return dirty << BDRV_SECTOR_BITS;
|
571 | 889ae39c | Liran Schour | } |
572 | 889ae39c | Liran Schour | |
573 | 32c835ba | Paolo Bonzini | /* Called with iothread lock taken. */
|
574 | 32c835ba | Paolo Bonzini | |
575 | 539de124 | Luiz Capitulino | static void blk_mig_cleanup(void) |
576 | 4ec7fcc7 | Jan Kiszka | { |
577 | 82801d8f | Jan Kiszka | BlkMigDevState *bmds; |
578 | 82801d8f | Jan Kiszka | BlkMigBlock *blk; |
579 | 4ec7fcc7 | Jan Kiszka | |
580 | 946d58be | Kevin Wolf | bdrv_drain_all(); |
581 | 946d58be | Kevin Wolf | |
582 | e4654d2d | Fam Zheng | unset_dirty_tracking(); |
583 | 8f794c55 | Marcelo Tosatti | |
584 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
585 | 82801d8f | Jan Kiszka | while ((bmds = QSIMPLEQ_FIRST(&block_mig_state.bmds_list)) != NULL) { |
586 | 82801d8f | Jan Kiszka | QSIMPLEQ_REMOVE_HEAD(&block_mig_state.bmds_list, entry); |
587 | 8591675f | Marcelo Tosatti | bdrv_set_in_use(bmds->bs, 0);
|
588 | 8442cfd0 | Fam Zheng | bdrv_unref(bmds->bs); |
589 | 7267c094 | Anthony Liguori | g_free(bmds->aio_bitmap); |
590 | 7267c094 | Anthony Liguori | g_free(bmds); |
591 | 4ec7fcc7 | Jan Kiszka | } |
592 | 4ec7fcc7 | Jan Kiszka | |
593 | 82801d8f | Jan Kiszka | while ((blk = QSIMPLEQ_FIRST(&block_mig_state.blk_list)) != NULL) { |
594 | 82801d8f | Jan Kiszka | QSIMPLEQ_REMOVE_HEAD(&block_mig_state.blk_list, entry); |
595 | 7267c094 | Anthony Liguori | g_free(blk->buf); |
596 | 7267c094 | Anthony Liguori | g_free(blk); |
597 | 4ec7fcc7 | Jan Kiszka | } |
598 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
599 | 4ec7fcc7 | Jan Kiszka | } |
600 | 4ec7fcc7 | Jan Kiszka | |
601 | 9b5bfab0 | Juan Quintela | static void block_migration_cancel(void *opaque) |
602 | 9b5bfab0 | Juan Quintela | { |
603 | 9b5bfab0 | Juan Quintela | blk_mig_cleanup(); |
604 | 9b5bfab0 | Juan Quintela | } |
605 | 9b5bfab0 | Juan Quintela | |
606 | d1315aac | Juan Quintela | static int block_save_setup(QEMUFile *f, void *opaque) |
607 | c163b5ca | lirans@il.ibm.com | { |
608 | 2975725f | Juan Quintela | int ret;
|
609 | 2975725f | Juan Quintela | |
610 | d1315aac | Juan Quintela | DPRINTF("Enter save live setup submitted %d transferred %d\n",
|
611 | d1315aac | Juan Quintela | block_mig_state.submitted, block_mig_state.transferred); |
612 | a55eb92c | Jan Kiszka | |
613 | 9b095037 | Paolo Bonzini | qemu_mutex_lock_iothread(); |
614 | d1315aac | Juan Quintela | init_blk_migration(f); |
615 | d1315aac | Juan Quintela | |
616 | d1315aac | Juan Quintela | /* start track dirty blocks */
|
617 | e4654d2d | Fam Zheng | set_dirty_tracking(); |
618 | 9b095037 | Paolo Bonzini | qemu_mutex_unlock_iothread(); |
619 | d1315aac | Juan Quintela | |
620 | 59feec42 | Juan Quintela | ret = flush_blks(f); |
621 | d1315aac | Juan Quintela | blk_mig_reset_dirty_cursor(); |
622 | d1315aac | Juan Quintela | qemu_put_be64(f, BLK_MIG_FLAG_EOS); |
623 | d1315aac | Juan Quintela | |
624 | d418cf57 | Paolo Bonzini | return ret;
|
625 | d1315aac | Juan Quintela | } |
626 | d1315aac | Juan Quintela | |
627 | 16310a3c | Juan Quintela | static int block_save_iterate(QEMUFile *f, void *opaque) |
628 | d1315aac | Juan Quintela | { |
629 | d1315aac | Juan Quintela | int ret;
|
630 | 6aaa9dae | Stefan Hajnoczi | int64_t last_ftell = qemu_ftell(f); |
631 | d1315aac | Juan Quintela | |
632 | 16310a3c | Juan Quintela | DPRINTF("Enter save live iterate submitted %d transferred %d\n",
|
633 | 16310a3c | Juan Quintela | block_mig_state.submitted, block_mig_state.transferred); |
634 | d1315aac | Juan Quintela | |
635 | 59feec42 | Juan Quintela | ret = flush_blks(f); |
636 | 2975725f | Juan Quintela | if (ret) {
|
637 | 2975725f | Juan Quintela | return ret;
|
638 | 4b640365 | Jan Kiszka | } |
639 | 4b640365 | Jan Kiszka | |
640 | d76cac7d | Liran Schour | blk_mig_reset_dirty_cursor(); |
641 | d76cac7d | Liran Schour | |
642 | 16310a3c | Juan Quintela | /* control the rate of transfer */
|
643 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
644 | 16310a3c | Juan Quintela | while ((block_mig_state.submitted +
|
645 | 16310a3c | Juan Quintela | block_mig_state.read_done) * BLOCK_SIZE < |
646 | 16310a3c | Juan Quintela | qemu_file_get_rate_limit(f)) { |
647 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
648 | 16310a3c | Juan Quintela | if (block_mig_state.bulk_completed == 0) { |
649 | 16310a3c | Juan Quintela | /* first finish the bulk phase */
|
650 | 16310a3c | Juan Quintela | if (blk_mig_save_bulked_block(f) == 0) { |
651 | 16310a3c | Juan Quintela | /* finished saving bulk on all devices */
|
652 | 16310a3c | Juan Quintela | block_mig_state.bulk_completed = 1;
|
653 | 16310a3c | Juan Quintela | } |
654 | 13197e3c | Paolo Bonzini | ret = 0;
|
655 | 16310a3c | Juan Quintela | } else {
|
656 | 32c835ba | Paolo Bonzini | /* Always called with iothread lock taken for
|
657 | 32c835ba | Paolo Bonzini | * simplicity, block_save_complete also calls it.
|
658 | 32c835ba | Paolo Bonzini | */
|
659 | 32c835ba | Paolo Bonzini | qemu_mutex_lock_iothread(); |
660 | 43be3a25 | Juan Quintela | ret = blk_mig_save_dirty_block(f, 1);
|
661 | 32c835ba | Paolo Bonzini | qemu_mutex_unlock_iothread(); |
662 | 13197e3c | Paolo Bonzini | } |
663 | 13197e3c | Paolo Bonzini | if (ret < 0) { |
664 | 13197e3c | Paolo Bonzini | return ret;
|
665 | 13197e3c | Paolo Bonzini | } |
666 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
667 | 13197e3c | Paolo Bonzini | if (ret != 0) { |
668 | 13197e3c | Paolo Bonzini | /* no more dirty blocks */
|
669 | 13197e3c | Paolo Bonzini | break;
|
670 | a55eb92c | Jan Kiszka | } |
671 | 16310a3c | Juan Quintela | } |
672 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
673 | a55eb92c | Jan Kiszka | |
674 | 59feec42 | Juan Quintela | ret = flush_blks(f); |
675 | 16310a3c | Juan Quintela | if (ret) {
|
676 | 16310a3c | Juan Quintela | return ret;
|
677 | 4b640365 | Jan Kiszka | } |
678 | 4b640365 | Jan Kiszka | |
679 | 16310a3c | Juan Quintela | qemu_put_be64(f, BLK_MIG_FLAG_EOS); |
680 | 6aaa9dae | Stefan Hajnoczi | return qemu_ftell(f) - last_ftell;
|
681 | 16310a3c | Juan Quintela | } |
682 | 16310a3c | Juan Quintela | |
683 | 32c835ba | Paolo Bonzini | /* Called with iothread lock taken. */
|
684 | 32c835ba | Paolo Bonzini | |
685 | 16310a3c | Juan Quintela | static int block_save_complete(QEMUFile *f, void *opaque) |
686 | 16310a3c | Juan Quintela | { |
687 | 16310a3c | Juan Quintela | int ret;
|
688 | 16310a3c | Juan Quintela | |
689 | 16310a3c | Juan Quintela | DPRINTF("Enter save live complete submitted %d transferred %d\n",
|
690 | 16310a3c | Juan Quintela | block_mig_state.submitted, block_mig_state.transferred); |
691 | 16310a3c | Juan Quintela | |
692 | 59feec42 | Juan Quintela | ret = flush_blks(f); |
693 | 16310a3c | Juan Quintela | if (ret) {
|
694 | 16310a3c | Juan Quintela | return ret;
|
695 | 16310a3c | Juan Quintela | } |
696 | a55eb92c | Jan Kiszka | |
697 | 16310a3c | Juan Quintela | blk_mig_reset_dirty_cursor(); |
698 | 01e61e2d | Jan Kiszka | |
699 | 16310a3c | Juan Quintela | /* we know for sure that save bulk is completed and
|
700 | 16310a3c | Juan Quintela | all async read completed */
|
701 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
702 | 16310a3c | Juan Quintela | assert(block_mig_state.submitted == 0);
|
703 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
704 | 16310a3c | Juan Quintela | |
705 | 43be3a25 | Juan Quintela | do {
|
706 | 43be3a25 | Juan Quintela | ret = blk_mig_save_dirty_block(f, 0);
|
707 | d418cf57 | Paolo Bonzini | if (ret < 0) { |
708 | d418cf57 | Paolo Bonzini | return ret;
|
709 | d418cf57 | Paolo Bonzini | } |
710 | 43be3a25 | Juan Quintela | } while (ret == 0); |
711 | 4b640365 | Jan Kiszka | |
712 | 43be3a25 | Juan Quintela | /* report completion */
|
713 | 43be3a25 | Juan Quintela | qemu_put_be64(f, (100 << BDRV_SECTOR_BITS) | BLK_MIG_FLAG_PROGRESS);
|
714 | a55eb92c | Jan Kiszka | |
715 | 16310a3c | Juan Quintela | DPRINTF("Block migration completed\n");
|
716 | 16310a3c | Juan Quintela | |
717 | a55eb92c | Jan Kiszka | qemu_put_be64(f, BLK_MIG_FLAG_EOS); |
718 | a55eb92c | Jan Kiszka | |
719 | d418cf57 | Paolo Bonzini | blk_mig_cleanup(); |
720 | 16310a3c | Juan Quintela | return 0; |
721 | c163b5ca | lirans@il.ibm.com | } |
722 | c163b5ca | lirans@il.ibm.com | |
723 | e4ed1541 | Juan Quintela | static uint64_t block_save_pending(QEMUFile *f, void *opaque, uint64_t max_size) |
724 | e4ed1541 | Juan Quintela | { |
725 | 6aaa9dae | Stefan Hajnoczi | /* Estimate pending number of bytes to send */
|
726 | 13197e3c | Paolo Bonzini | uint64_t pending; |
727 | 13197e3c | Paolo Bonzini | |
728 | 32c835ba | Paolo Bonzini | qemu_mutex_lock_iothread(); |
729 | 52e850de | Paolo Bonzini | blk_mig_lock(); |
730 | 13197e3c | Paolo Bonzini | pending = get_remaining_dirty() + |
731 | 6aaa9dae | Stefan Hajnoczi | block_mig_state.submitted * BLOCK_SIZE + |
732 | 6aaa9dae | Stefan Hajnoczi | block_mig_state.read_done * BLOCK_SIZE; |
733 | 6aaa9dae | Stefan Hajnoczi | |
734 | 6aaa9dae | Stefan Hajnoczi | /* Report at least one block pending during bulk phase */
|
735 | 6aaa9dae | Stefan Hajnoczi | if (pending == 0 && !block_mig_state.bulk_completed) { |
736 | 6aaa9dae | Stefan Hajnoczi | pending = BLOCK_SIZE; |
737 | 6aaa9dae | Stefan Hajnoczi | } |
738 | 52e850de | Paolo Bonzini | blk_mig_unlock(); |
739 | 32c835ba | Paolo Bonzini | qemu_mutex_unlock_iothread(); |
740 | e4ed1541 | Juan Quintela | |
741 | 6aaa9dae | Stefan Hajnoczi | DPRINTF("Enter save live pending %" PRIu64 "\n", pending); |
742 | 6aaa9dae | Stefan Hajnoczi | return pending;
|
743 | e4ed1541 | Juan Quintela | } |
744 | e4ed1541 | Juan Quintela | |
745 | c163b5ca | lirans@il.ibm.com | static int block_load(QEMUFile *f, void *opaque, int version_id) |
746 | c163b5ca | lirans@il.ibm.com | { |
747 | 01e61e2d | Jan Kiszka | static int banner_printed; |
748 | c163b5ca | lirans@il.ibm.com | int len, flags;
|
749 | c163b5ca | lirans@il.ibm.com | char device_name[256]; |
750 | c163b5ca | lirans@il.ibm.com | int64_t addr; |
751 | 77358b59 | Pierre Riteau | BlockDriverState *bs, *bs_prev = NULL;
|
752 | c163b5ca | lirans@il.ibm.com | uint8_t *buf; |
753 | 77358b59 | Pierre Riteau | int64_t total_sectors = 0;
|
754 | 77358b59 | Pierre Riteau | int nr_sectors;
|
755 | 42802d47 | Juan Quintela | int ret;
|
756 | a55eb92c | Jan Kiszka | |
757 | c163b5ca | lirans@il.ibm.com | do {
|
758 | c163b5ca | lirans@il.ibm.com | addr = qemu_get_be64(f); |
759 | a55eb92c | Jan Kiszka | |
760 | 6ea44308 | Jan Kiszka | flags = addr & ~BDRV_SECTOR_MASK; |
761 | 6ea44308 | Jan Kiszka | addr >>= BDRV_SECTOR_BITS; |
762 | a55eb92c | Jan Kiszka | |
763 | a55eb92c | Jan Kiszka | if (flags & BLK_MIG_FLAG_DEVICE_BLOCK) {
|
764 | c163b5ca | lirans@il.ibm.com | /* get device name */
|
765 | c163b5ca | lirans@il.ibm.com | len = qemu_get_byte(f); |
766 | c163b5ca | lirans@il.ibm.com | qemu_get_buffer(f, (uint8_t *)device_name, len); |
767 | c163b5ca | lirans@il.ibm.com | device_name[len] = '\0';
|
768 | a55eb92c | Jan Kiszka | |
769 | c163b5ca | lirans@il.ibm.com | bs = bdrv_find(device_name); |
770 | 4b640365 | Jan Kiszka | if (!bs) {
|
771 | 4b640365 | Jan Kiszka | fprintf(stderr, "Error unknown block device %s\n",
|
772 | 4b640365 | Jan Kiszka | device_name); |
773 | 4b640365 | Jan Kiszka | return -EINVAL;
|
774 | 4b640365 | Jan Kiszka | } |
775 | a55eb92c | Jan Kiszka | |
776 | 77358b59 | Pierre Riteau | if (bs != bs_prev) {
|
777 | 77358b59 | Pierre Riteau | bs_prev = bs; |
778 | 77358b59 | Pierre Riteau | total_sectors = bdrv_getlength(bs) >> BDRV_SECTOR_BITS; |
779 | 77358b59 | Pierre Riteau | if (total_sectors <= 0) { |
780 | 6daf194d | Markus Armbruster | error_report("Error getting length of block device %s",
|
781 | 77358b59 | Pierre Riteau | device_name); |
782 | 77358b59 | Pierre Riteau | return -EINVAL;
|
783 | 77358b59 | Pierre Riteau | } |
784 | 77358b59 | Pierre Riteau | } |
785 | 77358b59 | Pierre Riteau | |
786 | 77358b59 | Pierre Riteau | if (total_sectors - addr < BDRV_SECTORS_PER_DIRTY_CHUNK) {
|
787 | 77358b59 | Pierre Riteau | nr_sectors = total_sectors - addr; |
788 | 77358b59 | Pierre Riteau | } else {
|
789 | 77358b59 | Pierre Riteau | nr_sectors = BDRV_SECTORS_PER_DIRTY_CHUNK; |
790 | 77358b59 | Pierre Riteau | } |
791 | 77358b59 | Pierre Riteau | |
792 | 323004a3 | Peter Lieven | if (flags & BLK_MIG_FLAG_ZERO_BLOCK) {
|
793 | d32f35cb | Peter Lieven | ret = bdrv_write_zeroes(bs, addr, nr_sectors, |
794 | d32f35cb | Peter Lieven | BDRV_REQ_MAY_UNMAP); |
795 | 323004a3 | Peter Lieven | } else {
|
796 | 323004a3 | Peter Lieven | buf = g_malloc(BLOCK_SIZE); |
797 | 323004a3 | Peter Lieven | qemu_get_buffer(f, buf, BLOCK_SIZE); |
798 | 323004a3 | Peter Lieven | ret = bdrv_write(bs, addr, buf, nr_sectors); |
799 | 323004a3 | Peter Lieven | g_free(buf); |
800 | 323004a3 | Peter Lieven | } |
801 | 575a58d7 | Jan Kiszka | |
802 | b02bea3a | Yoshiaki Tamura | if (ret < 0) { |
803 | b02bea3a | Yoshiaki Tamura | return ret;
|
804 | b02bea3a | Yoshiaki Tamura | } |
805 | 01e61e2d | Jan Kiszka | } else if (flags & BLK_MIG_FLAG_PROGRESS) { |
806 | 01e61e2d | Jan Kiszka | if (!banner_printed) {
|
807 | 01e61e2d | Jan Kiszka | printf("Receiving block device images\n");
|
808 | 01e61e2d | Jan Kiszka | banner_printed = 1;
|
809 | 01e61e2d | Jan Kiszka | } |
810 | 01e61e2d | Jan Kiszka | printf("Completed %d %%%c", (int)addr, |
811 | 01e61e2d | Jan Kiszka | (addr == 100) ? '\n' : '\r'); |
812 | 01e61e2d | Jan Kiszka | fflush(stdout); |
813 | a55eb92c | Jan Kiszka | } else if (!(flags & BLK_MIG_FLAG_EOS)) { |
814 | d5f1f286 | Stefan Hajnoczi | fprintf(stderr, "Unknown block migration flags: %#x\n", flags);
|
815 | 4b640365 | Jan Kiszka | return -EINVAL;
|
816 | 4b640365 | Jan Kiszka | } |
817 | 42802d47 | Juan Quintela | ret = qemu_file_get_error(f); |
818 | 42802d47 | Juan Quintela | if (ret != 0) { |
819 | 42802d47 | Juan Quintela | return ret;
|
820 | c163b5ca | lirans@il.ibm.com | } |
821 | a55eb92c | Jan Kiszka | } while (!(flags & BLK_MIG_FLAG_EOS));
|
822 | a55eb92c | Jan Kiszka | |
823 | c163b5ca | lirans@il.ibm.com | return 0; |
824 | c163b5ca | lirans@il.ibm.com | } |
825 | c163b5ca | lirans@il.ibm.com | |
826 | 6607ae23 | Isaku Yamahata | static void block_set_params(const MigrationParams *params, void *opaque) |
827 | c163b5ca | lirans@il.ibm.com | { |
828 | 6607ae23 | Isaku Yamahata | block_mig_state.blk_enable = params->blk; |
829 | 6607ae23 | Isaku Yamahata | block_mig_state.shared_base = params->shared; |
830 | a55eb92c | Jan Kiszka | |
831 | c163b5ca | lirans@il.ibm.com | /* shared base means that blk_enable = 1 */
|
832 | 6607ae23 | Isaku Yamahata | block_mig_state.blk_enable |= params->shared; |
833 | c163b5ca | lirans@il.ibm.com | } |
834 | c163b5ca | lirans@il.ibm.com | |
835 | 6bd68781 | Juan Quintela | static bool block_is_active(void *opaque) |
836 | 6bd68781 | Juan Quintela | { |
837 | 6bd68781 | Juan Quintela | return block_mig_state.blk_enable == 1; |
838 | 6bd68781 | Juan Quintela | } |
839 | 6bd68781 | Juan Quintela | |
840 | 7908c78d | Juan Quintela | SaveVMHandlers savevm_block_handlers = { |
841 | 7908c78d | Juan Quintela | .set_params = block_set_params, |
842 | d1315aac | Juan Quintela | .save_live_setup = block_save_setup, |
843 | 16310a3c | Juan Quintela | .save_live_iterate = block_save_iterate, |
844 | 16310a3c | Juan Quintela | .save_live_complete = block_save_complete, |
845 | e4ed1541 | Juan Quintela | .save_live_pending = block_save_pending, |
846 | 7908c78d | Juan Quintela | .load_state = block_load, |
847 | 9b5bfab0 | Juan Quintela | .cancel = block_migration_cancel, |
848 | 6bd68781 | Juan Quintela | .is_active = block_is_active, |
849 | 7908c78d | Juan Quintela | }; |
850 | 7908c78d | Juan Quintela | |
851 | c163b5ca | lirans@il.ibm.com | void blk_mig_init(void) |
852 | a55eb92c | Jan Kiszka | { |
853 | 5e5328be | Jan Kiszka | QSIMPLEQ_INIT(&block_mig_state.bmds_list); |
854 | 5e5328be | Jan Kiszka | QSIMPLEQ_INIT(&block_mig_state.blk_list); |
855 | 52e850de | Paolo Bonzini | qemu_mutex_init(&block_mig_state.lock); |
856 | 5e5328be | Jan Kiszka | |
857 | 7908c78d | Juan Quintela | register_savevm_live(NULL, "block", 0, 1, &savevm_block_handlers, |
858 | 7908c78d | Juan Quintela | &block_mig_state); |
859 | c163b5ca | lirans@il.ibm.com | } |