root / include / block / block.h @ bf736fe3
History | View | Annotate | Download (17.4 kB)
1 |
#ifndef BLOCK_H
|
---|---|
2 |
#define BLOCK_H
|
3 |
|
4 |
#include "block/aio.h" |
5 |
#include "qemu-common.h" |
6 |
#include "qemu/option.h" |
7 |
#include "block/coroutine.h" |
8 |
#include "qapi/qmp/qobject.h" |
9 |
#include "qapi-types.h" |
10 |
|
11 |
/* block.c */
|
12 |
typedef struct BlockDriver BlockDriver; |
13 |
typedef struct BlockJob BlockJob; |
14 |
|
15 |
typedef struct BlockDriverInfo { |
16 |
/* in bytes, 0 if irrelevant */
|
17 |
int cluster_size;
|
18 |
/* offset at which the VM state can be saved (0 if not possible) */
|
19 |
int64_t vm_state_offset; |
20 |
bool is_dirty;
|
21 |
} BlockDriverInfo; |
22 |
|
23 |
typedef struct BlockFragInfo { |
24 |
uint64_t allocated_clusters; |
25 |
uint64_t total_clusters; |
26 |
uint64_t fragmented_clusters; |
27 |
uint64_t compressed_clusters; |
28 |
} BlockFragInfo; |
29 |
|
30 |
/* Callbacks for block device models */
|
31 |
typedef struct BlockDevOps { |
32 |
/*
|
33 |
* Runs when virtual media changed (monitor commands eject, change)
|
34 |
* Argument load is true on load and false on eject.
|
35 |
* Beware: doesn't run when a host device's physical media
|
36 |
* changes. Sure would be useful if it did.
|
37 |
* Device models with removable media must implement this callback.
|
38 |
*/
|
39 |
void (*change_media_cb)(void *opaque, bool load); |
40 |
/*
|
41 |
* Runs when an eject request is issued from the monitor, the tray
|
42 |
* is closed, and the medium is locked.
|
43 |
* Device models that do not implement is_medium_locked will not need
|
44 |
* this callback. Device models that can lock the medium or tray might
|
45 |
* want to implement the callback and unlock the tray when "force" is
|
46 |
* true, even if they do not support eject requests.
|
47 |
*/
|
48 |
void (*eject_request_cb)(void *opaque, bool force); |
49 |
/*
|
50 |
* Is the virtual tray open?
|
51 |
* Device models implement this only when the device has a tray.
|
52 |
*/
|
53 |
bool (*is_tray_open)(void *opaque); |
54 |
/*
|
55 |
* Is the virtual medium locked into the device?
|
56 |
* Device models implement this only when device has such a lock.
|
57 |
*/
|
58 |
bool (*is_medium_locked)(void *opaque); |
59 |
/*
|
60 |
* Runs when the size changed (e.g. monitor command block_resize)
|
61 |
*/
|
62 |
void (*resize_cb)(void *opaque); |
63 |
} BlockDevOps; |
64 |
|
65 |
#define BDRV_O_RDWR 0x0002 |
66 |
#define BDRV_O_SNAPSHOT 0x0008 /* open the file read only and save writes in a snapshot */ |
67 |
#define BDRV_O_NOCACHE 0x0020 /* do not use the host page cache */ |
68 |
#define BDRV_O_CACHE_WB 0x0040 /* use write-back caching */ |
69 |
#define BDRV_O_NATIVE_AIO 0x0080 /* use native AIO instead of the thread pool */ |
70 |
#define BDRV_O_NO_BACKING 0x0100 /* don't open the backing file */ |
71 |
#define BDRV_O_NO_FLUSH 0x0200 /* disable flushing on this disk */ |
72 |
#define BDRV_O_COPY_ON_READ 0x0400 /* copy read backing sectors into image */ |
73 |
#define BDRV_O_INCOMING 0x0800 /* consistency hint for incoming migration */ |
74 |
#define BDRV_O_CHECK 0x1000 /* open solely for consistency check */ |
75 |
#define BDRV_O_ALLOW_RDWR 0x2000 /* allow reopen to change from r/o to r/w */ |
76 |
#define BDRV_O_UNMAP 0x4000 /* execute guest UNMAP/TRIM operations */ |
77 |
|
78 |
#define BDRV_O_CACHE_MASK (BDRV_O_NOCACHE | BDRV_O_CACHE_WB | BDRV_O_NO_FLUSH)
|
79 |
|
80 |
#define BDRV_SECTOR_BITS 9 |
81 |
#define BDRV_SECTOR_SIZE (1ULL << BDRV_SECTOR_BITS) |
82 |
#define BDRV_SECTOR_MASK ~(BDRV_SECTOR_SIZE - 1) |
83 |
|
84 |
typedef enum { |
85 |
BDRV_ACTION_REPORT, BDRV_ACTION_IGNORE, BDRV_ACTION_STOP |
86 |
} BlockErrorAction; |
87 |
|
88 |
typedef QSIMPLEQ_HEAD(BlockReopenQueue, BlockReopenQueueEntry) BlockReopenQueue;
|
89 |
|
90 |
typedef struct BDRVReopenState { |
91 |
BlockDriverState *bs; |
92 |
int flags;
|
93 |
void *opaque;
|
94 |
} BDRVReopenState; |
95 |
|
96 |
|
97 |
void bdrv_iostatus_enable(BlockDriverState *bs);
|
98 |
void bdrv_iostatus_reset(BlockDriverState *bs);
|
99 |
void bdrv_iostatus_disable(BlockDriverState *bs);
|
100 |
bool bdrv_iostatus_is_enabled(const BlockDriverState *bs); |
101 |
void bdrv_iostatus_set_err(BlockDriverState *bs, int error); |
102 |
void bdrv_info_print(Monitor *mon, const QObject *data); |
103 |
void bdrv_info(Monitor *mon, QObject **ret_data);
|
104 |
void bdrv_stats_print(Monitor *mon, const QObject *data); |
105 |
void bdrv_info_stats(Monitor *mon, QObject **ret_data);
|
106 |
|
107 |
/* disk I/O throttling */
|
108 |
void bdrv_io_limits_enable(BlockDriverState *bs);
|
109 |
void bdrv_io_limits_disable(BlockDriverState *bs);
|
110 |
bool bdrv_io_limits_enabled(BlockDriverState *bs);
|
111 |
|
112 |
void bdrv_init(void); |
113 |
void bdrv_init_with_whitelist(void); |
114 |
BlockDriver *bdrv_find_protocol(const char *filename); |
115 |
BlockDriver *bdrv_find_format(const char *format_name); |
116 |
BlockDriver *bdrv_find_whitelisted_format(const char *format_name, |
117 |
bool readonly);
|
118 |
int bdrv_create(BlockDriver *drv, const char* filename, |
119 |
QEMUOptionParameter *options); |
120 |
int bdrv_create_file(const char* filename, QEMUOptionParameter *options); |
121 |
BlockDriverState *bdrv_new(const char *device_name); |
122 |
void bdrv_make_anon(BlockDriverState *bs);
|
123 |
void bdrv_swap(BlockDriverState *bs_new, BlockDriverState *bs_old);
|
124 |
void bdrv_append(BlockDriverState *bs_new, BlockDriverState *bs_top);
|
125 |
void bdrv_delete(BlockDriverState *bs);
|
126 |
int bdrv_parse_cache_flags(const char *mode, int *flags); |
127 |
int bdrv_parse_discard_flags(const char *mode, int *flags); |
128 |
int bdrv_file_open(BlockDriverState **pbs, const char *filename, |
129 |
QDict *options, int flags);
|
130 |
int bdrv_open_backing_file(BlockDriverState *bs, QDict *options);
|
131 |
int bdrv_open(BlockDriverState *bs, const char *filename, QDict *options, |
132 |
int flags, BlockDriver *drv);
|
133 |
BlockReopenQueue *bdrv_reopen_queue(BlockReopenQueue *bs_queue, |
134 |
BlockDriverState *bs, int flags);
|
135 |
int bdrv_reopen_multiple(BlockReopenQueue *bs_queue, Error **errp);
|
136 |
int bdrv_reopen(BlockDriverState *bs, int bdrv_flags, Error **errp); |
137 |
int bdrv_reopen_prepare(BDRVReopenState *reopen_state,
|
138 |
BlockReopenQueue *queue, Error **errp); |
139 |
void bdrv_reopen_commit(BDRVReopenState *reopen_state);
|
140 |
void bdrv_reopen_abort(BDRVReopenState *reopen_state);
|
141 |
void bdrv_close(BlockDriverState *bs);
|
142 |
void bdrv_add_close_notifier(BlockDriverState *bs, Notifier *notify);
|
143 |
int bdrv_attach_dev(BlockDriverState *bs, void *dev); |
144 |
void bdrv_attach_dev_nofail(BlockDriverState *bs, void *dev); |
145 |
void bdrv_detach_dev(BlockDriverState *bs, void *dev); |
146 |
void *bdrv_get_attached_dev(BlockDriverState *bs);
|
147 |
void bdrv_set_dev_ops(BlockDriverState *bs, const BlockDevOps *ops, |
148 |
void *opaque);
|
149 |
void bdrv_dev_eject_request(BlockDriverState *bs, bool force); |
150 |
bool bdrv_dev_has_removable_media(BlockDriverState *bs);
|
151 |
bool bdrv_dev_is_tray_open(BlockDriverState *bs);
|
152 |
bool bdrv_dev_is_medium_locked(BlockDriverState *bs);
|
153 |
int bdrv_read(BlockDriverState *bs, int64_t sector_num,
|
154 |
uint8_t *buf, int nb_sectors);
|
155 |
int bdrv_read_unthrottled(BlockDriverState *bs, int64_t sector_num,
|
156 |
uint8_t *buf, int nb_sectors);
|
157 |
int bdrv_write(BlockDriverState *bs, int64_t sector_num,
|
158 |
const uint8_t *buf, int nb_sectors); |
159 |
int bdrv_writev(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov);
|
160 |
int bdrv_pread(BlockDriverState *bs, int64_t offset,
|
161 |
void *buf, int count); |
162 |
int bdrv_pwrite(BlockDriverState *bs, int64_t offset,
|
163 |
const void *buf, int count); |
164 |
int bdrv_pwritev(BlockDriverState *bs, int64_t offset, QEMUIOVector *qiov);
|
165 |
int bdrv_pwrite_sync(BlockDriverState *bs, int64_t offset,
|
166 |
const void *buf, int count); |
167 |
int coroutine_fn bdrv_co_readv(BlockDriverState *bs, int64_t sector_num,
|
168 |
int nb_sectors, QEMUIOVector *qiov);
|
169 |
int coroutine_fn bdrv_co_copy_on_readv(BlockDriverState *bs,
|
170 |
int64_t sector_num, int nb_sectors, QEMUIOVector *qiov);
|
171 |
int coroutine_fn bdrv_co_writev(BlockDriverState *bs, int64_t sector_num,
|
172 |
int nb_sectors, QEMUIOVector *qiov);
|
173 |
/*
|
174 |
* Efficiently zero a region of the disk image. Note that this is a regular
|
175 |
* I/O request like read or write and should have a reasonable size. This
|
176 |
* function is not suitable for zeroing the entire image in a single request
|
177 |
* because it may allocate memory for the entire region.
|
178 |
*/
|
179 |
int coroutine_fn bdrv_co_write_zeroes(BlockDriverState *bs, int64_t sector_num,
|
180 |
int nb_sectors);
|
181 |
int coroutine_fn bdrv_co_is_allocated(BlockDriverState *bs, int64_t sector_num,
|
182 |
int nb_sectors, int *pnum); |
183 |
int coroutine_fn bdrv_co_is_allocated_above(BlockDriverState *top,
|
184 |
BlockDriverState *base, |
185 |
int64_t sector_num, |
186 |
int nb_sectors, int *pnum); |
187 |
BlockDriverState *bdrv_find_backing_image(BlockDriverState *bs, |
188 |
const char *backing_file); |
189 |
int bdrv_get_backing_file_depth(BlockDriverState *bs);
|
190 |
int bdrv_truncate(BlockDriverState *bs, int64_t offset);
|
191 |
int64_t bdrv_getlength(BlockDriverState *bs); |
192 |
int64_t bdrv_get_allocated_file_size(BlockDriverState *bs); |
193 |
void bdrv_get_geometry(BlockDriverState *bs, uint64_t *nb_sectors_ptr);
|
194 |
int bdrv_commit(BlockDriverState *bs);
|
195 |
int bdrv_commit_all(void); |
196 |
int bdrv_change_backing_file(BlockDriverState *bs,
|
197 |
const char *backing_file, const char *backing_fmt); |
198 |
void bdrv_register(BlockDriver *bdrv);
|
199 |
int bdrv_drop_intermediate(BlockDriverState *active, BlockDriverState *top,
|
200 |
BlockDriverState *base); |
201 |
BlockDriverState *bdrv_find_overlay(BlockDriverState *active, |
202 |
BlockDriverState *bs); |
203 |
BlockDriverState *bdrv_find_base(BlockDriverState *bs); |
204 |
|
205 |
|
206 |
typedef struct BdrvCheckResult { |
207 |
int corruptions;
|
208 |
int leaks;
|
209 |
int check_errors;
|
210 |
int corruptions_fixed;
|
211 |
int leaks_fixed;
|
212 |
int64_t image_end_offset; |
213 |
BlockFragInfo bfi; |
214 |
} BdrvCheckResult; |
215 |
|
216 |
typedef enum { |
217 |
BDRV_FIX_LEAKS = 1,
|
218 |
BDRV_FIX_ERRORS = 2,
|
219 |
} BdrvCheckMode; |
220 |
|
221 |
int bdrv_check(BlockDriverState *bs, BdrvCheckResult *res, BdrvCheckMode fix);
|
222 |
|
223 |
/* async block I/O */
|
224 |
typedef void BlockDriverDirtyHandler(BlockDriverState *bs, int64_t sector, |
225 |
int sector_num);
|
226 |
BlockDriverAIOCB *bdrv_aio_readv(BlockDriverState *bs, int64_t sector_num, |
227 |
QEMUIOVector *iov, int nb_sectors,
|
228 |
BlockDriverCompletionFunc *cb, void *opaque);
|
229 |
BlockDriverAIOCB *bdrv_aio_writev(BlockDriverState *bs, int64_t sector_num, |
230 |
QEMUIOVector *iov, int nb_sectors,
|
231 |
BlockDriverCompletionFunc *cb, void *opaque);
|
232 |
BlockDriverAIOCB *bdrv_aio_flush(BlockDriverState *bs, |
233 |
BlockDriverCompletionFunc *cb, void *opaque);
|
234 |
BlockDriverAIOCB *bdrv_aio_discard(BlockDriverState *bs, |
235 |
int64_t sector_num, int nb_sectors,
|
236 |
BlockDriverCompletionFunc *cb, void *opaque);
|
237 |
void bdrv_aio_cancel(BlockDriverAIOCB *acb);
|
238 |
|
239 |
typedef struct BlockRequest { |
240 |
/* Fields to be filled by multiwrite caller */
|
241 |
int64_t sector; |
242 |
int nb_sectors;
|
243 |
QEMUIOVector *qiov; |
244 |
BlockDriverCompletionFunc *cb; |
245 |
void *opaque;
|
246 |
|
247 |
/* Filled by multiwrite implementation */
|
248 |
int error;
|
249 |
} BlockRequest; |
250 |
|
251 |
int bdrv_aio_multiwrite(BlockDriverState *bs, BlockRequest *reqs,
|
252 |
int num_reqs);
|
253 |
|
254 |
/* sg packet commands */
|
255 |
int bdrv_ioctl(BlockDriverState *bs, unsigned long int req, void *buf); |
256 |
BlockDriverAIOCB *bdrv_aio_ioctl(BlockDriverState *bs, |
257 |
unsigned long int req, void *buf, |
258 |
BlockDriverCompletionFunc *cb, void *opaque);
|
259 |
|
260 |
/* Invalidate any cached metadata used by image formats */
|
261 |
void bdrv_invalidate_cache(BlockDriverState *bs);
|
262 |
void bdrv_invalidate_cache_all(void); |
263 |
|
264 |
void bdrv_clear_incoming_migration_all(void); |
265 |
|
266 |
/* Ensure contents are flushed to disk. */
|
267 |
int bdrv_flush(BlockDriverState *bs);
|
268 |
int coroutine_fn bdrv_co_flush(BlockDriverState *bs);
|
269 |
void bdrv_flush_all(void); |
270 |
void bdrv_close_all(void); |
271 |
void bdrv_drain_all(void); |
272 |
|
273 |
int bdrv_discard(BlockDriverState *bs, int64_t sector_num, int nb_sectors); |
274 |
int bdrv_co_discard(BlockDriverState *bs, int64_t sector_num, int nb_sectors); |
275 |
int bdrv_has_zero_init(BlockDriverState *bs);
|
276 |
int bdrv_is_allocated(BlockDriverState *bs, int64_t sector_num, int nb_sectors, |
277 |
int *pnum);
|
278 |
int bdrv_is_allocated_above(BlockDriverState *top, BlockDriverState *base,
|
279 |
int64_t sector_num, int nb_sectors, int *pnum); |
280 |
|
281 |
void bdrv_set_on_error(BlockDriverState *bs, BlockdevOnError on_read_error,
|
282 |
BlockdevOnError on_write_error); |
283 |
BlockdevOnError bdrv_get_on_error(BlockDriverState *bs, bool is_read);
|
284 |
BlockErrorAction bdrv_get_error_action(BlockDriverState *bs, bool is_read, int error); |
285 |
void bdrv_error_action(BlockDriverState *bs, BlockErrorAction action,
|
286 |
bool is_read, int error); |
287 |
int bdrv_is_read_only(BlockDriverState *bs);
|
288 |
int bdrv_is_sg(BlockDriverState *bs);
|
289 |
int bdrv_enable_write_cache(BlockDriverState *bs);
|
290 |
void bdrv_set_enable_write_cache(BlockDriverState *bs, bool wce); |
291 |
int bdrv_is_inserted(BlockDriverState *bs);
|
292 |
int bdrv_media_changed(BlockDriverState *bs);
|
293 |
void bdrv_lock_medium(BlockDriverState *bs, bool locked); |
294 |
void bdrv_eject(BlockDriverState *bs, bool eject_flag); |
295 |
const char *bdrv_get_format_name(BlockDriverState *bs); |
296 |
BlockDriverState *bdrv_find(const char *name); |
297 |
BlockDriverState *bdrv_next(BlockDriverState *bs); |
298 |
void bdrv_iterate(void (*it)(void *opaque, BlockDriverState *bs), |
299 |
void *opaque);
|
300 |
int bdrv_is_encrypted(BlockDriverState *bs);
|
301 |
int bdrv_key_required(BlockDriverState *bs);
|
302 |
int bdrv_set_key(BlockDriverState *bs, const char *key); |
303 |
int bdrv_query_missing_keys(void); |
304 |
void bdrv_iterate_format(void (*it)(void *opaque, const char *name), |
305 |
void *opaque);
|
306 |
const char *bdrv_get_device_name(BlockDriverState *bs); |
307 |
int bdrv_get_flags(BlockDriverState *bs);
|
308 |
int bdrv_write_compressed(BlockDriverState *bs, int64_t sector_num,
|
309 |
const uint8_t *buf, int nb_sectors); |
310 |
int bdrv_get_info(BlockDriverState *bs, BlockDriverInfo *bdi);
|
311 |
void bdrv_round_to_clusters(BlockDriverState *bs,
|
312 |
int64_t sector_num, int nb_sectors,
|
313 |
int64_t *cluster_sector_num, |
314 |
int *cluster_nb_sectors);
|
315 |
|
316 |
const char *bdrv_get_encrypted_filename(BlockDriverState *bs); |
317 |
void bdrv_get_backing_filename(BlockDriverState *bs,
|
318 |
char *filename, int filename_size); |
319 |
void bdrv_get_full_backing_filename(BlockDriverState *bs,
|
320 |
char *dest, size_t sz);
|
321 |
int bdrv_is_snapshot(BlockDriverState *bs);
|
322 |
|
323 |
int path_is_absolute(const char *path); |
324 |
void path_combine(char *dest, int dest_size, |
325 |
const char *base_path, |
326 |
const char *filename); |
327 |
|
328 |
int bdrv_writev_vmstate(BlockDriverState *bs, QEMUIOVector *qiov, int64_t pos);
|
329 |
int bdrv_save_vmstate(BlockDriverState *bs, const uint8_t *buf, |
330 |
int64_t pos, int size);
|
331 |
|
332 |
int bdrv_load_vmstate(BlockDriverState *bs, uint8_t *buf,
|
333 |
int64_t pos, int size);
|
334 |
|
335 |
void bdrv_img_create(const char *filename, const char *fmt, |
336 |
const char *base_filename, const char *base_fmt, |
337 |
char *options, uint64_t img_size, int flags, |
338 |
Error **errp, bool quiet);
|
339 |
|
340 |
void bdrv_set_buffer_alignment(BlockDriverState *bs, int align); |
341 |
void *qemu_blockalign(BlockDriverState *bs, size_t size);
|
342 |
bool bdrv_qiov_is_aligned(BlockDriverState *bs, QEMUIOVector *qiov);
|
343 |
|
344 |
struct HBitmapIter;
|
345 |
void bdrv_set_dirty_tracking(BlockDriverState *bs, int granularity); |
346 |
int bdrv_get_dirty(BlockDriverState *bs, int64_t sector);
|
347 |
void bdrv_set_dirty(BlockDriverState *bs, int64_t cur_sector, int nr_sectors); |
348 |
void bdrv_reset_dirty(BlockDriverState *bs, int64_t cur_sector, int nr_sectors); |
349 |
void bdrv_dirty_iter_init(BlockDriverState *bs, struct HBitmapIter *hbi); |
350 |
int64_t bdrv_get_dirty_count(BlockDriverState *bs); |
351 |
|
352 |
void bdrv_enable_copy_on_read(BlockDriverState *bs);
|
353 |
void bdrv_disable_copy_on_read(BlockDriverState *bs);
|
354 |
|
355 |
void bdrv_set_in_use(BlockDriverState *bs, int in_use); |
356 |
int bdrv_in_use(BlockDriverState *bs);
|
357 |
|
358 |
#ifdef CONFIG_LINUX_AIO
|
359 |
int raw_get_aio_fd(BlockDriverState *bs);
|
360 |
#else
|
361 |
static inline int raw_get_aio_fd(BlockDriverState *bs) |
362 |
{ |
363 |
return -ENOTSUP;
|
364 |
} |
365 |
#endif
|
366 |
|
367 |
enum BlockAcctType {
|
368 |
BDRV_ACCT_READ, |
369 |
BDRV_ACCT_WRITE, |
370 |
BDRV_ACCT_FLUSH, |
371 |
BDRV_MAX_IOTYPE, |
372 |
}; |
373 |
|
374 |
typedef struct BlockAcctCookie { |
375 |
int64_t bytes; |
376 |
int64_t start_time_ns; |
377 |
enum BlockAcctType type;
|
378 |
} BlockAcctCookie; |
379 |
|
380 |
void bdrv_acct_start(BlockDriverState *bs, BlockAcctCookie *cookie,
|
381 |
int64_t bytes, enum BlockAcctType type);
|
382 |
void bdrv_acct_done(BlockDriverState *bs, BlockAcctCookie *cookie);
|
383 |
|
384 |
typedef enum { |
385 |
BLKDBG_L1_UPDATE, |
386 |
|
387 |
BLKDBG_L1_GROW_ALLOC_TABLE, |
388 |
BLKDBG_L1_GROW_WRITE_TABLE, |
389 |
BLKDBG_L1_GROW_ACTIVATE_TABLE, |
390 |
|
391 |
BLKDBG_L2_LOAD, |
392 |
BLKDBG_L2_UPDATE, |
393 |
BLKDBG_L2_UPDATE_COMPRESSED, |
394 |
BLKDBG_L2_ALLOC_COW_READ, |
395 |
BLKDBG_L2_ALLOC_WRITE, |
396 |
|
397 |
BLKDBG_READ_AIO, |
398 |
BLKDBG_READ_BACKING_AIO, |
399 |
BLKDBG_READ_COMPRESSED, |
400 |
|
401 |
BLKDBG_WRITE_AIO, |
402 |
BLKDBG_WRITE_COMPRESSED, |
403 |
|
404 |
BLKDBG_VMSTATE_LOAD, |
405 |
BLKDBG_VMSTATE_SAVE, |
406 |
|
407 |
BLKDBG_COW_READ, |
408 |
BLKDBG_COW_WRITE, |
409 |
|
410 |
BLKDBG_REFTABLE_LOAD, |
411 |
BLKDBG_REFTABLE_GROW, |
412 |
|
413 |
BLKDBG_REFBLOCK_LOAD, |
414 |
BLKDBG_REFBLOCK_UPDATE, |
415 |
BLKDBG_REFBLOCK_UPDATE_PART, |
416 |
BLKDBG_REFBLOCK_ALLOC, |
417 |
BLKDBG_REFBLOCK_ALLOC_HOOKUP, |
418 |
BLKDBG_REFBLOCK_ALLOC_WRITE, |
419 |
BLKDBG_REFBLOCK_ALLOC_WRITE_BLOCKS, |
420 |
BLKDBG_REFBLOCK_ALLOC_WRITE_TABLE, |
421 |
BLKDBG_REFBLOCK_ALLOC_SWITCH_TABLE, |
422 |
|
423 |
BLKDBG_CLUSTER_ALLOC, |
424 |
BLKDBG_CLUSTER_ALLOC_BYTES, |
425 |
BLKDBG_CLUSTER_FREE, |
426 |
|
427 |
BLKDBG_FLUSH_TO_OS, |
428 |
BLKDBG_FLUSH_TO_DISK, |
429 |
|
430 |
BLKDBG_EVENT_MAX, |
431 |
} BlkDebugEvent; |
432 |
|
433 |
#define BLKDBG_EVENT(bs, evt) bdrv_debug_event(bs, evt)
|
434 |
void bdrv_debug_event(BlockDriverState *bs, BlkDebugEvent event);
|
435 |
|
436 |
int bdrv_debug_breakpoint(BlockDriverState *bs, const char *event, |
437 |
const char *tag); |
438 |
int bdrv_debug_resume(BlockDriverState *bs, const char *tag); |
439 |
bool bdrv_debug_is_suspended(BlockDriverState *bs, const char *tag); |
440 |
|
441 |
#endif
|