root / include / block / block.h @ ddf5636d
History | View | Annotate | Download (22.5 kB)
1 |
#ifndef BLOCK_H
|
---|---|
2 |
#define BLOCK_H
|
3 |
|
4 |
#include "block/aio.h" |
5 |
#include "qemu-common.h" |
6 |
#include "qemu/option.h" |
7 |
#include "block/coroutine.h" |
8 |
#include "qapi/qmp/qobject.h" |
9 |
#include "qapi-types.h" |
10 |
|
11 |
/* block.c */
|
12 |
typedef struct BlockDriver BlockDriver; |
13 |
typedef struct BlockJob BlockJob; |
14 |
|
15 |
typedef struct BlockDriverInfo { |
16 |
/* in bytes, 0 if irrelevant */
|
17 |
int cluster_size;
|
18 |
/* offset at which the VM state can be saved (0 if not possible) */
|
19 |
int64_t vm_state_offset; |
20 |
bool is_dirty;
|
21 |
/*
|
22 |
* True if unallocated blocks read back as zeroes. This is equivalent
|
23 |
* to the the LBPRZ flag in the SCSI logical block provisioning page.
|
24 |
*/
|
25 |
bool unallocated_blocks_are_zero;
|
26 |
/*
|
27 |
* True if the driver can optimize writing zeroes by unmapping
|
28 |
* sectors. This is equivalent to the BLKDISCARDZEROES ioctl in Linux
|
29 |
* with the difference that in qemu a discard is allowed to silently
|
30 |
* fail. Therefore we have to use bdrv_write_zeroes with the
|
31 |
* BDRV_REQ_MAY_UNMAP flag for an optimized zero write with unmapping.
|
32 |
* After this call the driver has to guarantee that the contents read
|
33 |
* back as zero. It is additionally required that the block device is
|
34 |
* opened with BDRV_O_UNMAP flag for this to work.
|
35 |
*/
|
36 |
bool can_write_zeroes_with_unmap;
|
37 |
} BlockDriverInfo; |
38 |
|
39 |
typedef struct BlockFragInfo { |
40 |
uint64_t allocated_clusters; |
41 |
uint64_t total_clusters; |
42 |
uint64_t fragmented_clusters; |
43 |
uint64_t compressed_clusters; |
44 |
} BlockFragInfo; |
45 |
|
46 |
/* Callbacks for block device models */
|
47 |
typedef struct BlockDevOps { |
48 |
/*
|
49 |
* Runs when virtual media changed (monitor commands eject, change)
|
50 |
* Argument load is true on load and false on eject.
|
51 |
* Beware: doesn't run when a host device's physical media
|
52 |
* changes. Sure would be useful if it did.
|
53 |
* Device models with removable media must implement this callback.
|
54 |
*/
|
55 |
void (*change_media_cb)(void *opaque, bool load); |
56 |
/*
|
57 |
* Runs when an eject request is issued from the monitor, the tray
|
58 |
* is closed, and the medium is locked.
|
59 |
* Device models that do not implement is_medium_locked will not need
|
60 |
* this callback. Device models that can lock the medium or tray might
|
61 |
* want to implement the callback and unlock the tray when "force" is
|
62 |
* true, even if they do not support eject requests.
|
63 |
*/
|
64 |
void (*eject_request_cb)(void *opaque, bool force); |
65 |
/*
|
66 |
* Is the virtual tray open?
|
67 |
* Device models implement this only when the device has a tray.
|
68 |
*/
|
69 |
bool (*is_tray_open)(void *opaque); |
70 |
/*
|
71 |
* Is the virtual medium locked into the device?
|
72 |
* Device models implement this only when device has such a lock.
|
73 |
*/
|
74 |
bool (*is_medium_locked)(void *opaque); |
75 |
/*
|
76 |
* Runs when the size changed (e.g. monitor command block_resize)
|
77 |
*/
|
78 |
void (*resize_cb)(void *opaque); |
79 |
} BlockDevOps; |
80 |
|
81 |
typedef enum { |
82 |
BDRV_REQ_COPY_ON_READ = 0x1,
|
83 |
BDRV_REQ_ZERO_WRITE = 0x2,
|
84 |
/* The BDRV_REQ_MAY_UNMAP flag is used to indicate that the block driver
|
85 |
* is allowed to optimize a write zeroes request by unmapping (discarding)
|
86 |
* blocks if it is guaranteed that the result will read back as
|
87 |
* zeroes. The flag is only passed to the driver if the block device is
|
88 |
* opened with BDRV_O_UNMAP.
|
89 |
*/
|
90 |
BDRV_REQ_MAY_UNMAP = 0x4,
|
91 |
} BdrvRequestFlags; |
92 |
|
93 |
#define BDRV_O_RDWR 0x0002 |
94 |
#define BDRV_O_SNAPSHOT 0x0008 /* open the file read only and save writes in a snapshot */ |
95 |
#define BDRV_O_NOCACHE 0x0020 /* do not use the host page cache */ |
96 |
#define BDRV_O_CACHE_WB 0x0040 /* use write-back caching */ |
97 |
#define BDRV_O_NATIVE_AIO 0x0080 /* use native AIO instead of the thread pool */ |
98 |
#define BDRV_O_NO_BACKING 0x0100 /* don't open the backing file */ |
99 |
#define BDRV_O_NO_FLUSH 0x0200 /* disable flushing on this disk */ |
100 |
#define BDRV_O_COPY_ON_READ 0x0400 /* copy read backing sectors into image */ |
101 |
#define BDRV_O_INCOMING 0x0800 /* consistency hint for incoming migration */ |
102 |
#define BDRV_O_CHECK 0x1000 /* open solely for consistency check */ |
103 |
#define BDRV_O_ALLOW_RDWR 0x2000 /* allow reopen to change from r/o to r/w */ |
104 |
#define BDRV_O_UNMAP 0x4000 /* execute guest UNMAP/TRIM operations */ |
105 |
|
106 |
#define BDRV_O_CACHE_MASK (BDRV_O_NOCACHE | BDRV_O_CACHE_WB | BDRV_O_NO_FLUSH)
|
107 |
|
108 |
#define BDRV_SECTOR_BITS 9 |
109 |
#define BDRV_SECTOR_SIZE (1ULL << BDRV_SECTOR_BITS) |
110 |
#define BDRV_SECTOR_MASK ~(BDRV_SECTOR_SIZE - 1) |
111 |
|
112 |
/* BDRV_BLOCK_DATA: data is read from bs->file or another file
|
113 |
* BDRV_BLOCK_ZERO: sectors read as zero
|
114 |
* BDRV_BLOCK_OFFSET_VALID: sector stored in bs->file as raw data
|
115 |
* BDRV_BLOCK_RAW: used internally to indicate that the request
|
116 |
* was answered by the raw driver and that one
|
117 |
* should look in bs->file directly.
|
118 |
*
|
119 |
* If BDRV_BLOCK_OFFSET_VALID is set, bits 9-62 represent the offset in
|
120 |
* bs->file where sector data can be read from as raw data.
|
121 |
*
|
122 |
* DATA == 0 && ZERO == 0 means that data is read from backing_hd if present.
|
123 |
*
|
124 |
* DATA ZERO OFFSET_VALID
|
125 |
* t t t sectors read as zero, bs->file is zero at offset
|
126 |
* t f t sectors read as valid from bs->file at offset
|
127 |
* f t t sectors preallocated, read as zero, bs->file not
|
128 |
* necessarily zero at offset
|
129 |
* f f t sectors preallocated but read from backing_hd,
|
130 |
* bs->file contains garbage at offset
|
131 |
* t t f sectors preallocated, read as zero, unknown offset
|
132 |
* t f f sectors read from unknown file or offset
|
133 |
* f t f not allocated or unknown offset, read as zero
|
134 |
* f f f not allocated or unknown offset, read from backing_hd
|
135 |
*/
|
136 |
#define BDRV_BLOCK_DATA 1 |
137 |
#define BDRV_BLOCK_ZERO 2 |
138 |
#define BDRV_BLOCK_OFFSET_VALID 4 |
139 |
#define BDRV_BLOCK_RAW 8 |
140 |
#define BDRV_BLOCK_OFFSET_MASK BDRV_SECTOR_MASK
|
141 |
|
142 |
typedef enum { |
143 |
BDRV_ACTION_REPORT, BDRV_ACTION_IGNORE, BDRV_ACTION_STOP |
144 |
} BlockErrorAction; |
145 |
|
146 |
typedef QSIMPLEQ_HEAD(BlockReopenQueue, BlockReopenQueueEntry) BlockReopenQueue;
|
147 |
|
148 |
typedef struct BDRVReopenState { |
149 |
BlockDriverState *bs; |
150 |
int flags;
|
151 |
void *opaque;
|
152 |
} BDRVReopenState; |
153 |
|
154 |
|
155 |
void bdrv_iostatus_enable(BlockDriverState *bs);
|
156 |
void bdrv_iostatus_reset(BlockDriverState *bs);
|
157 |
void bdrv_iostatus_disable(BlockDriverState *bs);
|
158 |
bool bdrv_iostatus_is_enabled(const BlockDriverState *bs); |
159 |
void bdrv_iostatus_set_err(BlockDriverState *bs, int error); |
160 |
void bdrv_info_print(Monitor *mon, const QObject *data); |
161 |
void bdrv_info(Monitor *mon, QObject **ret_data);
|
162 |
void bdrv_stats_print(Monitor *mon, const QObject *data); |
163 |
void bdrv_info_stats(Monitor *mon, QObject **ret_data);
|
164 |
|
165 |
/* disk I/O throttling */
|
166 |
void bdrv_io_limits_enable(BlockDriverState *bs);
|
167 |
void bdrv_io_limits_disable(BlockDriverState *bs);
|
168 |
|
169 |
void bdrv_init(void); |
170 |
void bdrv_init_with_whitelist(void); |
171 |
BlockDriver *bdrv_find_protocol(const char *filename, |
172 |
bool allow_protocol_prefix);
|
173 |
BlockDriver *bdrv_find_format(const char *format_name); |
174 |
BlockDriver *bdrv_find_whitelisted_format(const char *format_name, |
175 |
bool readonly);
|
176 |
int bdrv_create(BlockDriver *drv, const char* filename, |
177 |
QEMUOptionParameter *options, Error **errp); |
178 |
int bdrv_create_file(const char* filename, QEMUOptionParameter *options, |
179 |
Error **errp); |
180 |
BlockDriverState *bdrv_new(const char *device_name); |
181 |
void bdrv_make_anon(BlockDriverState *bs);
|
182 |
void bdrv_swap(BlockDriverState *bs_new, BlockDriverState *bs_old);
|
183 |
void bdrv_append(BlockDriverState *bs_new, BlockDriverState *bs_top);
|
184 |
int bdrv_parse_cache_flags(const char *mode, int *flags); |
185 |
int bdrv_parse_discard_flags(const char *mode, int *flags); |
186 |
int bdrv_file_open(BlockDriverState **pbs, const char *filename, |
187 |
const char *reference, QDict *options, int flags, |
188 |
Error **errp); |
189 |
int bdrv_open_image(BlockDriverState **pbs, const char *filename, |
190 |
QDict *options, const char *bdref_key, int flags, |
191 |
bool force_raw, bool allow_none, Error **errp); |
192 |
int bdrv_open_backing_file(BlockDriverState *bs, QDict *options, Error **errp);
|
193 |
int bdrv_open(BlockDriverState **pbs, const char *filename, |
194 |
const char *reference, QDict *options, int flags, |
195 |
BlockDriver *drv, Error **errp); |
196 |
BlockReopenQueue *bdrv_reopen_queue(BlockReopenQueue *bs_queue, |
197 |
BlockDriverState *bs, int flags);
|
198 |
int bdrv_reopen_multiple(BlockReopenQueue *bs_queue, Error **errp);
|
199 |
int bdrv_reopen(BlockDriverState *bs, int bdrv_flags, Error **errp); |
200 |
int bdrv_reopen_prepare(BDRVReopenState *reopen_state,
|
201 |
BlockReopenQueue *queue, Error **errp); |
202 |
void bdrv_reopen_commit(BDRVReopenState *reopen_state);
|
203 |
void bdrv_reopen_abort(BDRVReopenState *reopen_state);
|
204 |
void bdrv_close(BlockDriverState *bs);
|
205 |
void bdrv_add_close_notifier(BlockDriverState *bs, Notifier *notify);
|
206 |
int bdrv_attach_dev(BlockDriverState *bs, void *dev); |
207 |
void bdrv_attach_dev_nofail(BlockDriverState *bs, void *dev); |
208 |
void bdrv_detach_dev(BlockDriverState *bs, void *dev); |
209 |
void *bdrv_get_attached_dev(BlockDriverState *bs);
|
210 |
void bdrv_set_dev_ops(BlockDriverState *bs, const BlockDevOps *ops, |
211 |
void *opaque);
|
212 |
void bdrv_dev_eject_request(BlockDriverState *bs, bool force); |
213 |
bool bdrv_dev_has_removable_media(BlockDriverState *bs);
|
214 |
bool bdrv_dev_is_tray_open(BlockDriverState *bs);
|
215 |
bool bdrv_dev_is_medium_locked(BlockDriverState *bs);
|
216 |
int bdrv_read(BlockDriverState *bs, int64_t sector_num,
|
217 |
uint8_t *buf, int nb_sectors);
|
218 |
int bdrv_read_unthrottled(BlockDriverState *bs, int64_t sector_num,
|
219 |
uint8_t *buf, int nb_sectors);
|
220 |
int bdrv_write(BlockDriverState *bs, int64_t sector_num,
|
221 |
const uint8_t *buf, int nb_sectors); |
222 |
int bdrv_write_zeroes(BlockDriverState *bs, int64_t sector_num,
|
223 |
int nb_sectors, BdrvRequestFlags flags);
|
224 |
BlockDriverAIOCB *bdrv_aio_write_zeroes(BlockDriverState *bs, int64_t sector_num, |
225 |
int nb_sectors, BdrvRequestFlags flags,
|
226 |
BlockDriverCompletionFunc *cb, void *opaque);
|
227 |
int bdrv_make_zero(BlockDriverState *bs, BdrvRequestFlags flags);
|
228 |
int bdrv_pread(BlockDriverState *bs, int64_t offset,
|
229 |
void *buf, int count); |
230 |
int bdrv_pwrite(BlockDriverState *bs, int64_t offset,
|
231 |
const void *buf, int count); |
232 |
int bdrv_pwritev(BlockDriverState *bs, int64_t offset, QEMUIOVector *qiov);
|
233 |
int bdrv_pwrite_sync(BlockDriverState *bs, int64_t offset,
|
234 |
const void *buf, int count); |
235 |
int coroutine_fn bdrv_co_readv(BlockDriverState *bs, int64_t sector_num,
|
236 |
int nb_sectors, QEMUIOVector *qiov);
|
237 |
int coroutine_fn bdrv_co_copy_on_readv(BlockDriverState *bs,
|
238 |
int64_t sector_num, int nb_sectors, QEMUIOVector *qiov);
|
239 |
int coroutine_fn bdrv_co_writev(BlockDriverState *bs, int64_t sector_num,
|
240 |
int nb_sectors, QEMUIOVector *qiov);
|
241 |
/*
|
242 |
* Efficiently zero a region of the disk image. Note that this is a regular
|
243 |
* I/O request like read or write and should have a reasonable size. This
|
244 |
* function is not suitable for zeroing the entire image in a single request
|
245 |
* because it may allocate memory for the entire region.
|
246 |
*/
|
247 |
int coroutine_fn bdrv_co_write_zeroes(BlockDriverState *bs, int64_t sector_num,
|
248 |
int nb_sectors, BdrvRequestFlags flags);
|
249 |
BlockDriverState *bdrv_find_backing_image(BlockDriverState *bs, |
250 |
const char *backing_file); |
251 |
int bdrv_get_backing_file_depth(BlockDriverState *bs);
|
252 |
int bdrv_truncate(BlockDriverState *bs, int64_t offset);
|
253 |
int64_t bdrv_getlength(BlockDriverState *bs); |
254 |
int64_t bdrv_get_allocated_file_size(BlockDriverState *bs); |
255 |
void bdrv_get_geometry(BlockDriverState *bs, uint64_t *nb_sectors_ptr);
|
256 |
int bdrv_refresh_limits(BlockDriverState *bs);
|
257 |
int bdrv_commit(BlockDriverState *bs);
|
258 |
int bdrv_commit_all(void); |
259 |
int bdrv_change_backing_file(BlockDriverState *bs,
|
260 |
const char *backing_file, const char *backing_fmt); |
261 |
void bdrv_register(BlockDriver *bdrv);
|
262 |
int bdrv_drop_intermediate(BlockDriverState *active, BlockDriverState *top,
|
263 |
BlockDriverState *base); |
264 |
BlockDriverState *bdrv_find_overlay(BlockDriverState *active, |
265 |
BlockDriverState *bs); |
266 |
BlockDriverState *bdrv_find_base(BlockDriverState *bs); |
267 |
|
268 |
|
269 |
typedef struct BdrvCheckResult { |
270 |
int corruptions;
|
271 |
int leaks;
|
272 |
int check_errors;
|
273 |
int corruptions_fixed;
|
274 |
int leaks_fixed;
|
275 |
int64_t image_end_offset; |
276 |
BlockFragInfo bfi; |
277 |
} BdrvCheckResult; |
278 |
|
279 |
typedef enum { |
280 |
BDRV_FIX_LEAKS = 1,
|
281 |
BDRV_FIX_ERRORS = 2,
|
282 |
} BdrvCheckMode; |
283 |
|
284 |
int bdrv_check(BlockDriverState *bs, BdrvCheckResult *res, BdrvCheckMode fix);
|
285 |
|
286 |
int bdrv_amend_options(BlockDriverState *bs_new, QEMUOptionParameter *options);
|
287 |
|
288 |
/* external snapshots */
|
289 |
|
290 |
typedef enum { |
291 |
BS_IS_A_FILTER, |
292 |
BS_FILTER_PASS_DOWN, |
293 |
BS_AUTHORIZATION_COUNT, |
294 |
} BsAuthorization; |
295 |
|
296 |
bool bdrv_generic_is_first_non_filter(BlockDriverState *bs,
|
297 |
BlockDriverState *candidate); |
298 |
bool bdrv_recurse_is_first_non_filter(BlockDriverState *bs,
|
299 |
BlockDriverState *candidate); |
300 |
bool bdrv_is_first_non_filter(BlockDriverState *candidate);
|
301 |
|
302 |
/* async block I/O */
|
303 |
typedef void BlockDriverDirtyHandler(BlockDriverState *bs, int64_t sector, |
304 |
int sector_num);
|
305 |
BlockDriverAIOCB *bdrv_aio_readv(BlockDriverState *bs, int64_t sector_num, |
306 |
QEMUIOVector *iov, int nb_sectors,
|
307 |
BlockDriverCompletionFunc *cb, void *opaque);
|
308 |
BlockDriverAIOCB *bdrv_aio_writev(BlockDriverState *bs, int64_t sector_num, |
309 |
QEMUIOVector *iov, int nb_sectors,
|
310 |
BlockDriverCompletionFunc *cb, void *opaque);
|
311 |
BlockDriverAIOCB *bdrv_aio_flush(BlockDriverState *bs, |
312 |
BlockDriverCompletionFunc *cb, void *opaque);
|
313 |
BlockDriverAIOCB *bdrv_aio_discard(BlockDriverState *bs, |
314 |
int64_t sector_num, int nb_sectors,
|
315 |
BlockDriverCompletionFunc *cb, void *opaque);
|
316 |
void bdrv_aio_cancel(BlockDriverAIOCB *acb);
|
317 |
|
318 |
typedef struct BlockRequest { |
319 |
/* Fields to be filled by multiwrite caller */
|
320 |
int64_t sector; |
321 |
int nb_sectors;
|
322 |
int flags;
|
323 |
QEMUIOVector *qiov; |
324 |
BlockDriverCompletionFunc *cb; |
325 |
void *opaque;
|
326 |
|
327 |
/* Filled by multiwrite implementation */
|
328 |
int error;
|
329 |
} BlockRequest; |
330 |
|
331 |
int bdrv_aio_multiwrite(BlockDriverState *bs, BlockRequest *reqs,
|
332 |
int num_reqs);
|
333 |
|
334 |
/* sg packet commands */
|
335 |
int bdrv_ioctl(BlockDriverState *bs, unsigned long int req, void *buf); |
336 |
BlockDriverAIOCB *bdrv_aio_ioctl(BlockDriverState *bs, |
337 |
unsigned long int req, void *buf, |
338 |
BlockDriverCompletionFunc *cb, void *opaque);
|
339 |
|
340 |
/* Invalidate any cached metadata used by image formats */
|
341 |
void bdrv_invalidate_cache(BlockDriverState *bs);
|
342 |
void bdrv_invalidate_cache_all(void); |
343 |
|
344 |
void bdrv_clear_incoming_migration_all(void); |
345 |
|
346 |
/* Ensure contents are flushed to disk. */
|
347 |
int bdrv_flush(BlockDriverState *bs);
|
348 |
int coroutine_fn bdrv_co_flush(BlockDriverState *bs);
|
349 |
int bdrv_flush_all(void); |
350 |
void bdrv_close_all(void); |
351 |
void bdrv_drain_all(void); |
352 |
|
353 |
int bdrv_discard(BlockDriverState *bs, int64_t sector_num, int nb_sectors); |
354 |
int bdrv_co_discard(BlockDriverState *bs, int64_t sector_num, int nb_sectors); |
355 |
int bdrv_has_zero_init_1(BlockDriverState *bs);
|
356 |
int bdrv_has_zero_init(BlockDriverState *bs);
|
357 |
bool bdrv_unallocated_blocks_are_zero(BlockDriverState *bs);
|
358 |
bool bdrv_can_write_zeroes_with_unmap(BlockDriverState *bs);
|
359 |
int64_t bdrv_get_block_status(BlockDriverState *bs, int64_t sector_num, |
360 |
int nb_sectors, int *pnum); |
361 |
int bdrv_is_allocated(BlockDriverState *bs, int64_t sector_num, int nb_sectors, |
362 |
int *pnum);
|
363 |
int bdrv_is_allocated_above(BlockDriverState *top, BlockDriverState *base,
|
364 |
int64_t sector_num, int nb_sectors, int *pnum); |
365 |
|
366 |
void bdrv_set_on_error(BlockDriverState *bs, BlockdevOnError on_read_error,
|
367 |
BlockdevOnError on_write_error); |
368 |
BlockdevOnError bdrv_get_on_error(BlockDriverState *bs, bool is_read);
|
369 |
BlockErrorAction bdrv_get_error_action(BlockDriverState *bs, bool is_read, int error); |
370 |
void bdrv_error_action(BlockDriverState *bs, BlockErrorAction action,
|
371 |
bool is_read, int error); |
372 |
int bdrv_is_read_only(BlockDriverState *bs);
|
373 |
int bdrv_is_sg(BlockDriverState *bs);
|
374 |
int bdrv_enable_write_cache(BlockDriverState *bs);
|
375 |
void bdrv_set_enable_write_cache(BlockDriverState *bs, bool wce); |
376 |
int bdrv_is_inserted(BlockDriverState *bs);
|
377 |
int bdrv_media_changed(BlockDriverState *bs);
|
378 |
void bdrv_lock_medium(BlockDriverState *bs, bool locked); |
379 |
void bdrv_eject(BlockDriverState *bs, bool eject_flag); |
380 |
const char *bdrv_get_format_name(BlockDriverState *bs); |
381 |
BlockDriverState *bdrv_find(const char *name); |
382 |
BlockDriverState *bdrv_find_node(const char *node_name); |
383 |
BlockDeviceInfoList *bdrv_named_nodes_list(void);
|
384 |
BlockDriverState *bdrv_lookup_bs(const char *device, |
385 |
const char *node_name, |
386 |
Error **errp); |
387 |
BlockDriverState *bdrv_next(BlockDriverState *bs); |
388 |
void bdrv_iterate(void (*it)(void *opaque, BlockDriverState *bs), |
389 |
void *opaque);
|
390 |
int bdrv_is_encrypted(BlockDriverState *bs);
|
391 |
int bdrv_key_required(BlockDriverState *bs);
|
392 |
int bdrv_set_key(BlockDriverState *bs, const char *key); |
393 |
int bdrv_query_missing_keys(void); |
394 |
void bdrv_iterate_format(void (*it)(void *opaque, const char *name), |
395 |
void *opaque);
|
396 |
const char *bdrv_get_device_name(BlockDriverState *bs); |
397 |
int bdrv_get_flags(BlockDriverState *bs);
|
398 |
int bdrv_write_compressed(BlockDriverState *bs, int64_t sector_num,
|
399 |
const uint8_t *buf, int nb_sectors); |
400 |
int bdrv_get_info(BlockDriverState *bs, BlockDriverInfo *bdi);
|
401 |
ImageInfoSpecific *bdrv_get_specific_info(BlockDriverState *bs); |
402 |
void bdrv_round_to_clusters(BlockDriverState *bs,
|
403 |
int64_t sector_num, int nb_sectors,
|
404 |
int64_t *cluster_sector_num, |
405 |
int *cluster_nb_sectors);
|
406 |
|
407 |
const char *bdrv_get_encrypted_filename(BlockDriverState *bs); |
408 |
void bdrv_get_backing_filename(BlockDriverState *bs,
|
409 |
char *filename, int filename_size); |
410 |
void bdrv_get_full_backing_filename(BlockDriverState *bs,
|
411 |
char *dest, size_t sz);
|
412 |
int bdrv_is_snapshot(BlockDriverState *bs);
|
413 |
|
414 |
int path_is_absolute(const char *path); |
415 |
void path_combine(char *dest, int dest_size, |
416 |
const char *base_path, |
417 |
const char *filename); |
418 |
|
419 |
int bdrv_writev_vmstate(BlockDriverState *bs, QEMUIOVector *qiov, int64_t pos);
|
420 |
int bdrv_save_vmstate(BlockDriverState *bs, const uint8_t *buf, |
421 |
int64_t pos, int size);
|
422 |
|
423 |
int bdrv_load_vmstate(BlockDriverState *bs, uint8_t *buf,
|
424 |
int64_t pos, int size);
|
425 |
|
426 |
void bdrv_img_create(const char *filename, const char *fmt, |
427 |
const char *base_filename, const char *base_fmt, |
428 |
char *options, uint64_t img_size, int flags, |
429 |
Error **errp, bool quiet);
|
430 |
|
431 |
/* Returns the alignment in bytes that is required so that no bounce buffer
|
432 |
* is required throughout the stack */
|
433 |
size_t bdrv_opt_mem_align(BlockDriverState *bs); |
434 |
void bdrv_set_guest_block_size(BlockDriverState *bs, int align); |
435 |
void *qemu_blockalign(BlockDriverState *bs, size_t size);
|
436 |
bool bdrv_qiov_is_aligned(BlockDriverState *bs, QEMUIOVector *qiov);
|
437 |
|
438 |
struct HBitmapIter;
|
439 |
typedef struct BdrvDirtyBitmap BdrvDirtyBitmap; |
440 |
BdrvDirtyBitmap *bdrv_create_dirty_bitmap(BlockDriverState *bs, int granularity);
|
441 |
void bdrv_release_dirty_bitmap(BlockDriverState *bs, BdrvDirtyBitmap *bitmap);
|
442 |
BlockDirtyInfoList *bdrv_query_dirty_bitmaps(BlockDriverState *bs); |
443 |
int bdrv_get_dirty(BlockDriverState *bs, BdrvDirtyBitmap *bitmap, int64_t sector);
|
444 |
void bdrv_set_dirty(BlockDriverState *bs, int64_t cur_sector, int nr_sectors); |
445 |
void bdrv_reset_dirty(BlockDriverState *bs, int64_t cur_sector, int nr_sectors); |
446 |
void bdrv_dirty_iter_init(BlockDriverState *bs,
|
447 |
BdrvDirtyBitmap *bitmap, struct HBitmapIter *hbi);
|
448 |
int64_t bdrv_get_dirty_count(BlockDriverState *bs, BdrvDirtyBitmap *bitmap); |
449 |
|
450 |
void bdrv_enable_copy_on_read(BlockDriverState *bs);
|
451 |
void bdrv_disable_copy_on_read(BlockDriverState *bs);
|
452 |
|
453 |
void bdrv_ref(BlockDriverState *bs);
|
454 |
void bdrv_unref(BlockDriverState *bs);
|
455 |
void bdrv_set_in_use(BlockDriverState *bs, int in_use); |
456 |
int bdrv_in_use(BlockDriverState *bs);
|
457 |
|
458 |
#ifdef CONFIG_LINUX_AIO
|
459 |
int raw_get_aio_fd(BlockDriverState *bs);
|
460 |
#else
|
461 |
static inline int raw_get_aio_fd(BlockDriverState *bs) |
462 |
{ |
463 |
return -ENOTSUP;
|
464 |
} |
465 |
#endif
|
466 |
|
467 |
enum BlockAcctType {
|
468 |
BDRV_ACCT_READ, |
469 |
BDRV_ACCT_WRITE, |
470 |
BDRV_ACCT_FLUSH, |
471 |
BDRV_MAX_IOTYPE, |
472 |
}; |
473 |
|
474 |
typedef struct BlockAcctCookie { |
475 |
int64_t bytes; |
476 |
int64_t start_time_ns; |
477 |
enum BlockAcctType type;
|
478 |
} BlockAcctCookie; |
479 |
|
480 |
void bdrv_acct_start(BlockDriverState *bs, BlockAcctCookie *cookie,
|
481 |
int64_t bytes, enum BlockAcctType type);
|
482 |
void bdrv_acct_done(BlockDriverState *bs, BlockAcctCookie *cookie);
|
483 |
|
484 |
typedef enum { |
485 |
BLKDBG_L1_UPDATE, |
486 |
|
487 |
BLKDBG_L1_GROW_ALLOC_TABLE, |
488 |
BLKDBG_L1_GROW_WRITE_TABLE, |
489 |
BLKDBG_L1_GROW_ACTIVATE_TABLE, |
490 |
|
491 |
BLKDBG_L2_LOAD, |
492 |
BLKDBG_L2_UPDATE, |
493 |
BLKDBG_L2_UPDATE_COMPRESSED, |
494 |
BLKDBG_L2_ALLOC_COW_READ, |
495 |
BLKDBG_L2_ALLOC_WRITE, |
496 |
|
497 |
BLKDBG_READ_AIO, |
498 |
BLKDBG_READ_BACKING_AIO, |
499 |
BLKDBG_READ_COMPRESSED, |
500 |
|
501 |
BLKDBG_WRITE_AIO, |
502 |
BLKDBG_WRITE_COMPRESSED, |
503 |
|
504 |
BLKDBG_VMSTATE_LOAD, |
505 |
BLKDBG_VMSTATE_SAVE, |
506 |
|
507 |
BLKDBG_COW_READ, |
508 |
BLKDBG_COW_WRITE, |
509 |
|
510 |
BLKDBG_REFTABLE_LOAD, |
511 |
BLKDBG_REFTABLE_GROW, |
512 |
BLKDBG_REFTABLE_UPDATE, |
513 |
|
514 |
BLKDBG_REFBLOCK_LOAD, |
515 |
BLKDBG_REFBLOCK_UPDATE, |
516 |
BLKDBG_REFBLOCK_UPDATE_PART, |
517 |
BLKDBG_REFBLOCK_ALLOC, |
518 |
BLKDBG_REFBLOCK_ALLOC_HOOKUP, |
519 |
BLKDBG_REFBLOCK_ALLOC_WRITE, |
520 |
BLKDBG_REFBLOCK_ALLOC_WRITE_BLOCKS, |
521 |
BLKDBG_REFBLOCK_ALLOC_WRITE_TABLE, |
522 |
BLKDBG_REFBLOCK_ALLOC_SWITCH_TABLE, |
523 |
|
524 |
BLKDBG_CLUSTER_ALLOC, |
525 |
BLKDBG_CLUSTER_ALLOC_BYTES, |
526 |
BLKDBG_CLUSTER_FREE, |
527 |
|
528 |
BLKDBG_FLUSH_TO_OS, |
529 |
BLKDBG_FLUSH_TO_DISK, |
530 |
|
531 |
BLKDBG_PWRITEV_RMW_HEAD, |
532 |
BLKDBG_PWRITEV_RMW_AFTER_HEAD, |
533 |
BLKDBG_PWRITEV_RMW_TAIL, |
534 |
BLKDBG_PWRITEV_RMW_AFTER_TAIL, |
535 |
BLKDBG_PWRITEV, |
536 |
BLKDBG_PWRITEV_ZERO, |
537 |
BLKDBG_PWRITEV_DONE, |
538 |
|
539 |
BLKDBG_EVENT_MAX, |
540 |
} BlkDebugEvent; |
541 |
|
542 |
#define BLKDBG_EVENT(bs, evt) bdrv_debug_event(bs, evt)
|
543 |
void bdrv_debug_event(BlockDriverState *bs, BlkDebugEvent event);
|
544 |
|
545 |
int bdrv_debug_breakpoint(BlockDriverState *bs, const char *event, |
546 |
const char *tag); |
547 |
int bdrv_debug_remove_breakpoint(BlockDriverState *bs, const char *tag); |
548 |
int bdrv_debug_resume(BlockDriverState *bs, const char *tag); |
549 |
bool bdrv_debug_is_suspended(BlockDriverState *bs, const char *tag); |
550 |
|
551 |
#endif
|