root / include / block / block.h @ e4654d2d
History | View | Annotate | Download (21.2 kB)
1 |
#ifndef BLOCK_H
|
---|---|
2 |
#define BLOCK_H
|
3 |
|
4 |
#include "block/aio.h" |
5 |
#include "qemu-common.h" |
6 |
#include "qemu/option.h" |
7 |
#include "block/coroutine.h" |
8 |
#include "qapi/qmp/qobject.h" |
9 |
#include "qapi-types.h" |
10 |
|
11 |
/* block.c */
|
12 |
typedef struct BlockDriver BlockDriver; |
13 |
typedef struct BlockJob BlockJob; |
14 |
|
15 |
typedef struct BlockDriverInfo { |
16 |
/* in bytes, 0 if irrelevant */
|
17 |
int cluster_size;
|
18 |
/* offset at which the VM state can be saved (0 if not possible) */
|
19 |
int64_t vm_state_offset; |
20 |
bool is_dirty;
|
21 |
/*
|
22 |
* True if unallocated blocks read back as zeroes. This is equivalent
|
23 |
* to the the LBPRZ flag in the SCSI logical block provisioning page.
|
24 |
*/
|
25 |
bool unallocated_blocks_are_zero;
|
26 |
/*
|
27 |
* True if the driver can optimize writing zeroes by unmapping
|
28 |
* sectors. This is equivalent to the BLKDISCARDZEROES ioctl in Linux
|
29 |
* with the difference that in qemu a discard is allowed to silently
|
30 |
* fail. Therefore we have to use bdrv_write_zeroes with the
|
31 |
* BDRV_REQ_MAY_UNMAP flag for an optimized zero write with unmapping.
|
32 |
* After this call the driver has to guarantee that the contents read
|
33 |
* back as zero. It is additionally required that the block device is
|
34 |
* opened with BDRV_O_UNMAP flag for this to work.
|
35 |
*/
|
36 |
bool can_write_zeroes_with_unmap;
|
37 |
} BlockDriverInfo; |
38 |
|
39 |
typedef struct BlockFragInfo { |
40 |
uint64_t allocated_clusters; |
41 |
uint64_t total_clusters; |
42 |
uint64_t fragmented_clusters; |
43 |
uint64_t compressed_clusters; |
44 |
} BlockFragInfo; |
45 |
|
46 |
/* Callbacks for block device models */
|
47 |
typedef struct BlockDevOps { |
48 |
/*
|
49 |
* Runs when virtual media changed (monitor commands eject, change)
|
50 |
* Argument load is true on load and false on eject.
|
51 |
* Beware: doesn't run when a host device's physical media
|
52 |
* changes. Sure would be useful if it did.
|
53 |
* Device models with removable media must implement this callback.
|
54 |
*/
|
55 |
void (*change_media_cb)(void *opaque, bool load); |
56 |
/*
|
57 |
* Runs when an eject request is issued from the monitor, the tray
|
58 |
* is closed, and the medium is locked.
|
59 |
* Device models that do not implement is_medium_locked will not need
|
60 |
* this callback. Device models that can lock the medium or tray might
|
61 |
* want to implement the callback and unlock the tray when "force" is
|
62 |
* true, even if they do not support eject requests.
|
63 |
*/
|
64 |
void (*eject_request_cb)(void *opaque, bool force); |
65 |
/*
|
66 |
* Is the virtual tray open?
|
67 |
* Device models implement this only when the device has a tray.
|
68 |
*/
|
69 |
bool (*is_tray_open)(void *opaque); |
70 |
/*
|
71 |
* Is the virtual medium locked into the device?
|
72 |
* Device models implement this only when device has such a lock.
|
73 |
*/
|
74 |
bool (*is_medium_locked)(void *opaque); |
75 |
/*
|
76 |
* Runs when the size changed (e.g. monitor command block_resize)
|
77 |
*/
|
78 |
void (*resize_cb)(void *opaque); |
79 |
} BlockDevOps; |
80 |
|
81 |
typedef enum { |
82 |
BDRV_REQ_COPY_ON_READ = 0x1,
|
83 |
BDRV_REQ_ZERO_WRITE = 0x2,
|
84 |
/* The BDRV_REQ_MAY_UNMAP flag is used to indicate that the block driver
|
85 |
* is allowed to optimize a write zeroes request by unmapping (discarding)
|
86 |
* blocks if it is guaranteed that the result will read back as
|
87 |
* zeroes. The flag is only passed to the driver if the block device is
|
88 |
* opened with BDRV_O_UNMAP.
|
89 |
*/
|
90 |
BDRV_REQ_MAY_UNMAP = 0x4,
|
91 |
} BdrvRequestFlags; |
92 |
|
93 |
#define BDRV_O_RDWR 0x0002 |
94 |
#define BDRV_O_SNAPSHOT 0x0008 /* open the file read only and save writes in a snapshot */ |
95 |
#define BDRV_O_NOCACHE 0x0020 /* do not use the host page cache */ |
96 |
#define BDRV_O_CACHE_WB 0x0040 /* use write-back caching */ |
97 |
#define BDRV_O_NATIVE_AIO 0x0080 /* use native AIO instead of the thread pool */ |
98 |
#define BDRV_O_NO_BACKING 0x0100 /* don't open the backing file */ |
99 |
#define BDRV_O_NO_FLUSH 0x0200 /* disable flushing on this disk */ |
100 |
#define BDRV_O_COPY_ON_READ 0x0400 /* copy read backing sectors into image */ |
101 |
#define BDRV_O_INCOMING 0x0800 /* consistency hint for incoming migration */ |
102 |
#define BDRV_O_CHECK 0x1000 /* open solely for consistency check */ |
103 |
#define BDRV_O_ALLOW_RDWR 0x2000 /* allow reopen to change from r/o to r/w */ |
104 |
#define BDRV_O_UNMAP 0x4000 /* execute guest UNMAP/TRIM operations */ |
105 |
|
106 |
#define BDRV_O_CACHE_MASK (BDRV_O_NOCACHE | BDRV_O_CACHE_WB | BDRV_O_NO_FLUSH)
|
107 |
|
108 |
#define BDRV_SECTOR_BITS 9 |
109 |
#define BDRV_SECTOR_SIZE (1ULL << BDRV_SECTOR_BITS) |
110 |
#define BDRV_SECTOR_MASK ~(BDRV_SECTOR_SIZE - 1) |
111 |
|
112 |
/* BDRV_BLOCK_DATA: data is read from bs->file or another file
|
113 |
* BDRV_BLOCK_ZERO: sectors read as zero
|
114 |
* BDRV_BLOCK_OFFSET_VALID: sector stored in bs->file as raw data
|
115 |
* BDRV_BLOCK_RAW: used internally to indicate that the request
|
116 |
* was answered by the raw driver and that one
|
117 |
* should look in bs->file directly.
|
118 |
*
|
119 |
* If BDRV_BLOCK_OFFSET_VALID is set, bits 9-62 represent the offset in
|
120 |
* bs->file where sector data can be read from as raw data.
|
121 |
*
|
122 |
* DATA == 0 && ZERO == 0 means that data is read from backing_hd if present.
|
123 |
*
|
124 |
* DATA ZERO OFFSET_VALID
|
125 |
* t t t sectors read as zero, bs->file is zero at offset
|
126 |
* t f t sectors read as valid from bs->file at offset
|
127 |
* f t t sectors preallocated, read as zero, bs->file not
|
128 |
* necessarily zero at offset
|
129 |
* f f t sectors preallocated but read from backing_hd,
|
130 |
* bs->file contains garbage at offset
|
131 |
* t t f sectors preallocated, read as zero, unknown offset
|
132 |
* t f f sectors read from unknown file or offset
|
133 |
* f t f not allocated or unknown offset, read as zero
|
134 |
* f f f not allocated or unknown offset, read from backing_hd
|
135 |
*/
|
136 |
#define BDRV_BLOCK_DATA 1 |
137 |
#define BDRV_BLOCK_ZERO 2 |
138 |
#define BDRV_BLOCK_OFFSET_VALID 4 |
139 |
#define BDRV_BLOCK_RAW 8 |
140 |
#define BDRV_BLOCK_OFFSET_MASK BDRV_SECTOR_MASK
|
141 |
|
142 |
typedef enum { |
143 |
BDRV_ACTION_REPORT, BDRV_ACTION_IGNORE, BDRV_ACTION_STOP |
144 |
} BlockErrorAction; |
145 |
|
146 |
typedef QSIMPLEQ_HEAD(BlockReopenQueue, BlockReopenQueueEntry) BlockReopenQueue;
|
147 |
|
148 |
typedef struct BDRVReopenState { |
149 |
BlockDriverState *bs; |
150 |
int flags;
|
151 |
void *opaque;
|
152 |
} BDRVReopenState; |
153 |
|
154 |
|
155 |
void bdrv_iostatus_enable(BlockDriverState *bs);
|
156 |
void bdrv_iostatus_reset(BlockDriverState *bs);
|
157 |
void bdrv_iostatus_disable(BlockDriverState *bs);
|
158 |
bool bdrv_iostatus_is_enabled(const BlockDriverState *bs); |
159 |
void bdrv_iostatus_set_err(BlockDriverState *bs, int error); |
160 |
void bdrv_info_print(Monitor *mon, const QObject *data); |
161 |
void bdrv_info(Monitor *mon, QObject **ret_data);
|
162 |
void bdrv_stats_print(Monitor *mon, const QObject *data); |
163 |
void bdrv_info_stats(Monitor *mon, QObject **ret_data);
|
164 |
|
165 |
/* disk I/O throttling */
|
166 |
void bdrv_io_limits_enable(BlockDriverState *bs);
|
167 |
void bdrv_io_limits_disable(BlockDriverState *bs);
|
168 |
|
169 |
void bdrv_init(void); |
170 |
void bdrv_init_with_whitelist(void); |
171 |
BlockDriver *bdrv_find_protocol(const char *filename, |
172 |
bool allow_protocol_prefix);
|
173 |
BlockDriver *bdrv_find_format(const char *format_name); |
174 |
BlockDriver *bdrv_find_whitelisted_format(const char *format_name, |
175 |
bool readonly);
|
176 |
int bdrv_create(BlockDriver *drv, const char* filename, |
177 |
QEMUOptionParameter *options, Error **errp); |
178 |
int bdrv_create_file(const char* filename, QEMUOptionParameter *options, |
179 |
Error **errp); |
180 |
BlockDriverState *bdrv_new(const char *device_name); |
181 |
void bdrv_make_anon(BlockDriverState *bs);
|
182 |
void bdrv_swap(BlockDriverState *bs_new, BlockDriverState *bs_old);
|
183 |
void bdrv_append(BlockDriverState *bs_new, BlockDriverState *bs_top);
|
184 |
int bdrv_parse_cache_flags(const char *mode, int *flags); |
185 |
int bdrv_parse_discard_flags(const char *mode, int *flags); |
186 |
int bdrv_file_open(BlockDriverState **pbs, const char *filename, |
187 |
QDict *options, int flags, Error **errp);
|
188 |
int bdrv_open_backing_file(BlockDriverState *bs, QDict *options, Error **errp);
|
189 |
int bdrv_open(BlockDriverState *bs, const char *filename, QDict *options, |
190 |
int flags, BlockDriver *drv, Error **errp);
|
191 |
BlockReopenQueue *bdrv_reopen_queue(BlockReopenQueue *bs_queue, |
192 |
BlockDriverState *bs, int flags);
|
193 |
int bdrv_reopen_multiple(BlockReopenQueue *bs_queue, Error **errp);
|
194 |
int bdrv_reopen(BlockDriverState *bs, int bdrv_flags, Error **errp); |
195 |
int bdrv_reopen_prepare(BDRVReopenState *reopen_state,
|
196 |
BlockReopenQueue *queue, Error **errp); |
197 |
void bdrv_reopen_commit(BDRVReopenState *reopen_state);
|
198 |
void bdrv_reopen_abort(BDRVReopenState *reopen_state);
|
199 |
void bdrv_close(BlockDriverState *bs);
|
200 |
void bdrv_add_close_notifier(BlockDriverState *bs, Notifier *notify);
|
201 |
int bdrv_attach_dev(BlockDriverState *bs, void *dev); |
202 |
void bdrv_attach_dev_nofail(BlockDriverState *bs, void *dev); |
203 |
void bdrv_detach_dev(BlockDriverState *bs, void *dev); |
204 |
void *bdrv_get_attached_dev(BlockDriverState *bs);
|
205 |
void bdrv_set_dev_ops(BlockDriverState *bs, const BlockDevOps *ops, |
206 |
void *opaque);
|
207 |
void bdrv_dev_eject_request(BlockDriverState *bs, bool force); |
208 |
bool bdrv_dev_has_removable_media(BlockDriverState *bs);
|
209 |
bool bdrv_dev_is_tray_open(BlockDriverState *bs);
|
210 |
bool bdrv_dev_is_medium_locked(BlockDriverState *bs);
|
211 |
int bdrv_read(BlockDriverState *bs, int64_t sector_num,
|
212 |
uint8_t *buf, int nb_sectors);
|
213 |
int bdrv_read_unthrottled(BlockDriverState *bs, int64_t sector_num,
|
214 |
uint8_t *buf, int nb_sectors);
|
215 |
int bdrv_write(BlockDriverState *bs, int64_t sector_num,
|
216 |
const uint8_t *buf, int nb_sectors); |
217 |
int bdrv_write_zeroes(BlockDriverState *bs, int64_t sector_num,
|
218 |
int nb_sectors, BdrvRequestFlags flags);
|
219 |
int bdrv_make_zero(BlockDriverState *bs, BdrvRequestFlags flags);
|
220 |
int bdrv_writev(BlockDriverState *bs, int64_t sector_num, QEMUIOVector *qiov);
|
221 |
int bdrv_pread(BlockDriverState *bs, int64_t offset,
|
222 |
void *buf, int count); |
223 |
int bdrv_pwrite(BlockDriverState *bs, int64_t offset,
|
224 |
const void *buf, int count); |
225 |
int bdrv_pwritev(BlockDriverState *bs, int64_t offset, QEMUIOVector *qiov);
|
226 |
int bdrv_pwrite_sync(BlockDriverState *bs, int64_t offset,
|
227 |
const void *buf, int count); |
228 |
int coroutine_fn bdrv_co_readv(BlockDriverState *bs, int64_t sector_num,
|
229 |
int nb_sectors, QEMUIOVector *qiov);
|
230 |
int coroutine_fn bdrv_co_copy_on_readv(BlockDriverState *bs,
|
231 |
int64_t sector_num, int nb_sectors, QEMUIOVector *qiov);
|
232 |
int coroutine_fn bdrv_co_writev(BlockDriverState *bs, int64_t sector_num,
|
233 |
int nb_sectors, QEMUIOVector *qiov);
|
234 |
/*
|
235 |
* Efficiently zero a region of the disk image. Note that this is a regular
|
236 |
* I/O request like read or write and should have a reasonable size. This
|
237 |
* function is not suitable for zeroing the entire image in a single request
|
238 |
* because it may allocate memory for the entire region.
|
239 |
*/
|
240 |
int coroutine_fn bdrv_co_write_zeroes(BlockDriverState *bs, int64_t sector_num,
|
241 |
int nb_sectors, BdrvRequestFlags flags);
|
242 |
BlockDriverState *bdrv_find_backing_image(BlockDriverState *bs, |
243 |
const char *backing_file); |
244 |
int bdrv_get_backing_file_depth(BlockDriverState *bs);
|
245 |
int bdrv_truncate(BlockDriverState *bs, int64_t offset);
|
246 |
int64_t bdrv_getlength(BlockDriverState *bs); |
247 |
int64_t bdrv_get_allocated_file_size(BlockDriverState *bs); |
248 |
void bdrv_get_geometry(BlockDriverState *bs, uint64_t *nb_sectors_ptr);
|
249 |
int bdrv_commit(BlockDriverState *bs);
|
250 |
int bdrv_commit_all(void); |
251 |
int bdrv_change_backing_file(BlockDriverState *bs,
|
252 |
const char *backing_file, const char *backing_fmt); |
253 |
void bdrv_register(BlockDriver *bdrv);
|
254 |
int bdrv_drop_intermediate(BlockDriverState *active, BlockDriverState *top,
|
255 |
BlockDriverState *base); |
256 |
BlockDriverState *bdrv_find_overlay(BlockDriverState *active, |
257 |
BlockDriverState *bs); |
258 |
BlockDriverState *bdrv_find_base(BlockDriverState *bs); |
259 |
|
260 |
|
261 |
typedef struct BdrvCheckResult { |
262 |
int corruptions;
|
263 |
int leaks;
|
264 |
int check_errors;
|
265 |
int corruptions_fixed;
|
266 |
int leaks_fixed;
|
267 |
int64_t image_end_offset; |
268 |
BlockFragInfo bfi; |
269 |
} BdrvCheckResult; |
270 |
|
271 |
typedef enum { |
272 |
BDRV_FIX_LEAKS = 1,
|
273 |
BDRV_FIX_ERRORS = 2,
|
274 |
} BdrvCheckMode; |
275 |
|
276 |
int bdrv_check(BlockDriverState *bs, BdrvCheckResult *res, BdrvCheckMode fix);
|
277 |
|
278 |
int bdrv_amend_options(BlockDriverState *bs_new, QEMUOptionParameter *options);
|
279 |
|
280 |
/* external snapshots */
|
281 |
|
282 |
typedef enum { |
283 |
EXT_SNAPSHOT_ALLOWED, |
284 |
EXT_SNAPSHOT_FORBIDDEN, |
285 |
} ExtSnapshotPerm; |
286 |
|
287 |
/* return EXT_SNAPSHOT_ALLOWED if external snapshot is allowed
|
288 |
* return EXT_SNAPSHOT_FORBIDDEN if external snapshot is forbidden
|
289 |
*/
|
290 |
ExtSnapshotPerm bdrv_check_ext_snapshot(BlockDriverState *bs); |
291 |
/* helper used to forbid external snapshots like in blkverify */
|
292 |
ExtSnapshotPerm bdrv_check_ext_snapshot_forbidden(BlockDriverState *bs); |
293 |
|
294 |
/* async block I/O */
|
295 |
typedef void BlockDriverDirtyHandler(BlockDriverState *bs, int64_t sector, |
296 |
int sector_num);
|
297 |
BlockDriverAIOCB *bdrv_aio_readv(BlockDriverState *bs, int64_t sector_num, |
298 |
QEMUIOVector *iov, int nb_sectors,
|
299 |
BlockDriverCompletionFunc *cb, void *opaque);
|
300 |
BlockDriverAIOCB *bdrv_aio_writev(BlockDriverState *bs, int64_t sector_num, |
301 |
QEMUIOVector *iov, int nb_sectors,
|
302 |
BlockDriverCompletionFunc *cb, void *opaque);
|
303 |
BlockDriverAIOCB *bdrv_aio_flush(BlockDriverState *bs, |
304 |
BlockDriverCompletionFunc *cb, void *opaque);
|
305 |
BlockDriverAIOCB *bdrv_aio_discard(BlockDriverState *bs, |
306 |
int64_t sector_num, int nb_sectors,
|
307 |
BlockDriverCompletionFunc *cb, void *opaque);
|
308 |
void bdrv_aio_cancel(BlockDriverAIOCB *acb);
|
309 |
|
310 |
typedef struct BlockRequest { |
311 |
/* Fields to be filled by multiwrite caller */
|
312 |
int64_t sector; |
313 |
int nb_sectors;
|
314 |
QEMUIOVector *qiov; |
315 |
BlockDriverCompletionFunc *cb; |
316 |
void *opaque;
|
317 |
|
318 |
/* Filled by multiwrite implementation */
|
319 |
int error;
|
320 |
} BlockRequest; |
321 |
|
322 |
int bdrv_aio_multiwrite(BlockDriverState *bs, BlockRequest *reqs,
|
323 |
int num_reqs);
|
324 |
|
325 |
/* sg packet commands */
|
326 |
int bdrv_ioctl(BlockDriverState *bs, unsigned long int req, void *buf); |
327 |
BlockDriverAIOCB *bdrv_aio_ioctl(BlockDriverState *bs, |
328 |
unsigned long int req, void *buf, |
329 |
BlockDriverCompletionFunc *cb, void *opaque);
|
330 |
|
331 |
/* Invalidate any cached metadata used by image formats */
|
332 |
void bdrv_invalidate_cache(BlockDriverState *bs);
|
333 |
void bdrv_invalidate_cache_all(void); |
334 |
|
335 |
void bdrv_clear_incoming_migration_all(void); |
336 |
|
337 |
/* Ensure contents are flushed to disk. */
|
338 |
int bdrv_flush(BlockDriverState *bs);
|
339 |
int coroutine_fn bdrv_co_flush(BlockDriverState *bs);
|
340 |
int bdrv_flush_all(void); |
341 |
void bdrv_close_all(void); |
342 |
void bdrv_drain_all(void); |
343 |
|
344 |
int bdrv_discard(BlockDriverState *bs, int64_t sector_num, int nb_sectors); |
345 |
int bdrv_co_discard(BlockDriverState *bs, int64_t sector_num, int nb_sectors); |
346 |
int bdrv_has_zero_init_1(BlockDriverState *bs);
|
347 |
int bdrv_has_zero_init(BlockDriverState *bs);
|
348 |
bool bdrv_unallocated_blocks_are_zero(BlockDriverState *bs);
|
349 |
bool bdrv_can_write_zeroes_with_unmap(BlockDriverState *bs);
|
350 |
int64_t bdrv_get_block_status(BlockDriverState *bs, int64_t sector_num, |
351 |
int nb_sectors, int *pnum); |
352 |
int bdrv_is_allocated(BlockDriverState *bs, int64_t sector_num, int nb_sectors, |
353 |
int *pnum);
|
354 |
int bdrv_is_allocated_above(BlockDriverState *top, BlockDriverState *base,
|
355 |
int64_t sector_num, int nb_sectors, int *pnum); |
356 |
|
357 |
void bdrv_set_on_error(BlockDriverState *bs, BlockdevOnError on_read_error,
|
358 |
BlockdevOnError on_write_error); |
359 |
BlockdevOnError bdrv_get_on_error(BlockDriverState *bs, bool is_read);
|
360 |
BlockErrorAction bdrv_get_error_action(BlockDriverState *bs, bool is_read, int error); |
361 |
void bdrv_error_action(BlockDriverState *bs, BlockErrorAction action,
|
362 |
bool is_read, int error); |
363 |
int bdrv_is_read_only(BlockDriverState *bs);
|
364 |
int bdrv_is_sg(BlockDriverState *bs);
|
365 |
int bdrv_enable_write_cache(BlockDriverState *bs);
|
366 |
void bdrv_set_enable_write_cache(BlockDriverState *bs, bool wce); |
367 |
int bdrv_is_inserted(BlockDriverState *bs);
|
368 |
int bdrv_media_changed(BlockDriverState *bs);
|
369 |
void bdrv_lock_medium(BlockDriverState *bs, bool locked); |
370 |
void bdrv_eject(BlockDriverState *bs, bool eject_flag); |
371 |
const char *bdrv_get_format_name(BlockDriverState *bs); |
372 |
BlockDriverState *bdrv_find(const char *name); |
373 |
BlockDriverState *bdrv_next(BlockDriverState *bs); |
374 |
void bdrv_iterate(void (*it)(void *opaque, BlockDriverState *bs), |
375 |
void *opaque);
|
376 |
int bdrv_is_encrypted(BlockDriverState *bs);
|
377 |
int bdrv_key_required(BlockDriverState *bs);
|
378 |
int bdrv_set_key(BlockDriverState *bs, const char *key); |
379 |
int bdrv_query_missing_keys(void); |
380 |
void bdrv_iterate_format(void (*it)(void *opaque, const char *name), |
381 |
void *opaque);
|
382 |
const char *bdrv_get_device_name(BlockDriverState *bs); |
383 |
int bdrv_get_flags(BlockDriverState *bs);
|
384 |
int bdrv_write_compressed(BlockDriverState *bs, int64_t sector_num,
|
385 |
const uint8_t *buf, int nb_sectors); |
386 |
int bdrv_get_info(BlockDriverState *bs, BlockDriverInfo *bdi);
|
387 |
ImageInfoSpecific *bdrv_get_specific_info(BlockDriverState *bs); |
388 |
void bdrv_round_to_clusters(BlockDriverState *bs,
|
389 |
int64_t sector_num, int nb_sectors,
|
390 |
int64_t *cluster_sector_num, |
391 |
int *cluster_nb_sectors);
|
392 |
|
393 |
const char *bdrv_get_encrypted_filename(BlockDriverState *bs); |
394 |
void bdrv_get_backing_filename(BlockDriverState *bs,
|
395 |
char *filename, int filename_size); |
396 |
void bdrv_get_full_backing_filename(BlockDriverState *bs,
|
397 |
char *dest, size_t sz);
|
398 |
int bdrv_is_snapshot(BlockDriverState *bs);
|
399 |
|
400 |
int path_is_absolute(const char *path); |
401 |
void path_combine(char *dest, int dest_size, |
402 |
const char *base_path, |
403 |
const char *filename); |
404 |
|
405 |
int bdrv_writev_vmstate(BlockDriverState *bs, QEMUIOVector *qiov, int64_t pos);
|
406 |
int bdrv_save_vmstate(BlockDriverState *bs, const uint8_t *buf, |
407 |
int64_t pos, int size);
|
408 |
|
409 |
int bdrv_load_vmstate(BlockDriverState *bs, uint8_t *buf,
|
410 |
int64_t pos, int size);
|
411 |
|
412 |
void bdrv_img_create(const char *filename, const char *fmt, |
413 |
const char *base_filename, const char *base_fmt, |
414 |
char *options, uint64_t img_size, int flags, |
415 |
Error **errp, bool quiet);
|
416 |
|
417 |
void bdrv_set_buffer_alignment(BlockDriverState *bs, int align); |
418 |
void *qemu_blockalign(BlockDriverState *bs, size_t size);
|
419 |
bool bdrv_qiov_is_aligned(BlockDriverState *bs, QEMUIOVector *qiov);
|
420 |
|
421 |
struct HBitmapIter;
|
422 |
typedef struct BdrvDirtyBitmap BdrvDirtyBitmap; |
423 |
BdrvDirtyBitmap *bdrv_create_dirty_bitmap(BlockDriverState *bs, int granularity);
|
424 |
void bdrv_release_dirty_bitmap(BlockDriverState *bs, BdrvDirtyBitmap *bitmap);
|
425 |
int bdrv_get_dirty(BlockDriverState *bs, BdrvDirtyBitmap *bitmap, int64_t sector);
|
426 |
void bdrv_set_dirty(BlockDriverState *bs, int64_t cur_sector, int nr_sectors); |
427 |
void bdrv_reset_dirty(BlockDriverState *bs, int64_t cur_sector, int nr_sectors); |
428 |
void bdrv_dirty_iter_init(BlockDriverState *bs,
|
429 |
BdrvDirtyBitmap *bitmap, struct HBitmapIter *hbi);
|
430 |
int64_t bdrv_get_dirty_count(BlockDriverState *bs, BdrvDirtyBitmap *bitmap); |
431 |
|
432 |
void bdrv_enable_copy_on_read(BlockDriverState *bs);
|
433 |
void bdrv_disable_copy_on_read(BlockDriverState *bs);
|
434 |
|
435 |
void bdrv_ref(BlockDriverState *bs);
|
436 |
void bdrv_unref(BlockDriverState *bs);
|
437 |
void bdrv_set_in_use(BlockDriverState *bs, int in_use); |
438 |
int bdrv_in_use(BlockDriverState *bs);
|
439 |
|
440 |
#ifdef CONFIG_LINUX_AIO
|
441 |
int raw_get_aio_fd(BlockDriverState *bs);
|
442 |
#else
|
443 |
static inline int raw_get_aio_fd(BlockDriverState *bs) |
444 |
{ |
445 |
return -ENOTSUP;
|
446 |
} |
447 |
#endif
|
448 |
|
449 |
enum BlockAcctType {
|
450 |
BDRV_ACCT_READ, |
451 |
BDRV_ACCT_WRITE, |
452 |
BDRV_ACCT_FLUSH, |
453 |
BDRV_MAX_IOTYPE, |
454 |
}; |
455 |
|
456 |
typedef struct BlockAcctCookie { |
457 |
int64_t bytes; |
458 |
int64_t start_time_ns; |
459 |
enum BlockAcctType type;
|
460 |
} BlockAcctCookie; |
461 |
|
462 |
void bdrv_acct_start(BlockDriverState *bs, BlockAcctCookie *cookie,
|
463 |
int64_t bytes, enum BlockAcctType type);
|
464 |
void bdrv_acct_done(BlockDriverState *bs, BlockAcctCookie *cookie);
|
465 |
|
466 |
typedef enum { |
467 |
BLKDBG_L1_UPDATE, |
468 |
|
469 |
BLKDBG_L1_GROW_ALLOC_TABLE, |
470 |
BLKDBG_L1_GROW_WRITE_TABLE, |
471 |
BLKDBG_L1_GROW_ACTIVATE_TABLE, |
472 |
|
473 |
BLKDBG_L2_LOAD, |
474 |
BLKDBG_L2_UPDATE, |
475 |
BLKDBG_L2_UPDATE_COMPRESSED, |
476 |
BLKDBG_L2_ALLOC_COW_READ, |
477 |
BLKDBG_L2_ALLOC_WRITE, |
478 |
|
479 |
BLKDBG_READ_AIO, |
480 |
BLKDBG_READ_BACKING_AIO, |
481 |
BLKDBG_READ_COMPRESSED, |
482 |
|
483 |
BLKDBG_WRITE_AIO, |
484 |
BLKDBG_WRITE_COMPRESSED, |
485 |
|
486 |
BLKDBG_VMSTATE_LOAD, |
487 |
BLKDBG_VMSTATE_SAVE, |
488 |
|
489 |
BLKDBG_COW_READ, |
490 |
BLKDBG_COW_WRITE, |
491 |
|
492 |
BLKDBG_REFTABLE_LOAD, |
493 |
BLKDBG_REFTABLE_GROW, |
494 |
BLKDBG_REFTABLE_UPDATE, |
495 |
|
496 |
BLKDBG_REFBLOCK_LOAD, |
497 |
BLKDBG_REFBLOCK_UPDATE, |
498 |
BLKDBG_REFBLOCK_UPDATE_PART, |
499 |
BLKDBG_REFBLOCK_ALLOC, |
500 |
BLKDBG_REFBLOCK_ALLOC_HOOKUP, |
501 |
BLKDBG_REFBLOCK_ALLOC_WRITE, |
502 |
BLKDBG_REFBLOCK_ALLOC_WRITE_BLOCKS, |
503 |
BLKDBG_REFBLOCK_ALLOC_WRITE_TABLE, |
504 |
BLKDBG_REFBLOCK_ALLOC_SWITCH_TABLE, |
505 |
|
506 |
BLKDBG_CLUSTER_ALLOC, |
507 |
BLKDBG_CLUSTER_ALLOC_BYTES, |
508 |
BLKDBG_CLUSTER_FREE, |
509 |
|
510 |
BLKDBG_FLUSH_TO_OS, |
511 |
BLKDBG_FLUSH_TO_DISK, |
512 |
|
513 |
BLKDBG_EVENT_MAX, |
514 |
} BlkDebugEvent; |
515 |
|
516 |
#define BLKDBG_EVENT(bs, evt) bdrv_debug_event(bs, evt)
|
517 |
void bdrv_debug_event(BlockDriverState *bs, BlkDebugEvent event);
|
518 |
|
519 |
int bdrv_debug_breakpoint(BlockDriverState *bs, const char *event, |
520 |
const char *tag); |
521 |
int bdrv_debug_resume(BlockDriverState *bs, const char *tag); |
522 |
bool bdrv_debug_is_suspended(BlockDriverState *bs, const char *tag); |
523 |
|
524 |
#endif
|