Revision 9a0be91c xseg/peers/user/cached.c
b/xseg/peers/user/cached.c | ||
---|---|---|
448 | 448 |
} |
449 | 449 |
|
450 | 450 |
/* |
451 |
* handle_read reads al buckets with a given request's range. If bucket is: |
|
451 |
* handle_read reads all buckets within a given request's range. |
|
452 |
* If a bucket is: |
|
452 | 453 |
* VALID || DIRTY || WRITING: it's good to read. |
453 | 454 |
* INVALID: we have to issue a request (via blocker) to read it from slower |
454 | 455 |
* media. |
... | ... | |
725 | 726 |
*/ |
726 | 727 |
struct cached *cached = __get_cached(peer); |
727 | 728 |
struct cache_io *cio = __get_cache_io(pr); |
729 |
/*assert there is a handler for received cio*/ |
|
728 | 730 |
struct ce *ce = get_cache_entry(cached->cache, cio->h); |
729 | 731 |
|
730 | 732 |
struct req_completion *rc; |
Also available in: Unified diff