struct hlist_head *tl_hash;
unsigned int tl_hash_s;
- /* Interval tree of pending local write requests */
+ /* Interval tree of pending local requests */
+ struct rb_root read_requests;
struct rb_root write_requests;
/* blocks to resync in this run [unit BM_BLOCK_SIZE] */
/* no need to lock access, we are still initializing this minor device. */
if (!tl_init(mdev))
goto out_no_tl;
+ mdev->read_requests = RB_ROOT;
mdev->write_requests = RB_ROOT;
mdev->app_reads_hash = kzalloc(APP_R_HSIZE*sizeof(void *), GFP_KERNEL);
/* remove the request from the conflict detection
* respective block_id verification hash */
- if (!hlist_unhashed(&req->collision)) {
+ if (!drbd_interval_empty(&req->i)) {
+ struct rb_root *root;
+
hlist_del(&req->collision);
- if (!drbd_interval_empty(&req->i))
- drbd_remove_interval(&mdev->write_requests, &req->i);
+ if (rw == WRITE)
+ root = &mdev->write_requests;
+ else
+ root = &mdev->read_requests;
+ drbd_remove_interval(root, &req->i);
} else
D_ASSERT((s & (RQ_NET_MASK & ~RQ_NET_DONE)) == 0);
struct hlist_head *slot;
D_ASSERT(hlist_unhashed(&req->collision));
+ D_ASSERT(drbd_interval_empty(&req->i));
if (!get_net_conf(mdev))
return 0;
/* so we can verify the handle in the answer packet
* corresponding hlist_del is in _req_may_be_done() */
hlist_add_head(&req->collision, ar_hash_slot(mdev, req->i.sector));
+ drbd_insert_interval(&mdev->read_requests, &req->i);
set_bit(UNPLUG_REMOTE, &mdev->flags);