request_end(fc, req);
goto restart;
}
+ spin_lock(&fpq->lock);
list_add(&req->list, &fpq->io);
+ spin_unlock(&fpq->lock);
spin_unlock(&fc->lock);
cs->req = req;
err = fuse_copy_one(cs, &in->h, sizeof(in->h));
(struct fuse_arg *) in->args, 0);
fuse_copy_finish(cs);
spin_lock(&fc->lock);
+ spin_lock(&fpq->lock);
clear_bit(FR_LOCKED, &req->flags);
if (!fpq->connected) {
err = -ENODEV;
goto out_end;
}
list_move_tail(&req->list, &fpq->processing);
+ spin_unlock(&fpq->lock);
set_bit(FR_SENT, &req->flags);
/* matches barrier in request_wait_answer() */
smp_mb__after_atomic();
out_end:
list_del_init(&req->list);
+ spin_unlock(&fpq->lock);
request_end(fc, req);
return err;
goto err_finish;
spin_lock(&fc->lock);
+ spin_lock(&fpq->lock);
err = -ENOENT;
if (!fpq->connected)
- goto err_unlock;
+ goto err_unlock_pq;
req = request_find(fpq, oh.unique);
if (!req)
- goto err_unlock;
+ goto err_unlock_pq;
/* Is it an interrupt reply? */
if (req->intr_unique == oh.unique) {
+ spin_unlock(&fpq->lock);
+
err = -EINVAL;
if (nbytes != sizeof(struct fuse_out_header))
goto err_unlock;
list_move(&req->list, &fpq->io);
req->out.h = oh;
set_bit(FR_LOCKED, &req->flags);
+ spin_unlock(&fpq->lock);
cs->req = req;
if (!req->out.page_replace)
cs->move_pages = 0;
fuse_copy_finish(cs);
spin_lock(&fc->lock);
+ spin_lock(&fpq->lock);
clear_bit(FR_LOCKED, &req->flags);
if (!fpq->connected)
err = -ENOENT;
else if (err)
req->out.h.error = -EIO;
list_del_init(&req->list);
+ spin_unlock(&fpq->lock);
request_end(fc, req);
return err ? err : nbytes;
+ err_unlock_pq:
+ spin_unlock(&fpq->lock);
err_unlock:
spin_unlock(&fc->lock);
err_finish:
fc->connected = 0;
fc->blocked = 0;
fuse_set_initialized(fc);
+ spin_lock(&fpq->lock);
fpq->connected = 0;
list_for_each_entry_safe(req, next, &fpq->io, list) {
req->out.h.error = -ECONNABORTED;
spin_unlock(&req->waitq.lock);
}
list_splice_init(&fpq->processing, &to_end2);
+ spin_unlock(&fpq->lock);
fc->max_background = UINT_MAX;
flush_bg_queue(fc);