nvme: switch delete SQ/CQ to blk_execute_rq_nowait
authorChristoph Hellwig <hch@lst.de>
Mon, 16 Nov 2015 09:28:47 +0000 (10:28 +0100)
committerJens Axboe <axboe@fb.com>
Tue, 22 Dec 2015 16:38:34 +0000 (09:38 -0700)
Signed-off-by: Christoph Hellwig <hch@lst.de>
Acked-by: Keith Busch <keith.busch@intel.com>
Signed-off-by: Jens Axboe <axboe@fb.com>
drivers/nvme/host/pci.c

index 808fb73556034269e1839fe05b86e3137ec540e7..d6d92b022f972f03de2e3a003090302f740e5e91 100644 (file)
@@ -86,8 +86,6 @@ static void nvme_dev_shutdown(struct nvme_dev *dev);
 struct async_cmd_info {
        struct kthread_work work;
        struct kthread_worker *worker;
-       struct request *req;
-       u32 result;
        int status;
        void *ctx;
 };
@@ -391,16 +389,6 @@ static void abort_completion(struct nvme_queue *nvmeq, void *ctx,
        atomic_inc(&nvmeq->dev->ctrl.abort_limit);
 }
 
-static void async_completion(struct nvme_queue *nvmeq, void *ctx,
-                                               struct nvme_completion *cqe)
-{
-       struct async_cmd_info *cmdinfo = ctx;
-       cmdinfo->result = le32_to_cpup(&cqe->result);
-       cmdinfo->status = le16_to_cpup(&cqe->status) >> 1;
-       queue_kthread_work(cmdinfo->worker, &cmdinfo->work);
-       blk_mq_free_request(cmdinfo->req);
-}
-
 static inline struct nvme_cmd_info *get_cmd_from_tag(struct nvme_queue *nvmeq,
                                  unsigned int tag)
 {
@@ -985,28 +973,13 @@ static int nvme_submit_async_admin_req(struct nvme_dev *dev)
        return 0;
 }
 
-static int nvme_submit_admin_async_cmd(struct nvme_dev *dev,
-                       struct nvme_command *cmd,
-                       struct async_cmd_info *cmdinfo, unsigned timeout)
+static void async_cmd_info_endio(struct request *req, int error)
 {
-       struct nvme_queue *nvmeq = dev->queues[0];
-       struct request *req;
-       struct nvme_cmd_info *cmd_rq;
-
-       req = blk_mq_alloc_request(dev->ctrl.admin_q, WRITE, 0);
-       if (IS_ERR(req))
-               return PTR_ERR(req);
+       struct async_cmd_info *cmdinfo = req->end_io_data;
 
-       req->timeout = timeout;
-       cmd_rq = blk_mq_rq_to_pdu(req);
-       cmdinfo->req = req;
-       nvme_set_info(cmd_rq, cmdinfo, async_completion);
-       cmdinfo->status = -EINTR;
-
-       cmd->common.command_id = req->tag;
-
-       nvme_submit_cmd(nvmeq, cmd);
-       return 0;
+       cmdinfo->status = req->errors;
+       queue_kthread_work(cmdinfo->worker, &cmdinfo->work);
+       blk_mq_free_request(req);
 }
 
 static int adapter_delete_queue(struct nvme_dev *dev, u8 opcode, u16 id)
@@ -1920,6 +1893,7 @@ static void nvme_del_queue_end(struct nvme_queue *nvmeq)
 static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode,
                                                kthread_work_func_t fn)
 {
+       struct request *req;
        struct nvme_command c;
 
        memset(&c, 0, sizeof(c));
@@ -1927,8 +1901,15 @@ static int adapter_async_del_queue(struct nvme_queue *nvmeq, u8 opcode,
        c.delete_queue.qid = cpu_to_le16(nvmeq->qid);
 
        init_kthread_work(&nvmeq->cmdinfo.work, fn);
-       return nvme_submit_admin_async_cmd(nvmeq->dev, &c, &nvmeq->cmdinfo,
-                                                               ADMIN_TIMEOUT);
+
+       req = nvme_alloc_request(nvmeq->dev->ctrl.admin_q, &c, 0);
+       if (IS_ERR(req))
+               return PTR_ERR(req);
+
+       req->timeout = ADMIN_TIMEOUT;
+       req->end_io_data = &nvmeq->cmdinfo;
+       blk_execute_rq_nowait(req->q, NULL, req, 0, async_cmd_info_endio);
+       return 0;
 }
 
 static void nvme_del_cq_work_handler(struct kthread_work *work)