summaryrefslogtreecommitdiffstats
path: root/drivers/block/nvme.c
diff options
context:
space:
mode:
authorMatthew Wilcox <matthew.r.wilcox@intel.com>2011-12-20 17:04:12 +0100
committerMatthew Wilcox <matthew.r.wilcox@intel.com>2012-01-10 20:49:18 +0100
commit040a93b52a9eee8177ebaf2ba0ee0f9f518d1bf8 (patch)
treebe1f319048580f40c2634bca46e2840d721a32f4 /drivers/block/nvme.c
parentNVMe: Simplify completion handling (diff)
downloadlinux-040a93b52a9eee8177ebaf2ba0ee0f9f518d1bf8.tar.xz
linux-040a93b52a9eee8177ebaf2ba0ee0f9f518d1bf8.zip
NVMe: Change get_nvmeq to take a dev instead of a namespace
Upcoming patches require calling get_nvmeq when we don't have a namespace. Some callers already have the device in a local variable anyway. Signed-off-by: Matthew Wilcox <matthew.r.wilcox@intel.com>
Diffstat (limited to 'drivers/block/nvme.c')
-rw-r--r--drivers/block/nvme.c8
1 files changed, 4 insertions, 4 deletions
diff --git a/drivers/block/nvme.c b/drivers/block/nvme.c
index 4724655a6ebf..aa2fd66aabd6 100644
--- a/drivers/block/nvme.c
+++ b/drivers/block/nvme.c
@@ -257,9 +257,9 @@ static void *cancel_cmdid(struct nvme_queue *nvmeq, int cmdid,
return ctx;
}
-static struct nvme_queue *get_nvmeq(struct nvme_ns *ns)
+static struct nvme_queue *get_nvmeq(struct nvme_dev *dev)
{
- return ns->dev->queues[get_cpu() + 1];
+ return dev->queues[get_cpu() + 1];
}
static void put_nvmeq(struct nvme_queue *nvmeq)
@@ -606,7 +606,7 @@ static int nvme_submit_bio_queue(struct nvme_queue *nvmeq, struct nvme_ns *ns,
static int nvme_make_request(struct request_queue *q, struct bio *bio)
{
struct nvme_ns *ns = q->queuedata;
- struct nvme_queue *nvmeq = get_nvmeq(ns);
+ struct nvme_queue *nvmeq = get_nvmeq(ns->dev);
int result = -EBUSY;
spin_lock_irq(&nvmeq->q_lock);
@@ -1103,7 +1103,7 @@ static int nvme_submit_io(struct nvme_ns *ns, struct nvme_user_io __user *uio)
/* XXX: metadata */
prps = nvme_setup_prps(dev, &c.common, sg, &length, GFP_KERNEL);
- nvmeq = get_nvmeq(ns);
+ nvmeq = get_nvmeq(dev);
/*
* Since nvme_submit_sync_cmd sleeps, we can't keep preemption
* disabled. We may be preempted at any point, and be rescheduled