summaryrefslogtreecommitdiffstats
path: root/drivers/block
diff options
context:
space:
mode:
authorMatthew Wilcox <matthew.r.wilcox@intel.com>2011-03-21 09:48:57 -0400
committerMatthew Wilcox <matthew.r.wilcox@intel.com>2011-11-04 15:53:01 -0400
commit6c7d49455ceb63064f992347d9185ff5bf43497a (patch)
tree8a5fde81a68cd41c12079d8c2baf7ea1cf3e40fe /drivers/block
parent9d4af1b7796ba02b73a79a8694399e5a3cd1c55d (diff)
NVMe: Change the definition of nvme_user_io
The read and write commands don't define a 'result', so there's no need to copy it back to userspace. Remove the ability of the ioctl to submit commands to a different namespace; it's just asking for trouble, and the use case I have in mind will be addressed througha different ioctl in the future. That removes the need for both the block_shift and nsid arguments. Check that the opcode is one of 'read' or 'write'. Future opcodes may be added in the future, but we will need a different structure definition for them. The nblocks field is redefined to be 0-based. This allows the user to request the full 65536 blocks. Don't byteswap the reftag, apptag and appmask. Martin Petersen tells me these are calculated in big-endian and are transmitted to the device in big-endian. Signed-off-by: Matthew Wilcox <matthew.r.wilcox@intel.com>
Diffstat (limited to 'drivers/block')
-rw-r--r--drivers/block/nvme.c27
1 files changed, 17 insertions, 10 deletions
diff --git a/drivers/block/nvme.c b/drivers/block/nvme.c
index d0b52622e26..90a96ec8a59 100644
--- a/drivers/block/nvme.c
+++ b/drivers/block/nvme.c
@@ -1035,29 +1035,37 @@ static int nvme_submit_io(struct nvme_ns *ns, struct nvme_user_io __user *uio)
struct nvme_user_io io;
struct nvme_command c;
unsigned length;
- u32 result;
int nents, status;
struct scatterlist *sg;
struct nvme_prps *prps;
if (copy_from_user(&io, uio, sizeof(io)))
return -EFAULT;
- length = io.nblocks << io.block_shift;
- nents = nvme_map_user_pages(dev, io.opcode & 1, io.addr, length, &sg);
+ length = (io.nblocks + 1) << ns->lba_shift;
+
+ switch (io.opcode) {
+ case nvme_cmd_write:
+ case nvme_cmd_read:
+ nents = nvme_map_user_pages(dev, io.opcode & 1, io.addr,
+ length, &sg);
+ default:
+ return -EFAULT;
+ }
+
if (nents < 0)
return nents;
memset(&c, 0, sizeof(c));
c.rw.opcode = io.opcode;
c.rw.flags = io.flags;
- c.rw.nsid = cpu_to_le32(io.nsid);
+ c.rw.nsid = cpu_to_le32(ns->ns_id);
c.rw.slba = cpu_to_le64(io.slba);
- c.rw.length = cpu_to_le16(io.nblocks - 1);
+ c.rw.length = cpu_to_le16(io.nblocks);
c.rw.control = cpu_to_le16(io.control);
c.rw.dsmgmt = cpu_to_le16(io.dsmgmt);
- c.rw.reftag = cpu_to_le32(io.reftag); /* XXX: endian? */
- c.rw.apptag = cpu_to_le16(io.apptag);
- c.rw.appmask = cpu_to_le16(io.appmask);
+ c.rw.reftag = io.reftag;
+ c.rw.apptag = io.apptag;
+ c.rw.appmask = io.appmask;
/* XXX: metadata */
prps = nvme_setup_prps(dev, &c.common, sg, length);
@@ -1069,11 +1077,10 @@ static int nvme_submit_io(struct nvme_ns *ns, struct nvme_user_io __user *uio)
* additional races since q_lock already protects against other CPUs.
*/
put_nvmeq(nvmeq);
- status = nvme_submit_sync_cmd(nvmeq, &c, &result, IO_TIMEOUT);
+ status = nvme_submit_sync_cmd(nvmeq, &c, NULL, IO_TIMEOUT);
nvme_unmap_user_pages(dev, io.opcode & 1, io.addr, length, sg, nents);
nvme_free_prps(dev, prps);
- put_user(result, &uio->result);
return status;
}