summaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
authorKevin O'Connor <kevin@koconnor.net>2022-01-19 13:01:19 -0500
committerKevin O'Connor <kevin@koconnor.net>2022-01-27 11:25:36 -0500
commit0a40653f3013a31f3f7aeeddd7633328469a3ad3 (patch)
tree95c4267bf8c5758903e8bd4418623669b9a308e4 /src
parent9404f597b2a6aeb2e1d2b3866122ac52e5a87aeb (diff)
downloadqemu-seabios-0a40653f3013a31f3f7aeeddd7633328469a3ad3.tar.gz
nvme: Pass prp1 and prp2 directly to nvme_io_xfer()
When using a prp2 parameter, build it in nvme_prpl_xfer() and pass it directly to nvme_io_xfer(). Signed-off-by: Kevin O'Connor <kevin@koconnor.net> Reviewed-by: Alexander Graf <graf@amazon.com>
Diffstat (limited to 'src')
-rw-r--r--src/hw/nvme.c39
1 files changed, 18 insertions, 21 deletions
diff --git a/src/hw/nvme.c b/src/hw/nvme.c
index eee7d17..20976fc 100644
--- a/src/hw/nvme.c
+++ b/src/hw/nvme.c
@@ -416,33 +416,19 @@ err:
/* Reads count sectors into buf. The buffer cannot cross page boundaries. */
static int
-nvme_io_xfer(struct nvme_namespace *ns, u64 lba, void *buf, u16 count,
- int write)
+nvme_io_xfer(struct nvme_namespace *ns, u64 lba, void *prp1, void *prp2,
+ u16 count, int write)
{
- u32 buf_addr = (u32)buf;
- void *prp2;
-
- if (buf_addr & 0x3) {
+ if (((u32)prp1 & 0x3) || ((u32)prp2 & 0x3)) {
/* Buffer is misaligned */
warn_internalerror();
return -1;
}
- if ((ns->block_size * count) > (NVME_PAGE_SIZE * 2)) {
- /* We need to describe more than 2 pages, rely on PRP List */
- prp2 = ns->prpl;
- } else if ((ns->block_size * count) > NVME_PAGE_SIZE) {
- /* Directly embed the 2nd page if we only need 2 pages */
- prp2 = (void *)(long)ns->prpl[0];
- } else {
- /* One page is enough, don't expose anything else */
- prp2 = NULL;
- }
-
struct nvme_sqe *io_read = nvme_get_next_sqe(&ns->ctrl->io_sq,
write ? NVME_SQE_OPC_IO_WRITE
: NVME_SQE_OPC_IO_READ,
- NULL, buf, prp2);
+ NULL, prp1, prp2);
io_read->nsid = ns->ns_id;
io_read->dword[10] = (u32)lba;
io_read->dword[11] = (u32)(lba >> 32);
@@ -475,7 +461,7 @@ nvme_bounce_xfer(struct nvme_namespace *ns, u64 lba, void *buf, u16 count,
if (write)
memcpy(ns->dma_buffer, buf, blocks * ns->block_size);
- int res = nvme_io_xfer(ns, lba, ns->dma_buffer, blocks, write);
+ int res = nvme_io_xfer(ns, lba, ns->dma_buffer, NULL, blocks, write);
if (!write && res >= 0)
memcpy(buf, ns->dma_buffer, res * ns->block_size);
@@ -515,7 +501,7 @@ nvme_prpl_xfer(struct nvme_namespace *ns, u64 lba, void *buf, u16 count,
size = count * ns->block_size;
/* Special case for transfers that fit into PRP1, but are unaligned */
if (((size + (base & ~NVME_PAGE_MASK)) <= NVME_PAGE_SIZE))
- return nvme_io_xfer(ns, lba, buf, count, write);
+ return nvme_io_xfer(ns, lba, buf, NULL, count, write);
/* Every request has to be page aligned */
if (base & ~NVME_PAGE_MASK)
@@ -535,7 +521,18 @@ nvme_prpl_xfer(struct nvme_namespace *ns, u64 lba, void *buf, u16 count,
goto bounce;
}
- return nvme_io_xfer(ns, lba, buf, count, write);
+ void *prp2;
+ if ((ns->block_size * count) > (NVME_PAGE_SIZE * 2)) {
+ /* We need to describe more than 2 pages, rely on PRP List */
+ prp2 = ns->prpl;
+ } else if ((ns->block_size * count) > NVME_PAGE_SIZE) {
+ /* Directly embed the 2nd page if we only need 2 pages */
+ prp2 = (void *)(long)ns->prpl[0];
+ } else {
+ /* One page is enough, don't expose anything else */
+ prp2 = NULL;
+ }
+ return nvme_io_xfer(ns, lba, buf, prp2, count, write);
bounce:
/* Use bounce buffer to make transfer */