diff options
author | Jean-Philippe Brucker <jean-philippe@linaro.org> | 2020-03-26 12:35:58 +0300 |
---|---|---|
committer | Joerg Roedel <jroedel@suse.de> | 2020-03-27 13:10:20 +0300 |
commit | 39b3b3c9cac1b4a64a6e546c9faf0c1011d775d4 (patch) | |
tree | e2790d0c80b65e2c208e83e7ecd7406f53083c05 /Documentation/memory-barriers.txt | |
parent | 7062af3ed2ba451029e3733d9f677c68f5ea9e77 (diff) | |
download | linux-39b3b3c9cac1b4a64a6e546c9faf0c1011d775d4.tar.xz |
iommu/virtio: Reject IOMMU page granule larger than PAGE_SIZE
We don't currently support IOMMUs with a page granule larger than the
system page size. The IOVA allocator has a BUG_ON() in this case, and
VFIO has a WARN_ON().
Removing these obstacles ranges doesn't seem possible without major
changes to the DMA API and VFIO. Some callers of iommu_map(), for
example, want to map multiple page-aligned regions adjacent to each
others for scatter-gather purposes. Even in simple DMA API uses, a call
to dma_map_page() would let the endpoint access neighbouring memory. And
VFIO users cannot ensure that their virtual address buffer is physically
contiguous at the IOMMU granule.
Rather than triggering the IOVA BUG_ON() on mismatched page sizes, abort
the vdomain finalise() with an error message. We could simply abort the
viommu probe(), but an upcoming extension to virtio-iommu will allow
setting different page masks for each endpoint.
Reported-by: Bharat Bhushan <bbhushan2@marvell.com>
Signed-off-by: Jean-Philippe Brucker <jean-philippe@linaro.org>
Reviewed-by: Bharat Bhushan <bbhushan2@marvell.com>
Reviewed-by: Eric Auger <eric.auger@redhat.com>
Reviewed-by: Robin Murphy <robin.murphy@arm.com>
Link: https://lore.kernel.org/r/20200326093558.2641019-4-jean-philippe@linaro.org
Signed-off-by: Joerg Roedel <jroedel@suse.de>
Diffstat (limited to 'Documentation/memory-barriers.txt')
0 files changed, 0 insertions, 0 deletions