@@ -26,6 +26,7 @@
#include "cpu.h"
#include "trace.h"
#include "qapi/error.h"
+#include "sysemu/iommufd.h"
/*
* Return one past the end of the end of section. Be careful with uint64_t
@@ -76,8 +77,9 @@ static bool vhost_vdpa_listener_skipped_section(MemoryRegionSection *section,
* The caller must set asid = 0 if the device does not support asid.
* This is not an ABI break since it is set to 0 by the initializer anyway.
*/
-int vhost_vdpa_dma_map(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
- hwaddr size, void *vaddr, bool readonly)
+
+int vhost_vdpa_leagcy_dma_map(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
+ hwaddr size, void *vaddr, bool readonly)
{
struct vhost_msg_v2 msg = {};
int fd = v->device_fd;
@@ -103,13 +105,32 @@ int vhost_vdpa_dma_map(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
return ret;
}
+int vhost_vdpa_dma_map(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
+ hwaddr size, void *vaddr, bool readonly)
+{
+ struct vhost_dev *dev = v->dev;
+
+ if ((v->enable_iommufd) && (v->ops == NULL)) {
+ vdpa_backend_iommufd_ops_class_init(v);
+ }
+
+ struct vdpa_iommu_backend_ops *ops = v->ops;
+ /* inoder to reuse the iotlb prcess to in kernel, still need to call leagcy
+ mode mapping but in kernel , the leagcy mode mapping was replace by
+ iommufd*/
+ if (v->enable_iommufd) {
+ ops->dma_map(dev, asid, iova, size, vaddr, readonly);
+ }
+ return vhost_vdpa_leagcy_dma_map(v, asid, iova, size, vaddr, readonly);
+}
/*
* The caller must set asid = 0 if the device does not support asid.
* This is not an ABI break since it is set to 0 by the initializer anyway.
*/
-int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
- hwaddr size)
+
+int vhost_vdpa_leagcy_dma_unmap(struct vhost_vdpa *v, uint32_t asid,
+ hwaddr iova, hwaddr size)
{
struct vhost_msg_v2 msg = {};
int fd = v->device_fd;
@@ -132,6 +153,26 @@ int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
return ret;
}
+int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
+ hwaddr size)
+{
+ struct vhost_dev *dev = v->dev;
+
+ if ((v->enable_iommufd) && (v->ops == NULL)) {
+ vdpa_backend_iommufd_ops_class_init(v);
+ }
+
+
+ /* inoder to reuse the iotlb prcess to in kernel, still need to call leagcy
+ mode mapping but in kernel , the leagcy mode mapping was replace by
+ iommufd*/
+ if (v->enable_iommufd) {
+ struct vdpa_iommu_backend_ops *ops = v->ops;
+
+ ops->dma_unmap(dev, asid, iova, size);
+ }
+ return vhost_vdpa_leagcy_dma_unmap(v, asid, iova, size);
+}
static void vhost_vdpa_listener_begin_batch(struct vhost_vdpa *v)
{
@@ -423,13 +464,14 @@ static void vhost_vdpa_init_svq(struct vhost_dev *hdev, struct vhost_vdpa *v)
v->shadow_vqs = g_steal_pointer(&shadow_vqs);
}
-
+int g_iommufd;
static int vhost_vdpa_init(struct vhost_dev *dev, void *opaque, Error **errp)
{
struct vhost_vdpa *v;
assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA);
trace_vhost_vdpa_init(dev, opaque);
int ret;
+ printf("[%s] %d called\n", __func__, __LINE__);
/*
* Similar to VFIO, we end up pinning all guest memory and have to
@@ -580,7 +622,9 @@ static int vhost_vdpa_cleanup(struct vhost_dev *dev)
vhost_vdpa_host_notifiers_uninit(dev, dev->nvqs);
memory_listener_unregister(&v->listener);
vhost_vdpa_svq_cleanup(dev);
-
+ if (vhost_vdpa_first_dev(dev)) {
+ v->ops->detach_device(v);
+ }
dev->opaque = NULL;
ram_block_discard_disable(false);
@@ -55,6 +55,10 @@ typedef struct vhost_vdpa {
void *shadow_vq_ops_opaque;
struct vhost_dev *dev;
VhostVDPAHostNotifier notifier[VIRTIO_QUEUE_MAX];
+ /*iommufd related*/
+ struct vdpa_iommu_backend_ops *ops;
+ bool enable_iommufd;
+
} VhostVDPA;
@@ -76,9 +80,29 @@ typedef struct vdpa_iommufd {
int vhost_vdpa_get_iova_range(int fd, struct vhost_vdpa_iova_range *iova_range);
+int vhost_vdpa_leagcy_dma_map(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
+ hwaddr size, void *vaddr, bool readonly);
+int vhost_vdpa_leagcy_dma_unmap(struct vhost_vdpa *v, uint32_t asid,
+ hwaddr iova, hwaddr size);
+
int vhost_vdpa_dma_map(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
hwaddr size, void *vaddr, bool readonly);
int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, uint32_t asid, hwaddr iova,
hwaddr size);
+struct vdpa_iommu_backend_ops {
+ /*< private >*/
+ ObjectClass parent_class;
+ int (*dma_map)(struct vhost_dev *dev, uint32_t asid, hwaddr iova,
+ hwaddr size, void *vaddr, bool readonly);
+ int (*dma_unmap)(struct vhost_dev *dev, uint32_t asid, hwaddr iova,
+ hwaddr size);
+ int (*attach_device)(struct vhost_vdpa *dev, AddressSpace *as,
+ Error **errp);
+ void (*detach_device)(struct vhost_vdpa *dev);
+ int (*reset)(VDPAIOMMUFDState *vdpa_iommufd);
+};
+
+void vdpa_backend_iommufd_ops_class_init(struct vhost_vdpa *v);
+
#endif
1.Change the map/umap function to legacy_map/unmap 2. Add the check for iommufd support, a>If support iommufd, call the iommufd-related function b>In order to use kernel's iotlb process. Still need to call the legacy mode iotlb message, Kernel will check and skip the legacy iotlb message if iommufd enable Signed-off-by: Cindy Lu <lulu@redhat.com> --- hw/virtio/vhost-vdpa.c | 56 ++++++++++++++++++++++++++++++---- include/hw/virtio/vhost-vdpa.h | 24 +++++++++++++++ 2 files changed, 74 insertions(+), 6 deletions(-)