Message ID | 20231102071302.1818071-13-zhenzhong.duan@intel.com |
---|---|
State | New |
Headers | show |
Series | vfio: Adopt iommufd | expand |
On 11/2/23 08:12, Zhenzhong Duan wrote: > From: Eric Auger <eric.auger@redhat.com> > > dirty_pages_supported field is also moved to the base container > > No fucntional change intended. > > Signed-off-by: Eric Auger <eric.auger@redhat.com> > Signed-off-by: Yi Liu <yi.l.liu@intel.com> > Signed-off-by: Yi Sun <yi.y.sun@linux.intel.com> > Signed-off-by: Zhenzhong Duan <zhenzhong.duan@intel.com> > --- > v4: use assert Reviewed-by: Cédric Le Goater <clg@redhat.com> Thanks, C. > > include/hw/vfio/vfio-common.h | 6 ------ > include/hw/vfio/vfio-container-base.h | 6 ++++++ > hw/vfio/common.c | 12 ++++++++---- > hw/vfio/container-base.c | 16 ++++++++++++++++ > hw/vfio/container.c | 21 ++++++++++++++------- > 5 files changed, 44 insertions(+), 17 deletions(-) > > diff --git a/include/hw/vfio/vfio-common.h b/include/hw/vfio/vfio-common.h > index bd4de6cb3a..60f2785fe0 100644 > --- a/include/hw/vfio/vfio-common.h > +++ b/include/hw/vfio/vfio-common.h > @@ -83,7 +83,6 @@ typedef struct VFIOContainer { > unsigned iommu_type; > Error *error; > bool initialized; > - bool dirty_pages_supported; > uint64_t dirty_pgsizes; > uint64_t max_dirty_bitmap_size; > unsigned long pgsizes; > @@ -190,11 +189,6 @@ VFIOAddressSpace *vfio_get_address_space(AddressSpace *as); > void vfio_put_address_space(VFIOAddressSpace *space); > bool vfio_devices_all_running_and_saving(VFIOContainer *container); > > -/* container->fd */ > -int vfio_set_dirty_page_tracking(VFIOContainer *container, bool start); > -int vfio_query_dirty_bitmap(VFIOContainer *container, VFIOBitmap *vbmap, > - hwaddr iova, hwaddr size); > - > /* SPAPR specific */ > int vfio_container_add_section_window(VFIOContainer *container, > MemoryRegionSection *section, > diff --git a/include/hw/vfio/vfio-container-base.h b/include/hw/vfio/vfio-container-base.h > index c7cc6ec9c5..f244f003d0 100644 > --- a/include/hw/vfio/vfio-container-base.h > +++ b/include/hw/vfio/vfio-container-base.h > @@ -36,6 +36,7 @@ typedef struct VFIOAddressSpace { > typedef struct VFIOContainerBase { > const VFIOIOMMUOps *ops; > VFIOAddressSpace *space; > + bool dirty_pages_supported; > QLIST_HEAD(, VFIOGuestIOMMU) giommu_list; > QLIST_ENTRY(VFIOContainerBase) next; > } VFIOContainerBase; > @@ -54,6 +55,11 @@ int vfio_container_dma_map(VFIOContainerBase *bcontainer, > int vfio_container_dma_unmap(VFIOContainerBase *bcontainer, > hwaddr iova, ram_addr_t size, > IOMMUTLBEntry *iotlb); > +int vfio_container_set_dirty_page_tracking(VFIOContainerBase *bcontainer, > + bool start); > +int vfio_container_query_dirty_bitmap(VFIOContainerBase *bcontainer, > + VFIOBitmap *vbmap, > + hwaddr iova, hwaddr size); > > void vfio_container_init(VFIOContainerBase *bcontainer, > VFIOAddressSpace *space, > diff --git a/hw/vfio/common.c b/hw/vfio/common.c > index 1d8202537e..b1a875ca93 100644 > --- a/hw/vfio/common.c > +++ b/hw/vfio/common.c > @@ -1079,7 +1079,8 @@ static void vfio_listener_log_global_start(MemoryListener *listener) > if (vfio_devices_all_device_dirty_tracking(container)) { > ret = vfio_devices_dma_logging_start(container); > } else { > - ret = vfio_set_dirty_page_tracking(container, true); > + ret = vfio_container_set_dirty_page_tracking(&container->bcontainer, > + true); > } > > if (ret) { > @@ -1097,7 +1098,8 @@ static void vfio_listener_log_global_stop(MemoryListener *listener) > if (vfio_devices_all_device_dirty_tracking(container)) { > vfio_devices_dma_logging_stop(container); > } else { > - ret = vfio_set_dirty_page_tracking(container, false); > + ret = vfio_container_set_dirty_page_tracking(&container->bcontainer, > + false); > } > > if (ret) { > @@ -1165,7 +1167,8 @@ int vfio_get_dirty_bitmap(VFIOContainer *container, uint64_t iova, > VFIOBitmap vbmap; > int ret; > > - if (!container->dirty_pages_supported && !all_device_dirty_tracking) { > + if (!container->bcontainer.dirty_pages_supported && > + !all_device_dirty_tracking) { > cpu_physical_memory_set_dirty_range(ram_addr, size, > tcg_enabled() ? DIRTY_CLIENTS_ALL : > DIRTY_CLIENTS_NOCODE); > @@ -1180,7 +1183,8 @@ int vfio_get_dirty_bitmap(VFIOContainer *container, uint64_t iova, > if (all_device_dirty_tracking) { > ret = vfio_devices_query_dirty_bitmap(container, &vbmap, iova, size); > } else { > - ret = vfio_query_dirty_bitmap(container, &vbmap, iova, size); > + ret = vfio_container_query_dirty_bitmap(&container->bcontainer, &vbmap, > + iova, size); > } > > if (ret) { > diff --git a/hw/vfio/container-base.c b/hw/vfio/container-base.c > index 3933391e0d..5d654ae172 100644 > --- a/hw/vfio/container-base.c > +++ b/hw/vfio/container-base.c > @@ -31,11 +31,27 @@ int vfio_container_dma_unmap(VFIOContainerBase *bcontainer, > return bcontainer->ops->dma_unmap(bcontainer, iova, size, iotlb); > } > > +int vfio_container_set_dirty_page_tracking(VFIOContainerBase *bcontainer, > + bool start) > +{ > + g_assert(bcontainer->ops->set_dirty_page_tracking); > + return bcontainer->ops->set_dirty_page_tracking(bcontainer, start); > +} > + > +int vfio_container_query_dirty_bitmap(VFIOContainerBase *bcontainer, > + VFIOBitmap *vbmap, > + hwaddr iova, hwaddr size) > +{ > + g_assert(bcontainer->ops->query_dirty_bitmap); > + return bcontainer->ops->query_dirty_bitmap(bcontainer, vbmap, iova, size); > +} > + > void vfio_container_init(VFIOContainerBase *bcontainer, VFIOAddressSpace *space, > const VFIOIOMMUOps *ops) > { > bcontainer->ops = ops; > bcontainer->space = space; > + bcontainer->dirty_pages_supported = false; > QLIST_INIT(&bcontainer->giommu_list); > } > > diff --git a/hw/vfio/container.c b/hw/vfio/container.c > index f12fcb6fe1..3ab74e2615 100644 > --- a/hw/vfio/container.c > +++ b/hw/vfio/container.c > @@ -131,7 +131,7 @@ static int vfio_legacy_dma_unmap(VFIOContainerBase *bcontainer, hwaddr iova, > > if (iotlb && vfio_devices_all_running_and_mig_active(container)) { > if (!vfio_devices_all_device_dirty_tracking(container) && > - container->dirty_pages_supported) { > + container->bcontainer.dirty_pages_supported) { > return vfio_dma_unmap_bitmap(container, iova, size, iotlb); > } > > @@ -205,14 +205,17 @@ static int vfio_legacy_dma_map(VFIOContainerBase *bcontainer, hwaddr iova, > return -errno; > } > > -int vfio_set_dirty_page_tracking(VFIOContainer *container, bool start) > +static int vfio_legacy_set_dirty_page_tracking(VFIOContainerBase *bcontainer, > + bool start) > { > + VFIOContainer *container = container_of(bcontainer, VFIOContainer, > + bcontainer); > int ret; > struct vfio_iommu_type1_dirty_bitmap dirty = { > .argsz = sizeof(dirty), > }; > > - if (!container->dirty_pages_supported) { > + if (!bcontainer->dirty_pages_supported) { > return 0; > } > > @@ -232,9 +235,12 @@ int vfio_set_dirty_page_tracking(VFIOContainer *container, bool start) > return ret; > } > > -int vfio_query_dirty_bitmap(VFIOContainer *container, VFIOBitmap *vbmap, > - hwaddr iova, hwaddr size) > +static int vfio_legacy_query_dirty_bitmap(VFIOContainerBase *bcontainer, > + VFIOBitmap *vbmap, > + hwaddr iova, hwaddr size) > { > + VFIOContainer *container = container_of(bcontainer, VFIOContainer, > + bcontainer); > struct vfio_iommu_type1_dirty_bitmap *dbitmap; > struct vfio_iommu_type1_dirty_bitmap_get *range; > int ret; > @@ -461,7 +467,7 @@ static void vfio_get_iommu_info_migration(VFIOContainer *container, > * qemu_real_host_page_size to mark those dirty. > */ > if (cap_mig->pgsize_bitmap & qemu_real_host_page_size()) { > - container->dirty_pages_supported = true; > + container->bcontainer.dirty_pages_supported = true; > container->max_dirty_bitmap_size = cap_mig->max_dirty_bitmap_size; > container->dirty_pgsizes = cap_mig->pgsize_bitmap; > } > @@ -553,7 +559,6 @@ static int vfio_connect_container(VFIOGroup *group, AddressSpace *as, > container = g_malloc0(sizeof(*container)); > container->fd = fd; > container->error = NULL; > - container->dirty_pages_supported = false; > container->dma_max_mappings = 0; > container->iova_ranges = NULL; > QLIST_INIT(&container->vrdl_list); > @@ -937,4 +942,6 @@ void vfio_detach_device(VFIODevice *vbasedev) > const VFIOIOMMUOps vfio_legacy_ops = { > .dma_map = vfio_legacy_dma_map, > .dma_unmap = vfio_legacy_dma_unmap, > + .set_dirty_page_tracking = vfio_legacy_set_dirty_page_tracking, > + .query_dirty_bitmap = vfio_legacy_query_dirty_bitmap, > };
diff --git a/include/hw/vfio/vfio-common.h b/include/hw/vfio/vfio-common.h index bd4de6cb3a..60f2785fe0 100644 --- a/include/hw/vfio/vfio-common.h +++ b/include/hw/vfio/vfio-common.h @@ -83,7 +83,6 @@ typedef struct VFIOContainer { unsigned iommu_type; Error *error; bool initialized; - bool dirty_pages_supported; uint64_t dirty_pgsizes; uint64_t max_dirty_bitmap_size; unsigned long pgsizes; @@ -190,11 +189,6 @@ VFIOAddressSpace *vfio_get_address_space(AddressSpace *as); void vfio_put_address_space(VFIOAddressSpace *space); bool vfio_devices_all_running_and_saving(VFIOContainer *container); -/* container->fd */ -int vfio_set_dirty_page_tracking(VFIOContainer *container, bool start); -int vfio_query_dirty_bitmap(VFIOContainer *container, VFIOBitmap *vbmap, - hwaddr iova, hwaddr size); - /* SPAPR specific */ int vfio_container_add_section_window(VFIOContainer *container, MemoryRegionSection *section, diff --git a/include/hw/vfio/vfio-container-base.h b/include/hw/vfio/vfio-container-base.h index c7cc6ec9c5..f244f003d0 100644 --- a/include/hw/vfio/vfio-container-base.h +++ b/include/hw/vfio/vfio-container-base.h @@ -36,6 +36,7 @@ typedef struct VFIOAddressSpace { typedef struct VFIOContainerBase { const VFIOIOMMUOps *ops; VFIOAddressSpace *space; + bool dirty_pages_supported; QLIST_HEAD(, VFIOGuestIOMMU) giommu_list; QLIST_ENTRY(VFIOContainerBase) next; } VFIOContainerBase; @@ -54,6 +55,11 @@ int vfio_container_dma_map(VFIOContainerBase *bcontainer, int vfio_container_dma_unmap(VFIOContainerBase *bcontainer, hwaddr iova, ram_addr_t size, IOMMUTLBEntry *iotlb); +int vfio_container_set_dirty_page_tracking(VFIOContainerBase *bcontainer, + bool start); +int vfio_container_query_dirty_bitmap(VFIOContainerBase *bcontainer, + VFIOBitmap *vbmap, + hwaddr iova, hwaddr size); void vfio_container_init(VFIOContainerBase *bcontainer, VFIOAddressSpace *space, diff --git a/hw/vfio/common.c b/hw/vfio/common.c index 1d8202537e..b1a875ca93 100644 --- a/hw/vfio/common.c +++ b/hw/vfio/common.c @@ -1079,7 +1079,8 @@ static void vfio_listener_log_global_start(MemoryListener *listener) if (vfio_devices_all_device_dirty_tracking(container)) { ret = vfio_devices_dma_logging_start(container); } else { - ret = vfio_set_dirty_page_tracking(container, true); + ret = vfio_container_set_dirty_page_tracking(&container->bcontainer, + true); } if (ret) { @@ -1097,7 +1098,8 @@ static void vfio_listener_log_global_stop(MemoryListener *listener) if (vfio_devices_all_device_dirty_tracking(container)) { vfio_devices_dma_logging_stop(container); } else { - ret = vfio_set_dirty_page_tracking(container, false); + ret = vfio_container_set_dirty_page_tracking(&container->bcontainer, + false); } if (ret) { @@ -1165,7 +1167,8 @@ int vfio_get_dirty_bitmap(VFIOContainer *container, uint64_t iova, VFIOBitmap vbmap; int ret; - if (!container->dirty_pages_supported && !all_device_dirty_tracking) { + if (!container->bcontainer.dirty_pages_supported && + !all_device_dirty_tracking) { cpu_physical_memory_set_dirty_range(ram_addr, size, tcg_enabled() ? DIRTY_CLIENTS_ALL : DIRTY_CLIENTS_NOCODE); @@ -1180,7 +1183,8 @@ int vfio_get_dirty_bitmap(VFIOContainer *container, uint64_t iova, if (all_device_dirty_tracking) { ret = vfio_devices_query_dirty_bitmap(container, &vbmap, iova, size); } else { - ret = vfio_query_dirty_bitmap(container, &vbmap, iova, size); + ret = vfio_container_query_dirty_bitmap(&container->bcontainer, &vbmap, + iova, size); } if (ret) { diff --git a/hw/vfio/container-base.c b/hw/vfio/container-base.c index 3933391e0d..5d654ae172 100644 --- a/hw/vfio/container-base.c +++ b/hw/vfio/container-base.c @@ -31,11 +31,27 @@ int vfio_container_dma_unmap(VFIOContainerBase *bcontainer, return bcontainer->ops->dma_unmap(bcontainer, iova, size, iotlb); } +int vfio_container_set_dirty_page_tracking(VFIOContainerBase *bcontainer, + bool start) +{ + g_assert(bcontainer->ops->set_dirty_page_tracking); + return bcontainer->ops->set_dirty_page_tracking(bcontainer, start); +} + +int vfio_container_query_dirty_bitmap(VFIOContainerBase *bcontainer, + VFIOBitmap *vbmap, + hwaddr iova, hwaddr size) +{ + g_assert(bcontainer->ops->query_dirty_bitmap); + return bcontainer->ops->query_dirty_bitmap(bcontainer, vbmap, iova, size); +} + void vfio_container_init(VFIOContainerBase *bcontainer, VFIOAddressSpace *space, const VFIOIOMMUOps *ops) { bcontainer->ops = ops; bcontainer->space = space; + bcontainer->dirty_pages_supported = false; QLIST_INIT(&bcontainer->giommu_list); } diff --git a/hw/vfio/container.c b/hw/vfio/container.c index f12fcb6fe1..3ab74e2615 100644 --- a/hw/vfio/container.c +++ b/hw/vfio/container.c @@ -131,7 +131,7 @@ static int vfio_legacy_dma_unmap(VFIOContainerBase *bcontainer, hwaddr iova, if (iotlb && vfio_devices_all_running_and_mig_active(container)) { if (!vfio_devices_all_device_dirty_tracking(container) && - container->dirty_pages_supported) { + container->bcontainer.dirty_pages_supported) { return vfio_dma_unmap_bitmap(container, iova, size, iotlb); } @@ -205,14 +205,17 @@ static int vfio_legacy_dma_map(VFIOContainerBase *bcontainer, hwaddr iova, return -errno; } -int vfio_set_dirty_page_tracking(VFIOContainer *container, bool start) +static int vfio_legacy_set_dirty_page_tracking(VFIOContainerBase *bcontainer, + bool start) { + VFIOContainer *container = container_of(bcontainer, VFIOContainer, + bcontainer); int ret; struct vfio_iommu_type1_dirty_bitmap dirty = { .argsz = sizeof(dirty), }; - if (!container->dirty_pages_supported) { + if (!bcontainer->dirty_pages_supported) { return 0; } @@ -232,9 +235,12 @@ int vfio_set_dirty_page_tracking(VFIOContainer *container, bool start) return ret; } -int vfio_query_dirty_bitmap(VFIOContainer *container, VFIOBitmap *vbmap, - hwaddr iova, hwaddr size) +static int vfio_legacy_query_dirty_bitmap(VFIOContainerBase *bcontainer, + VFIOBitmap *vbmap, + hwaddr iova, hwaddr size) { + VFIOContainer *container = container_of(bcontainer, VFIOContainer, + bcontainer); struct vfio_iommu_type1_dirty_bitmap *dbitmap; struct vfio_iommu_type1_dirty_bitmap_get *range; int ret; @@ -461,7 +467,7 @@ static void vfio_get_iommu_info_migration(VFIOContainer *container, * qemu_real_host_page_size to mark those dirty. */ if (cap_mig->pgsize_bitmap & qemu_real_host_page_size()) { - container->dirty_pages_supported = true; + container->bcontainer.dirty_pages_supported = true; container->max_dirty_bitmap_size = cap_mig->max_dirty_bitmap_size; container->dirty_pgsizes = cap_mig->pgsize_bitmap; } @@ -553,7 +559,6 @@ static int vfio_connect_container(VFIOGroup *group, AddressSpace *as, container = g_malloc0(sizeof(*container)); container->fd = fd; container->error = NULL; - container->dirty_pages_supported = false; container->dma_max_mappings = 0; container->iova_ranges = NULL; QLIST_INIT(&container->vrdl_list); @@ -937,4 +942,6 @@ void vfio_detach_device(VFIODevice *vbasedev) const VFIOIOMMUOps vfio_legacy_ops = { .dma_map = vfio_legacy_dma_map, .dma_unmap = vfio_legacy_dma_unmap, + .set_dirty_page_tracking = vfio_legacy_set_dirty_page_tracking, + .query_dirty_bitmap = vfio_legacy_query_dirty_bitmap, };