ram_addr_t size, void *vaddr, bool readonly,
MemoryRegion *mr)
{
- const VFIOIOMMUFDContainer *container =
- container_of(bcontainer, VFIOIOMMUFDContainer, bcontainer);
+ const VFIOIOMMUFDContainer *container = VFIO_IOMMU_IOMMUFD(bcontainer);
return iommufd_backend_map_dma(container->be,
container->ioas_id,
hwaddr iova, ram_addr_t size,
int fd, unsigned long start, bool readonly)
{
- const VFIOIOMMUFDContainer *container =
- container_of(bcontainer, VFIOIOMMUFDContainer, bcontainer);
+ const VFIOIOMMUFDContainer *container = VFIO_IOMMU_IOMMUFD(bcontainer);
return iommufd_backend_map_file_dma(container->be,
container->ioas_id,
hwaddr iova, ram_addr_t size,
IOMMUTLBEntry *iotlb, bool unmap_all)
{
- const VFIOIOMMUFDContainer *container =
- container_of(bcontainer, VFIOIOMMUFDContainer, bcontainer);
+ const VFIOIOMMUFDContainer *container = VFIO_IOMMU_IOMMUFD(bcontainer);
/* unmap in halves */
if (unmap_all) {
static int iommufd_set_dirty_page_tracking(const VFIOContainer *bcontainer,
bool start, Error **errp)
{
- const VFIOIOMMUFDContainer *container =
- container_of(bcontainer, VFIOIOMMUFDContainer, bcontainer);
+ const VFIOIOMMUFDContainer *container = VFIO_IOMMU_IOMMUFD(bcontainer);
VFIOIOASHwpt *hwpt;
QLIST_FOREACH(hwpt, &container->hwpt_list, next) {
VFIOBitmap *vbmap, hwaddr iova,
hwaddr size, Error **errp)
{
- VFIOIOMMUFDContainer *container = container_of(bcontainer,
- VFIOIOMMUFDContainer,
- bcontainer);
+ VFIOIOMMUFDContainer *container = VFIO_IOMMU_IOMMUFD(bcontainer);
unsigned long page_size = qemu_real_host_page_size();
VFIOIOASHwpt *hwpt;
{
ERRP_GUARD();
IOMMUFDBackend *iommufd = vbasedev->iommufd;
+ VFIOContainer *bcontainer = VFIO_IOMMU(container);
uint32_t type, flags = 0;
uint64_t hw_caps;
VFIOIOASHwpt *hwpt;
vbasedev->iommu_dirty_tracking = iommufd_hwpt_dirty_tracking(hwpt);
QLIST_INSERT_HEAD(&hwpt->device_list, vbasedev, hwpt_next);
QLIST_INSERT_HEAD(&container->hwpt_list, hwpt, next);
- container->bcontainer.dirty_pages_supported |=
+ bcontainer->dirty_pages_supported |=
vbasedev->iommu_dirty_tracking;
- if (container->bcontainer.dirty_pages_supported &&
+ if (bcontainer->dirty_pages_supported &&
!vbasedev->iommu_dirty_tracking) {
warn_report("IOMMU instance for device %s doesn't support dirty tracking",
vbasedev->name);
static void iommufd_cdev_container_destroy(VFIOIOMMUFDContainer *container)
{
- VFIOContainer *bcontainer = &container->bcontainer;
+ VFIOContainer *bcontainer = VFIO_IOMMU(container);
if (!QLIST_EMPTY(&bcontainer->device_list)) {
return;
static bool iommufd_cdev_get_info_iova_range(VFIOIOMMUFDContainer *container,
uint32_t ioas_id, Error **errp)
{
- VFIOContainer *bcontainer = &container->bcontainer;
+ VFIOContainer *bcontainer = VFIO_IOMMU(container);
g_autofree struct iommu_ioas_iova_ranges *info = NULL;
struct iommu_iova_range *iova_ranges;
int sz, fd = container->be->fd;
/* try to attach to an existing container in this space */
QLIST_FOREACH(bcontainer, &space->containers, next) {
- container = container_of(bcontainer, VFIOIOMMUFDContainer, bcontainer);
+ container = VFIO_IOMMU_IOMMUFD(bcontainer);
if (VFIO_IOMMU_GET_CLASS(bcontainer) != iommufd_vioc ||
vbasedev->iommufd != container->be) {
continue;
QLIST_INIT(&container->hwpt_list);
vbasedev->cpr.ioas_id = ioas_id;
- bcontainer = &container->bcontainer;
+ bcontainer = VFIO_IOMMU(container);
vfio_address_space_insert(space, bcontainer);
if (!iommufd_cdev_attach_container(vbasedev, container, errp)) {
{
VFIOContainer *bcontainer = vbasedev->bcontainer;
VFIOAddressSpace *space = bcontainer->space;
- VFIOIOMMUFDContainer *container = container_of(bcontainer,
- VFIOIOMMUFDContainer,
- bcontainer);
+ VFIOIOMMUFDContainer *container = VFIO_IOMMU_IOMMUFD(bcontainer);
+
vfio_device_unprepare(vbasedev);
if (!vbasedev->ram_block_discard_allowed) {