| /* |
| * iommufd container backend |
| * |
| * Copyright (C) 2023 Intel Corporation. |
| * Copyright Red Hat, Inc. 2023 |
| * |
| * Authors: Yi Liu <yi.l.liu@intel.com> |
| * Eric Auger <eric.auger@redhat.com> |
| * |
| * SPDX-License-Identifier: GPL-2.0-or-later |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "sysemu/iommufd.h" |
| #include "qapi/error.h" |
| #include "qemu/module.h" |
| #include "qom/object_interfaces.h" |
| #include "qemu/error-report.h" |
| #include "monitor/monitor.h" |
| #include "trace.h" |
| #include "hw/vfio/vfio-common.h" |
| #include <sys/ioctl.h> |
| #include <linux/iommufd.h> |
| |
| static void iommufd_backend_init(Object *obj) |
| { |
| IOMMUFDBackend *be = IOMMUFD_BACKEND(obj); |
| |
| be->fd = -1; |
| be->users = 0; |
| be->owned = true; |
| } |
| |
| static void iommufd_backend_finalize(Object *obj) |
| { |
| IOMMUFDBackend *be = IOMMUFD_BACKEND(obj); |
| |
| if (be->owned) { |
| close(be->fd); |
| be->fd = -1; |
| } |
| } |
| |
| static void iommufd_backend_set_fd(Object *obj, const char *str, Error **errp) |
| { |
| ERRP_GUARD(); |
| IOMMUFDBackend *be = IOMMUFD_BACKEND(obj); |
| int fd = -1; |
| |
| fd = monitor_fd_param(monitor_cur(), str, errp); |
| if (fd == -1) { |
| error_prepend(errp, "Could not parse remote object fd %s:", str); |
| return; |
| } |
| be->fd = fd; |
| be->owned = false; |
| trace_iommu_backend_set_fd(be->fd); |
| } |
| |
| static bool iommufd_backend_can_be_deleted(UserCreatable *uc) |
| { |
| IOMMUFDBackend *be = IOMMUFD_BACKEND(uc); |
| |
| return !be->users; |
| } |
| |
| static void iommufd_backend_class_init(ObjectClass *oc, void *data) |
| { |
| UserCreatableClass *ucc = USER_CREATABLE_CLASS(oc); |
| |
| ucc->can_be_deleted = iommufd_backend_can_be_deleted; |
| |
| object_class_property_add_str(oc, "fd", NULL, iommufd_backend_set_fd); |
| } |
| |
| bool iommufd_backend_connect(IOMMUFDBackend *be, Error **errp) |
| { |
| int fd; |
| |
| if (be->owned && !be->users) { |
| fd = qemu_open("/dev/iommu", O_RDWR, errp); |
| if (fd < 0) { |
| return false; |
| } |
| be->fd = fd; |
| } |
| be->users++; |
| |
| trace_iommufd_backend_connect(be->fd, be->owned, be->users); |
| return true; |
| } |
| |
| void iommufd_backend_disconnect(IOMMUFDBackend *be) |
| { |
| if (!be->users) { |
| goto out; |
| } |
| be->users--; |
| if (!be->users && be->owned) { |
| close(be->fd); |
| be->fd = -1; |
| } |
| out: |
| trace_iommufd_backend_disconnect(be->fd, be->users); |
| } |
| |
| bool iommufd_backend_alloc_ioas(IOMMUFDBackend *be, uint32_t *ioas_id, |
| Error **errp) |
| { |
| int fd = be->fd; |
| struct iommu_ioas_alloc alloc_data = { |
| .size = sizeof(alloc_data), |
| .flags = 0, |
| }; |
| |
| if (ioctl(fd, IOMMU_IOAS_ALLOC, &alloc_data)) { |
| error_setg_errno(errp, errno, "Failed to allocate ioas"); |
| return false; |
| } |
| |
| *ioas_id = alloc_data.out_ioas_id; |
| trace_iommufd_backend_alloc_ioas(fd, *ioas_id); |
| |
| return true; |
| } |
| |
| void iommufd_backend_free_id(IOMMUFDBackend *be, uint32_t id) |
| { |
| int ret, fd = be->fd; |
| struct iommu_destroy des = { |
| .size = sizeof(des), |
| .id = id, |
| }; |
| |
| ret = ioctl(fd, IOMMU_DESTROY, &des); |
| trace_iommufd_backend_free_id(fd, id, ret); |
| if (ret) { |
| error_report("Failed to free id: %u %m", id); |
| } |
| } |
| |
| int iommufd_backend_map_dma(IOMMUFDBackend *be, uint32_t ioas_id, hwaddr iova, |
| ram_addr_t size, void *vaddr, bool readonly) |
| { |
| int ret, fd = be->fd; |
| struct iommu_ioas_map map = { |
| .size = sizeof(map), |
| .flags = IOMMU_IOAS_MAP_READABLE | |
| IOMMU_IOAS_MAP_FIXED_IOVA, |
| .ioas_id = ioas_id, |
| .__reserved = 0, |
| .user_va = (uintptr_t)vaddr, |
| .iova = iova, |
| .length = size, |
| }; |
| |
| if (!readonly) { |
| map.flags |= IOMMU_IOAS_MAP_WRITEABLE; |
| } |
| |
| ret = ioctl(fd, IOMMU_IOAS_MAP, &map); |
| trace_iommufd_backend_map_dma(fd, ioas_id, iova, size, |
| vaddr, readonly, ret); |
| if (ret) { |
| ret = -errno; |
| |
| /* TODO: Not support mapping hardware PCI BAR region for now. */ |
| if (errno == EFAULT) { |
| warn_report("IOMMU_IOAS_MAP failed: %m, PCI BAR?"); |
| } else { |
| error_report("IOMMU_IOAS_MAP failed: %m"); |
| } |
| } |
| return ret; |
| } |
| |
| int iommufd_backend_unmap_dma(IOMMUFDBackend *be, uint32_t ioas_id, |
| hwaddr iova, ram_addr_t size) |
| { |
| int ret, fd = be->fd; |
| struct iommu_ioas_unmap unmap = { |
| .size = sizeof(unmap), |
| .ioas_id = ioas_id, |
| .iova = iova, |
| .length = size, |
| }; |
| |
| ret = ioctl(fd, IOMMU_IOAS_UNMAP, &unmap); |
| /* |
| * IOMMUFD takes mapping as some kind of object, unmapping |
| * nonexistent mapping is treated as deleting a nonexistent |
| * object and return ENOENT. This is different from legacy |
| * backend which allows it. vIOMMU may trigger a lot of |
| * redundant unmapping, to avoid flush the log, treat them |
| * as succeess for IOMMUFD just like legacy backend. |
| */ |
| if (ret && errno == ENOENT) { |
| trace_iommufd_backend_unmap_dma_non_exist(fd, ioas_id, iova, size, ret); |
| ret = 0; |
| } else { |
| trace_iommufd_backend_unmap_dma(fd, ioas_id, iova, size, ret); |
| } |
| |
| if (ret) { |
| ret = -errno; |
| error_report("IOMMU_IOAS_UNMAP failed: %m"); |
| } |
| return ret; |
| } |
| |
| bool iommufd_backend_alloc_hwpt(IOMMUFDBackend *be, uint32_t dev_id, |
| uint32_t pt_id, uint32_t flags, |
| uint32_t data_type, uint32_t data_len, |
| void *data_ptr, uint32_t *out_hwpt, |
| Error **errp) |
| { |
| int ret, fd = be->fd; |
| struct iommu_hwpt_alloc alloc_hwpt = { |
| .size = sizeof(struct iommu_hwpt_alloc), |
| .flags = flags, |
| .dev_id = dev_id, |
| .pt_id = pt_id, |
| .data_type = data_type, |
| .data_len = data_len, |
| .data_uptr = (uintptr_t)data_ptr, |
| }; |
| |
| ret = ioctl(fd, IOMMU_HWPT_ALLOC, &alloc_hwpt); |
| trace_iommufd_backend_alloc_hwpt(fd, dev_id, pt_id, flags, data_type, |
| data_len, (uintptr_t)data_ptr, |
| alloc_hwpt.out_hwpt_id, ret); |
| if (ret) { |
| error_setg_errno(errp, errno, "Failed to allocate hwpt"); |
| return false; |
| } |
| |
| *out_hwpt = alloc_hwpt.out_hwpt_id; |
| return true; |
| } |
| |
| bool iommufd_backend_get_device_info(IOMMUFDBackend *be, uint32_t devid, |
| uint32_t *type, void *data, uint32_t len, |
| uint64_t *caps, Error **errp) |
| { |
| struct iommu_hw_info info = { |
| .size = sizeof(info), |
| .dev_id = devid, |
| .data_len = len, |
| .data_uptr = (uintptr_t)data, |
| }; |
| |
| if (ioctl(be->fd, IOMMU_GET_HW_INFO, &info)) { |
| error_setg_errno(errp, errno, "Failed to get hardware info"); |
| return false; |
| } |
| |
| g_assert(type); |
| *type = info.out_data_type; |
| g_assert(caps); |
| *caps = info.out_capabilities; |
| |
| return true; |
| } |
| |
| static int hiod_iommufd_get_cap(HostIOMMUDevice *hiod, int cap, Error **errp) |
| { |
| HostIOMMUDeviceCaps *caps = &hiod->caps; |
| |
| switch (cap) { |
| case HOST_IOMMU_DEVICE_CAP_IOMMU_TYPE: |
| return caps->type; |
| case HOST_IOMMU_DEVICE_CAP_AW_BITS: |
| return vfio_device_get_aw_bits(hiod->agent); |
| default: |
| error_setg(errp, "%s: unsupported capability %x", hiod->name, cap); |
| return -EINVAL; |
| } |
| } |
| |
| static void hiod_iommufd_class_init(ObjectClass *oc, void *data) |
| { |
| HostIOMMUDeviceClass *hioc = HOST_IOMMU_DEVICE_CLASS(oc); |
| |
| hioc->get_cap = hiod_iommufd_get_cap; |
| }; |
| |
| static const TypeInfo types[] = { |
| { |
| .name = TYPE_IOMMUFD_BACKEND, |
| .parent = TYPE_OBJECT, |
| .instance_size = sizeof(IOMMUFDBackend), |
| .instance_init = iommufd_backend_init, |
| .instance_finalize = iommufd_backend_finalize, |
| .class_size = sizeof(IOMMUFDBackendClass), |
| .class_init = iommufd_backend_class_init, |
| .interfaces = (InterfaceInfo[]) { |
| { TYPE_USER_CREATABLE }, |
| { } |
| } |
| }, { |
| .name = TYPE_HOST_IOMMU_DEVICE_IOMMUFD, |
| .parent = TYPE_HOST_IOMMU_DEVICE, |
| .class_init = hiod_iommufd_class_init, |
| .abstract = true, |
| } |
| }; |
| |
| DEFINE_TYPES(types) |