| /* |
| * Migration support for VFIO devices |
| * |
| * Copyright NVIDIA, Inc. 2020 |
| * |
| * This work is licensed under the terms of the GNU GPL, version 2. See |
| * the COPYING file in the top-level directory. |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "qemu/main-loop.h" |
| #include "qemu/cutils.h" |
| #include <linux/vfio.h> |
| #include <sys/ioctl.h> |
| |
| #include "sysemu/runstate.h" |
| #include "hw/vfio/vfio-common.h" |
| #include "cpu.h" |
| #include "migration/migration.h" |
| #include "migration/vmstate.h" |
| #include "migration/qemu-file.h" |
| #include "migration/register.h" |
| #include "migration/blocker.h" |
| #include "migration/misc.h" |
| #include "qapi/error.h" |
| #include "exec/ramlist.h" |
| #include "exec/ram_addr.h" |
| #include "pci.h" |
| #include "trace.h" |
| #include "hw/hw.h" |
| |
| /* |
| * Flags to be used as unique delimiters for VFIO devices in the migration |
| * stream. These flags are composed as: |
| * 0xffffffff => MSB 32-bit all 1s |
| * 0xef10 => Magic ID, represents emulated (virtual) function IO |
| * 0x0000 => 16-bits reserved for flags |
| * |
| * The beginning of state information is marked by _DEV_CONFIG_STATE, |
| * _DEV_SETUP_STATE, or _DEV_DATA_STATE, respectively. The end of a |
| * certain state information is marked by _END_OF_STATE. |
| */ |
| #define VFIO_MIG_FLAG_END_OF_STATE (0xffffffffef100001ULL) |
| #define VFIO_MIG_FLAG_DEV_CONFIG_STATE (0xffffffffef100002ULL) |
| #define VFIO_MIG_FLAG_DEV_SETUP_STATE (0xffffffffef100003ULL) |
| #define VFIO_MIG_FLAG_DEV_DATA_STATE (0xffffffffef100004ULL) |
| |
| static int64_t bytes_transferred; |
| |
| static inline int vfio_mig_access(VFIODevice *vbasedev, void *val, int count, |
| off_t off, bool iswrite) |
| { |
| int ret; |
| |
| ret = iswrite ? pwrite(vbasedev->fd, val, count, off) : |
| pread(vbasedev->fd, val, count, off); |
| if (ret < count) { |
| error_report("vfio_mig_%s %d byte %s: failed at offset 0x%" |
| HWADDR_PRIx", err: %s", iswrite ? "write" : "read", count, |
| vbasedev->name, off, strerror(errno)); |
| return (ret < 0) ? ret : -EINVAL; |
| } |
| return 0; |
| } |
| |
| static int vfio_mig_rw(VFIODevice *vbasedev, __u8 *buf, size_t count, |
| off_t off, bool iswrite) |
| { |
| int ret, done = 0; |
| __u8 *tbuf = buf; |
| |
| while (count) { |
| int bytes = 0; |
| |
| if (count >= 8 && !(off % 8)) { |
| bytes = 8; |
| } else if (count >= 4 && !(off % 4)) { |
| bytes = 4; |
| } else if (count >= 2 && !(off % 2)) { |
| bytes = 2; |
| } else { |
| bytes = 1; |
| } |
| |
| ret = vfio_mig_access(vbasedev, tbuf, bytes, off, iswrite); |
| if (ret) { |
| return ret; |
| } |
| |
| count -= bytes; |
| done += bytes; |
| off += bytes; |
| tbuf += bytes; |
| } |
| return done; |
| } |
| |
| #define vfio_mig_read(f, v, c, o) vfio_mig_rw(f, (__u8 *)v, c, o, false) |
| #define vfio_mig_write(f, v, c, o) vfio_mig_rw(f, (__u8 *)v, c, o, true) |
| |
| #define VFIO_MIG_STRUCT_OFFSET(f) \ |
| offsetof(struct vfio_device_migration_info, f) |
| /* |
| * Change the device_state register for device @vbasedev. Bits set in @mask |
| * are preserved, bits set in @value are set, and bits not set in either @mask |
| * or @value are cleared in device_state. If the register cannot be accessed, |
| * the resulting state would be invalid, or the device enters an error state, |
| * an error is returned. |
| */ |
| |
| static int vfio_migration_set_state(VFIODevice *vbasedev, uint32_t mask, |
| uint32_t value) |
| { |
| VFIOMigration *migration = vbasedev->migration; |
| VFIORegion *region = &migration->region; |
| off_t dev_state_off = region->fd_offset + |
| VFIO_MIG_STRUCT_OFFSET(device_state); |
| uint32_t device_state; |
| int ret; |
| |
| ret = vfio_mig_read(vbasedev, &device_state, sizeof(device_state), |
| dev_state_off); |
| if (ret < 0) { |
| return ret; |
| } |
| |
| device_state = (device_state & mask) | value; |
| |
| if (!VFIO_DEVICE_STATE_VALID(device_state)) { |
| return -EINVAL; |
| } |
| |
| ret = vfio_mig_write(vbasedev, &device_state, sizeof(device_state), |
| dev_state_off); |
| if (ret < 0) { |
| int rret; |
| |
| rret = vfio_mig_read(vbasedev, &device_state, sizeof(device_state), |
| dev_state_off); |
| |
| if ((rret < 0) || (VFIO_DEVICE_STATE_IS_ERROR(device_state))) { |
| hw_error("%s: Device in error state 0x%x", vbasedev->name, |
| device_state); |
| return rret ? rret : -EIO; |
| } |
| return ret; |
| } |
| |
| migration->device_state = device_state; |
| trace_vfio_migration_set_state(vbasedev->name, device_state); |
| return 0; |
| } |
| |
| static void *get_data_section_size(VFIORegion *region, uint64_t data_offset, |
| uint64_t data_size, uint64_t *size) |
| { |
| void *ptr = NULL; |
| uint64_t limit = 0; |
| int i; |
| |
| if (!region->mmaps) { |
| if (size) { |
| *size = MIN(data_size, region->size - data_offset); |
| } |
| return ptr; |
| } |
| |
| for (i = 0; i < region->nr_mmaps; i++) { |
| VFIOMmap *map = region->mmaps + i; |
| |
| if ((data_offset >= map->offset) && |
| (data_offset < map->offset + map->size)) { |
| |
| /* check if data_offset is within sparse mmap areas */ |
| ptr = map->mmap + data_offset - map->offset; |
| if (size) { |
| *size = MIN(data_size, map->offset + map->size - data_offset); |
| } |
| break; |
| } else if ((data_offset < map->offset) && |
| (!limit || limit > map->offset)) { |
| /* |
| * data_offset is not within sparse mmap areas, find size of |
| * non-mapped area. Check through all list since region->mmaps list |
| * is not sorted. |
| */ |
| limit = map->offset; |
| } |
| } |
| |
| if (!ptr && size) { |
| *size = limit ? MIN(data_size, limit - data_offset) : data_size; |
| } |
| return ptr; |
| } |
| |
| static int vfio_save_buffer(QEMUFile *f, VFIODevice *vbasedev, uint64_t *size) |
| { |
| VFIOMigration *migration = vbasedev->migration; |
| VFIORegion *region = &migration->region; |
| uint64_t data_offset = 0, data_size = 0, sz; |
| int ret; |
| |
| ret = vfio_mig_read(vbasedev, &data_offset, sizeof(data_offset), |
| region->fd_offset + VFIO_MIG_STRUCT_OFFSET(data_offset)); |
| if (ret < 0) { |
| return ret; |
| } |
| |
| ret = vfio_mig_read(vbasedev, &data_size, sizeof(data_size), |
| region->fd_offset + VFIO_MIG_STRUCT_OFFSET(data_size)); |
| if (ret < 0) { |
| return ret; |
| } |
| |
| trace_vfio_save_buffer(vbasedev->name, data_offset, data_size, |
| migration->pending_bytes); |
| |
| qemu_put_be64(f, data_size); |
| sz = data_size; |
| |
| while (sz) { |
| void *buf; |
| uint64_t sec_size; |
| bool buf_allocated = false; |
| |
| buf = get_data_section_size(region, data_offset, sz, &sec_size); |
| |
| if (!buf) { |
| buf = g_try_malloc(sec_size); |
| if (!buf) { |
| error_report("%s: Error allocating buffer ", __func__); |
| return -ENOMEM; |
| } |
| buf_allocated = true; |
| |
| ret = vfio_mig_read(vbasedev, buf, sec_size, |
| region->fd_offset + data_offset); |
| if (ret < 0) { |
| g_free(buf); |
| return ret; |
| } |
| } |
| |
| qemu_put_buffer(f, buf, sec_size); |
| |
| if (buf_allocated) { |
| g_free(buf); |
| } |
| sz -= sec_size; |
| data_offset += sec_size; |
| } |
| |
| ret = qemu_file_get_error(f); |
| |
| if (!ret && size) { |
| *size = data_size; |
| } |
| |
| bytes_transferred += data_size; |
| return ret; |
| } |
| |
| static int vfio_load_buffer(QEMUFile *f, VFIODevice *vbasedev, |
| uint64_t data_size) |
| { |
| VFIORegion *region = &vbasedev->migration->region; |
| uint64_t data_offset = 0, size, report_size; |
| int ret; |
| |
| do { |
| ret = vfio_mig_read(vbasedev, &data_offset, sizeof(data_offset), |
| region->fd_offset + VFIO_MIG_STRUCT_OFFSET(data_offset)); |
| if (ret < 0) { |
| return ret; |
| } |
| |
| if (data_offset + data_size > region->size) { |
| /* |
| * If data_size is greater than the data section of migration region |
| * then iterate the write buffer operation. This case can occur if |
| * size of migration region at destination is smaller than size of |
| * migration region at source. |
| */ |
| report_size = size = region->size - data_offset; |
| data_size -= size; |
| } else { |
| report_size = size = data_size; |
| data_size = 0; |
| } |
| |
| trace_vfio_load_state_device_data(vbasedev->name, data_offset, size); |
| |
| while (size) { |
| void *buf; |
| uint64_t sec_size; |
| bool buf_alloc = false; |
| |
| buf = get_data_section_size(region, data_offset, size, &sec_size); |
| |
| if (!buf) { |
| buf = g_try_malloc(sec_size); |
| if (!buf) { |
| error_report("%s: Error allocating buffer ", __func__); |
| return -ENOMEM; |
| } |
| buf_alloc = true; |
| } |
| |
| qemu_get_buffer(f, buf, sec_size); |
| |
| if (buf_alloc) { |
| ret = vfio_mig_write(vbasedev, buf, sec_size, |
| region->fd_offset + data_offset); |
| g_free(buf); |
| |
| if (ret < 0) { |
| return ret; |
| } |
| } |
| size -= sec_size; |
| data_offset += sec_size; |
| } |
| |
| ret = vfio_mig_write(vbasedev, &report_size, sizeof(report_size), |
| region->fd_offset + VFIO_MIG_STRUCT_OFFSET(data_size)); |
| if (ret < 0) { |
| return ret; |
| } |
| } while (data_size); |
| |
| return 0; |
| } |
| |
| static int vfio_update_pending(VFIODevice *vbasedev) |
| { |
| VFIOMigration *migration = vbasedev->migration; |
| VFIORegion *region = &migration->region; |
| uint64_t pending_bytes = 0; |
| int ret; |
| |
| ret = vfio_mig_read(vbasedev, &pending_bytes, sizeof(pending_bytes), |
| region->fd_offset + VFIO_MIG_STRUCT_OFFSET(pending_bytes)); |
| if (ret < 0) { |
| migration->pending_bytes = 0; |
| return ret; |
| } |
| |
| migration->pending_bytes = pending_bytes; |
| trace_vfio_update_pending(vbasedev->name, pending_bytes); |
| return 0; |
| } |
| |
| static int vfio_save_device_config_state(QEMUFile *f, void *opaque) |
| { |
| VFIODevice *vbasedev = opaque; |
| |
| qemu_put_be64(f, VFIO_MIG_FLAG_DEV_CONFIG_STATE); |
| |
| if (vbasedev->ops && vbasedev->ops->vfio_save_config) { |
| vbasedev->ops->vfio_save_config(vbasedev, f); |
| } |
| |
| qemu_put_be64(f, VFIO_MIG_FLAG_END_OF_STATE); |
| |
| trace_vfio_save_device_config_state(vbasedev->name); |
| |
| return qemu_file_get_error(f); |
| } |
| |
| static int vfio_load_device_config_state(QEMUFile *f, void *opaque) |
| { |
| VFIODevice *vbasedev = opaque; |
| uint64_t data; |
| |
| if (vbasedev->ops && vbasedev->ops->vfio_load_config) { |
| int ret; |
| |
| ret = vbasedev->ops->vfio_load_config(vbasedev, f); |
| if (ret) { |
| error_report("%s: Failed to load device config space", |
| vbasedev->name); |
| return ret; |
| } |
| } |
| |
| data = qemu_get_be64(f); |
| if (data != VFIO_MIG_FLAG_END_OF_STATE) { |
| error_report("%s: Failed loading device config space, " |
| "end flag incorrect 0x%"PRIx64, vbasedev->name, data); |
| return -EINVAL; |
| } |
| |
| trace_vfio_load_device_config_state(vbasedev->name); |
| return qemu_file_get_error(f); |
| } |
| |
| static int vfio_set_dirty_page_tracking(VFIODevice *vbasedev, bool start) |
| { |
| int ret; |
| VFIOMigration *migration = vbasedev->migration; |
| VFIOContainer *container = vbasedev->group->container; |
| struct vfio_iommu_type1_dirty_bitmap dirty = { |
| .argsz = sizeof(dirty), |
| }; |
| |
| if (start) { |
| if (migration->device_state & VFIO_DEVICE_STATE_SAVING) { |
| dirty.flags = VFIO_IOMMU_DIRTY_PAGES_FLAG_START; |
| } else { |
| return -EINVAL; |
| } |
| } else { |
| dirty.flags = VFIO_IOMMU_DIRTY_PAGES_FLAG_STOP; |
| } |
| |
| ret = ioctl(container->fd, VFIO_IOMMU_DIRTY_PAGES, &dirty); |
| if (ret) { |
| error_report("Failed to set dirty tracking flag 0x%x errno: %d", |
| dirty.flags, errno); |
| return -errno; |
| } |
| return ret; |
| } |
| |
| static void vfio_migration_cleanup(VFIODevice *vbasedev) |
| { |
| VFIOMigration *migration = vbasedev->migration; |
| |
| vfio_set_dirty_page_tracking(vbasedev, false); |
| |
| if (migration->region.mmaps) { |
| vfio_region_unmap(&migration->region); |
| } |
| } |
| |
| /* ---------------------------------------------------------------------- */ |
| |
| static int vfio_save_setup(QEMUFile *f, void *opaque) |
| { |
| VFIODevice *vbasedev = opaque; |
| VFIOMigration *migration = vbasedev->migration; |
| int ret; |
| |
| trace_vfio_save_setup(vbasedev->name); |
| |
| qemu_put_be64(f, VFIO_MIG_FLAG_DEV_SETUP_STATE); |
| |
| if (migration->region.mmaps) { |
| /* |
| * Calling vfio_region_mmap() from migration thread. Memory API called |
| * from this function require locking the iothread when called from |
| * outside the main loop thread. |
| */ |
| qemu_mutex_lock_iothread(); |
| ret = vfio_region_mmap(&migration->region); |
| qemu_mutex_unlock_iothread(); |
| if (ret) { |
| error_report("%s: Failed to mmap VFIO migration region: %s", |
| vbasedev->name, strerror(-ret)); |
| error_report("%s: Falling back to slow path", vbasedev->name); |
| } |
| } |
| |
| ret = vfio_migration_set_state(vbasedev, VFIO_DEVICE_STATE_MASK, |
| VFIO_DEVICE_STATE_SAVING); |
| if (ret) { |
| error_report("%s: Failed to set state SAVING", vbasedev->name); |
| return ret; |
| } |
| |
| ret = vfio_set_dirty_page_tracking(vbasedev, true); |
| if (ret) { |
| return ret; |
| } |
| |
| qemu_put_be64(f, VFIO_MIG_FLAG_END_OF_STATE); |
| |
| ret = qemu_file_get_error(f); |
| if (ret) { |
| return ret; |
| } |
| |
| return 0; |
| } |
| |
| static void vfio_save_cleanup(void *opaque) |
| { |
| VFIODevice *vbasedev = opaque; |
| |
| vfio_migration_cleanup(vbasedev); |
| trace_vfio_save_cleanup(vbasedev->name); |
| } |
| |
| static void vfio_save_pending(QEMUFile *f, void *opaque, |
| uint64_t threshold_size, |
| uint64_t *res_precopy_only, |
| uint64_t *res_compatible, |
| uint64_t *res_postcopy_only) |
| { |
| VFIODevice *vbasedev = opaque; |
| VFIOMigration *migration = vbasedev->migration; |
| int ret; |
| |
| ret = vfio_update_pending(vbasedev); |
| if (ret) { |
| return; |
| } |
| |
| *res_precopy_only += migration->pending_bytes; |
| |
| trace_vfio_save_pending(vbasedev->name, *res_precopy_only, |
| *res_postcopy_only, *res_compatible); |
| } |
| |
| static int vfio_save_iterate(QEMUFile *f, void *opaque) |
| { |
| VFIODevice *vbasedev = opaque; |
| VFIOMigration *migration = vbasedev->migration; |
| uint64_t data_size; |
| int ret; |
| |
| qemu_put_be64(f, VFIO_MIG_FLAG_DEV_DATA_STATE); |
| |
| if (migration->pending_bytes == 0) { |
| ret = vfio_update_pending(vbasedev); |
| if (ret) { |
| return ret; |
| } |
| |
| if (migration->pending_bytes == 0) { |
| qemu_put_be64(f, 0); |
| qemu_put_be64(f, VFIO_MIG_FLAG_END_OF_STATE); |
| /* indicates data finished, goto complete phase */ |
| return 1; |
| } |
| } |
| |
| ret = vfio_save_buffer(f, vbasedev, &data_size); |
| if (ret) { |
| error_report("%s: vfio_save_buffer failed %s", vbasedev->name, |
| strerror(errno)); |
| return ret; |
| } |
| |
| qemu_put_be64(f, VFIO_MIG_FLAG_END_OF_STATE); |
| |
| ret = qemu_file_get_error(f); |
| if (ret) { |
| return ret; |
| } |
| |
| /* |
| * Reset pending_bytes as .save_live_pending is not called during savevm or |
| * snapshot case, in such case vfio_update_pending() at the start of this |
| * function updates pending_bytes. |
| */ |
| migration->pending_bytes = 0; |
| trace_vfio_save_iterate(vbasedev->name, data_size); |
| return 0; |
| } |
| |
| static int vfio_save_complete_precopy(QEMUFile *f, void *opaque) |
| { |
| VFIODevice *vbasedev = opaque; |
| VFIOMigration *migration = vbasedev->migration; |
| uint64_t data_size; |
| int ret; |
| |
| ret = vfio_migration_set_state(vbasedev, ~VFIO_DEVICE_STATE_RUNNING, |
| VFIO_DEVICE_STATE_SAVING); |
| if (ret) { |
| error_report("%s: Failed to set state STOP and SAVING", |
| vbasedev->name); |
| return ret; |
| } |
| |
| ret = vfio_save_device_config_state(f, opaque); |
| if (ret) { |
| return ret; |
| } |
| |
| ret = vfio_update_pending(vbasedev); |
| if (ret) { |
| return ret; |
| } |
| |
| while (migration->pending_bytes > 0) { |
| qemu_put_be64(f, VFIO_MIG_FLAG_DEV_DATA_STATE); |
| ret = vfio_save_buffer(f, vbasedev, &data_size); |
| if (ret < 0) { |
| error_report("%s: Failed to save buffer", vbasedev->name); |
| return ret; |
| } |
| |
| if (data_size == 0) { |
| break; |
| } |
| |
| ret = vfio_update_pending(vbasedev); |
| if (ret) { |
| return ret; |
| } |
| } |
| |
| qemu_put_be64(f, VFIO_MIG_FLAG_END_OF_STATE); |
| |
| ret = qemu_file_get_error(f); |
| if (ret) { |
| return ret; |
| } |
| |
| ret = vfio_migration_set_state(vbasedev, ~VFIO_DEVICE_STATE_SAVING, 0); |
| if (ret) { |
| error_report("%s: Failed to set state STOPPED", vbasedev->name); |
| return ret; |
| } |
| |
| trace_vfio_save_complete_precopy(vbasedev->name); |
| return ret; |
| } |
| |
| static int vfio_load_setup(QEMUFile *f, void *opaque) |
| { |
| VFIODevice *vbasedev = opaque; |
| VFIOMigration *migration = vbasedev->migration; |
| int ret = 0; |
| |
| if (migration->region.mmaps) { |
| ret = vfio_region_mmap(&migration->region); |
| if (ret) { |
| error_report("%s: Failed to mmap VFIO migration region %d: %s", |
| vbasedev->name, migration->region.nr, |
| strerror(-ret)); |
| error_report("%s: Falling back to slow path", vbasedev->name); |
| } |
| } |
| |
| ret = vfio_migration_set_state(vbasedev, ~VFIO_DEVICE_STATE_MASK, |
| VFIO_DEVICE_STATE_RESUMING); |
| if (ret) { |
| error_report("%s: Failed to set state RESUMING", vbasedev->name); |
| if (migration->region.mmaps) { |
| vfio_region_unmap(&migration->region); |
| } |
| } |
| return ret; |
| } |
| |
| static int vfio_load_cleanup(void *opaque) |
| { |
| VFIODevice *vbasedev = opaque; |
| |
| vfio_migration_cleanup(vbasedev); |
| trace_vfio_load_cleanup(vbasedev->name); |
| return 0; |
| } |
| |
| static int vfio_load_state(QEMUFile *f, void *opaque, int version_id) |
| { |
| VFIODevice *vbasedev = opaque; |
| int ret = 0; |
| uint64_t data; |
| |
| data = qemu_get_be64(f); |
| while (data != VFIO_MIG_FLAG_END_OF_STATE) { |
| |
| trace_vfio_load_state(vbasedev->name, data); |
| |
| switch (data) { |
| case VFIO_MIG_FLAG_DEV_CONFIG_STATE: |
| { |
| ret = vfio_load_device_config_state(f, opaque); |
| if (ret) { |
| return ret; |
| } |
| break; |
| } |
| case VFIO_MIG_FLAG_DEV_SETUP_STATE: |
| { |
| data = qemu_get_be64(f); |
| if (data == VFIO_MIG_FLAG_END_OF_STATE) { |
| return ret; |
| } else { |
| error_report("%s: SETUP STATE: EOS not found 0x%"PRIx64, |
| vbasedev->name, data); |
| return -EINVAL; |
| } |
| break; |
| } |
| case VFIO_MIG_FLAG_DEV_DATA_STATE: |
| { |
| uint64_t data_size = qemu_get_be64(f); |
| |
| if (data_size) { |
| ret = vfio_load_buffer(f, vbasedev, data_size); |
| if (ret < 0) { |
| return ret; |
| } |
| } |
| break; |
| } |
| default: |
| error_report("%s: Unknown tag 0x%"PRIx64, vbasedev->name, data); |
| return -EINVAL; |
| } |
| |
| data = qemu_get_be64(f); |
| ret = qemu_file_get_error(f); |
| if (ret) { |
| return ret; |
| } |
| } |
| return ret; |
| } |
| |
| static SaveVMHandlers savevm_vfio_handlers = { |
| .save_setup = vfio_save_setup, |
| .save_cleanup = vfio_save_cleanup, |
| .save_live_pending = vfio_save_pending, |
| .save_live_iterate = vfio_save_iterate, |
| .save_live_complete_precopy = vfio_save_complete_precopy, |
| .load_setup = vfio_load_setup, |
| .load_cleanup = vfio_load_cleanup, |
| .load_state = vfio_load_state, |
| }; |
| |
| /* ---------------------------------------------------------------------- */ |
| |
| static void vfio_vmstate_change(void *opaque, int running, RunState state) |
| { |
| VFIODevice *vbasedev = opaque; |
| VFIOMigration *migration = vbasedev->migration; |
| uint32_t value, mask; |
| int ret; |
| |
| if (vbasedev->migration->vm_running == running) { |
| return; |
| } |
| |
| if (running) { |
| /* |
| * Here device state can have one of _SAVING, _RESUMING or _STOP bit. |
| * Transition from _SAVING to _RUNNING can happen if there is migration |
| * failure, in that case clear _SAVING bit. |
| * Transition from _RESUMING to _RUNNING occurs during resuming |
| * phase, in that case clear _RESUMING bit. |
| * In both the above cases, set _RUNNING bit. |
| */ |
| mask = ~VFIO_DEVICE_STATE_MASK; |
| value = VFIO_DEVICE_STATE_RUNNING; |
| } else { |
| /* |
| * Here device state could be either _RUNNING or _SAVING|_RUNNING. Reset |
| * _RUNNING bit |
| */ |
| mask = ~VFIO_DEVICE_STATE_RUNNING; |
| value = 0; |
| } |
| |
| ret = vfio_migration_set_state(vbasedev, mask, value); |
| if (ret) { |
| /* |
| * Migration should be aborted in this case, but vm_state_notify() |
| * currently does not support reporting failures. |
| */ |
| error_report("%s: Failed to set device state 0x%x", vbasedev->name, |
| (migration->device_state & mask) | value); |
| qemu_file_set_error(migrate_get_current()->to_dst_file, ret); |
| } |
| vbasedev->migration->vm_running = running; |
| trace_vfio_vmstate_change(vbasedev->name, running, RunState_str(state), |
| (migration->device_state & mask) | value); |
| } |
| |
| static void vfio_migration_state_notifier(Notifier *notifier, void *data) |
| { |
| MigrationState *s = data; |
| VFIOMigration *migration = container_of(notifier, VFIOMigration, |
| migration_state); |
| VFIODevice *vbasedev = migration->vbasedev; |
| int ret; |
| |
| trace_vfio_migration_state_notifier(vbasedev->name, |
| MigrationStatus_str(s->state)); |
| |
| switch (s->state) { |
| case MIGRATION_STATUS_CANCELLING: |
| case MIGRATION_STATUS_CANCELLED: |
| case MIGRATION_STATUS_FAILED: |
| bytes_transferred = 0; |
| ret = vfio_migration_set_state(vbasedev, |
| ~(VFIO_DEVICE_STATE_SAVING | VFIO_DEVICE_STATE_RESUMING), |
| VFIO_DEVICE_STATE_RUNNING); |
| if (ret) { |
| error_report("%s: Failed to set state RUNNING", vbasedev->name); |
| } |
| } |
| } |
| |
| static void vfio_migration_exit(VFIODevice *vbasedev) |
| { |
| VFIOMigration *migration = vbasedev->migration; |
| |
| vfio_region_exit(&migration->region); |
| vfio_region_finalize(&migration->region); |
| g_free(vbasedev->migration); |
| vbasedev->migration = NULL; |
| } |
| |
| static int vfio_migration_init(VFIODevice *vbasedev, |
| struct vfio_region_info *info) |
| { |
| int ret; |
| Object *obj; |
| VFIOMigration *migration; |
| char id[256] = ""; |
| g_autofree char *path = NULL, *oid = NULL; |
| |
| if (!vbasedev->ops->vfio_get_object) { |
| return -EINVAL; |
| } |
| |
| obj = vbasedev->ops->vfio_get_object(vbasedev); |
| if (!obj) { |
| return -EINVAL; |
| } |
| |
| vbasedev->migration = g_new0(VFIOMigration, 1); |
| |
| ret = vfio_region_setup(obj, vbasedev, &vbasedev->migration->region, |
| info->index, "migration"); |
| if (ret) { |
| error_report("%s: Failed to setup VFIO migration region %d: %s", |
| vbasedev->name, info->index, strerror(-ret)); |
| goto err; |
| } |
| |
| if (!vbasedev->migration->region.size) { |
| error_report("%s: Invalid zero-sized VFIO migration region %d", |
| vbasedev->name, info->index); |
| ret = -EINVAL; |
| goto err; |
| } |
| |
| migration = vbasedev->migration; |
| migration->vbasedev = vbasedev; |
| |
| oid = vmstate_if_get_id(VMSTATE_IF(DEVICE(obj))); |
| if (oid) { |
| path = g_strdup_printf("%s/vfio", oid); |
| } else { |
| path = g_strdup("vfio"); |
| } |
| strpadcpy(id, sizeof(id), path, '\0'); |
| |
| register_savevm_live(id, VMSTATE_INSTANCE_ID_ANY, 1, &savevm_vfio_handlers, |
| vbasedev); |
| |
| migration->vm_state = qemu_add_vm_change_state_handler(vfio_vmstate_change, |
| vbasedev); |
| migration->migration_state.notify = vfio_migration_state_notifier; |
| add_migration_state_change_notifier(&migration->migration_state); |
| return 0; |
| |
| err: |
| vfio_migration_exit(vbasedev); |
| return ret; |
| } |
| |
| /* ---------------------------------------------------------------------- */ |
| |
| int64_t vfio_mig_bytes_transferred(void) |
| { |
| return bytes_transferred; |
| } |
| |
| int vfio_migration_probe(VFIODevice *vbasedev, Error **errp) |
| { |
| VFIOContainer *container = vbasedev->group->container; |
| struct vfio_region_info *info = NULL; |
| Error *local_err = NULL; |
| int ret = -ENOTSUP; |
| |
| if (!vbasedev->enable_migration || !container->dirty_pages_supported) { |
| goto add_blocker; |
| } |
| |
| ret = vfio_get_dev_region_info(vbasedev, VFIO_REGION_TYPE_MIGRATION, |
| VFIO_REGION_SUBTYPE_MIGRATION, &info); |
| if (ret) { |
| goto add_blocker; |
| } |
| |
| ret = vfio_migration_init(vbasedev, info); |
| if (ret) { |
| goto add_blocker; |
| } |
| |
| trace_vfio_migration_probe(vbasedev->name, info->index); |
| g_free(info); |
| return 0; |
| |
| add_blocker: |
| error_setg(&vbasedev->migration_blocker, |
| "VFIO device doesn't support migration"); |
| g_free(info); |
| |
| ret = migrate_add_blocker(vbasedev->migration_blocker, &local_err); |
| if (local_err) { |
| error_propagate(errp, local_err); |
| error_free(vbasedev->migration_blocker); |
| vbasedev->migration_blocker = NULL; |
| } |
| return ret; |
| } |
| |
| void vfio_migration_finalize(VFIODevice *vbasedev) |
| { |
| if (vbasedev->migration) { |
| VFIOMigration *migration = vbasedev->migration; |
| |
| remove_migration_state_change_notifier(&migration->migration_state); |
| qemu_del_vm_change_state_handler(migration->vm_state); |
| vfio_migration_exit(vbasedev); |
| } |
| |
| if (vbasedev->migration_blocker) { |
| migrate_del_blocker(vbasedev->migration_blocker); |
| error_free(vbasedev->migration_blocker); |
| vbasedev->migration_blocker = NULL; |
| } |
| } |