|  | /* | 
|  | * Migration support for VFIO devices | 
|  | * | 
|  | * Copyright NVIDIA, Inc. 2020 | 
|  | * | 
|  | * This work is licensed under the terms of the GNU GPL, version 2. See | 
|  | * the COPYING file in the top-level directory. | 
|  | */ | 
|  |  | 
|  | #include "qemu/osdep.h" | 
|  | #include "qemu/main-loop.h" | 
|  | #include "qemu/cutils.h" | 
|  | #include "qemu/units.h" | 
|  | #include "qemu/error-report.h" | 
|  | #include <linux/vfio.h> | 
|  | #include <sys/ioctl.h> | 
|  |  | 
|  | #include "system/runstate.h" | 
|  | #include "hw/vfio/vfio-common.h" | 
|  | #include "migration/misc.h" | 
|  | #include "migration/savevm.h" | 
|  | #include "migration/vmstate.h" | 
|  | #include "migration/qemu-file.h" | 
|  | #include "migration/register.h" | 
|  | #include "migration/blocker.h" | 
|  | #include "qapi/error.h" | 
|  | #include "qapi/qapi-events-vfio.h" | 
|  | #include "exec/ramlist.h" | 
|  | #include "exec/ram_addr.h" | 
|  | #include "pci.h" | 
|  | #include "trace.h" | 
|  | #include "hw/hw.h" | 
|  |  | 
|  | /* | 
|  | * Flags to be used as unique delimiters for VFIO devices in the migration | 
|  | * stream. These flags are composed as: | 
|  | * 0xffffffff => MSB 32-bit all 1s | 
|  | * 0xef10     => Magic ID, represents emulated (virtual) function IO | 
|  | * 0x0000     => 16-bits reserved for flags | 
|  | * | 
|  | * The beginning of state information is marked by _DEV_CONFIG_STATE, | 
|  | * _DEV_SETUP_STATE, or _DEV_DATA_STATE, respectively. The end of a | 
|  | * certain state information is marked by _END_OF_STATE. | 
|  | */ | 
|  | #define VFIO_MIG_FLAG_END_OF_STATE      (0xffffffffef100001ULL) | 
|  | #define VFIO_MIG_FLAG_DEV_CONFIG_STATE  (0xffffffffef100002ULL) | 
|  | #define VFIO_MIG_FLAG_DEV_SETUP_STATE   (0xffffffffef100003ULL) | 
|  | #define VFIO_MIG_FLAG_DEV_DATA_STATE    (0xffffffffef100004ULL) | 
|  | #define VFIO_MIG_FLAG_DEV_INIT_DATA_SENT (0xffffffffef100005ULL) | 
|  |  | 
|  | /* | 
|  | * This is an arbitrary size based on migration of mlx5 devices, where typically | 
|  | * total device migration size is on the order of 100s of MB. Testing with | 
|  | * larger values, e.g. 128MB and 1GB, did not show a performance improvement. | 
|  | */ | 
|  | #define VFIO_MIG_DEFAULT_DATA_BUFFER_SIZE (1 * MiB) | 
|  |  | 
|  | static int64_t bytes_transferred; | 
|  |  | 
|  | static const char *mig_state_to_str(enum vfio_device_mig_state state) | 
|  | { | 
|  | switch (state) { | 
|  | case VFIO_DEVICE_STATE_ERROR: | 
|  | return "ERROR"; | 
|  | case VFIO_DEVICE_STATE_STOP: | 
|  | return "STOP"; | 
|  | case VFIO_DEVICE_STATE_RUNNING: | 
|  | return "RUNNING"; | 
|  | case VFIO_DEVICE_STATE_STOP_COPY: | 
|  | return "STOP_COPY"; | 
|  | case VFIO_DEVICE_STATE_RESUMING: | 
|  | return "RESUMING"; | 
|  | case VFIO_DEVICE_STATE_RUNNING_P2P: | 
|  | return "RUNNING_P2P"; | 
|  | case VFIO_DEVICE_STATE_PRE_COPY: | 
|  | return "PRE_COPY"; | 
|  | case VFIO_DEVICE_STATE_PRE_COPY_P2P: | 
|  | return "PRE_COPY_P2P"; | 
|  | default: | 
|  | return "UNKNOWN STATE"; | 
|  | } | 
|  | } | 
|  |  | 
|  | static QapiVfioMigrationState | 
|  | mig_state_to_qapi_state(enum vfio_device_mig_state state) | 
|  | { | 
|  | switch (state) { | 
|  | case VFIO_DEVICE_STATE_STOP: | 
|  | return QAPI_VFIO_MIGRATION_STATE_STOP; | 
|  | case VFIO_DEVICE_STATE_RUNNING: | 
|  | return QAPI_VFIO_MIGRATION_STATE_RUNNING; | 
|  | case VFIO_DEVICE_STATE_STOP_COPY: | 
|  | return QAPI_VFIO_MIGRATION_STATE_STOP_COPY; | 
|  | case VFIO_DEVICE_STATE_RESUMING: | 
|  | return QAPI_VFIO_MIGRATION_STATE_RESUMING; | 
|  | case VFIO_DEVICE_STATE_RUNNING_P2P: | 
|  | return QAPI_VFIO_MIGRATION_STATE_RUNNING_P2P; | 
|  | case VFIO_DEVICE_STATE_PRE_COPY: | 
|  | return QAPI_VFIO_MIGRATION_STATE_PRE_COPY; | 
|  | case VFIO_DEVICE_STATE_PRE_COPY_P2P: | 
|  | return QAPI_VFIO_MIGRATION_STATE_PRE_COPY_P2P; | 
|  | default: | 
|  | g_assert_not_reached(); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void vfio_migration_send_event(VFIODevice *vbasedev) | 
|  | { | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  | DeviceState *dev = vbasedev->dev; | 
|  | g_autofree char *qom_path = NULL; | 
|  | Object *obj; | 
|  |  | 
|  | if (!vbasedev->migration_events) { | 
|  | return; | 
|  | } | 
|  |  | 
|  | g_assert(vbasedev->ops->vfio_get_object); | 
|  | obj = vbasedev->ops->vfio_get_object(vbasedev); | 
|  | g_assert(obj); | 
|  | qom_path = object_get_canonical_path(obj); | 
|  |  | 
|  | qapi_event_send_vfio_migration( | 
|  | dev->id, qom_path, mig_state_to_qapi_state(migration->device_state)); | 
|  | } | 
|  |  | 
|  | static void vfio_migration_set_device_state(VFIODevice *vbasedev, | 
|  | enum vfio_device_mig_state state) | 
|  | { | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  |  | 
|  | trace_vfio_migration_set_device_state(vbasedev->name, | 
|  | mig_state_to_str(state)); | 
|  |  | 
|  | migration->device_state = state; | 
|  | vfio_migration_send_event(vbasedev); | 
|  | } | 
|  |  | 
|  | static int vfio_migration_set_state(VFIODevice *vbasedev, | 
|  | enum vfio_device_mig_state new_state, | 
|  | enum vfio_device_mig_state recover_state, | 
|  | Error **errp) | 
|  | { | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  | uint64_t buf[DIV_ROUND_UP(sizeof(struct vfio_device_feature) + | 
|  | sizeof(struct vfio_device_feature_mig_state), | 
|  | sizeof(uint64_t))] = {}; | 
|  | struct vfio_device_feature *feature = (struct vfio_device_feature *)buf; | 
|  | struct vfio_device_feature_mig_state *mig_state = | 
|  | (struct vfio_device_feature_mig_state *)feature->data; | 
|  | int ret; | 
|  | g_autofree char *error_prefix = | 
|  | g_strdup_printf("%s: Failed setting device state to %s.", | 
|  | vbasedev->name, mig_state_to_str(new_state)); | 
|  |  | 
|  | trace_vfio_migration_set_state(vbasedev->name, mig_state_to_str(new_state), | 
|  | mig_state_to_str(recover_state)); | 
|  |  | 
|  | if (new_state == migration->device_state) { | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | feature->argsz = sizeof(buf); | 
|  | feature->flags = | 
|  | VFIO_DEVICE_FEATURE_SET | VFIO_DEVICE_FEATURE_MIG_DEVICE_STATE; | 
|  | mig_state->device_state = new_state; | 
|  | if (ioctl(vbasedev->fd, VFIO_DEVICE_FEATURE, feature)) { | 
|  | /* Try to set the device in some good state */ | 
|  | ret = -errno; | 
|  |  | 
|  | if (recover_state == VFIO_DEVICE_STATE_ERROR) { | 
|  | error_setg_errno(errp, errno, | 
|  | "%s Recover state is ERROR. Resetting device", | 
|  | error_prefix); | 
|  |  | 
|  | goto reset_device; | 
|  | } | 
|  |  | 
|  | error_setg_errno(errp, errno, | 
|  | "%s Setting device in recover state %s", | 
|  | error_prefix, mig_state_to_str(recover_state)); | 
|  |  | 
|  | mig_state->device_state = recover_state; | 
|  | if (ioctl(vbasedev->fd, VFIO_DEVICE_FEATURE, feature)) { | 
|  | ret = -errno; | 
|  | /* | 
|  | * If setting the device in recover state fails, report | 
|  | * the error here and propagate the first error. | 
|  | */ | 
|  | error_report( | 
|  | "%s: Failed setting device in recover state, err: %s. Resetting device", | 
|  | vbasedev->name, strerror(errno)); | 
|  |  | 
|  | goto reset_device; | 
|  | } | 
|  |  | 
|  | vfio_migration_set_device_state(vbasedev, recover_state); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | vfio_migration_set_device_state(vbasedev, new_state); | 
|  | if (mig_state->data_fd != -1) { | 
|  | if (migration->data_fd != -1) { | 
|  | /* | 
|  | * This can happen if the device is asynchronously reset and | 
|  | * terminates a data transfer. | 
|  | */ | 
|  | error_setg(errp, "%s: data_fd out of sync", vbasedev->name); | 
|  | close(mig_state->data_fd); | 
|  |  | 
|  | return -EBADF; | 
|  | } | 
|  |  | 
|  | migration->data_fd = mig_state->data_fd; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  |  | 
|  | reset_device: | 
|  | if (ioctl(vbasedev->fd, VFIO_DEVICE_RESET)) { | 
|  | hw_error("%s: Failed resetting device, err: %s", vbasedev->name, | 
|  | strerror(errno)); | 
|  | } | 
|  |  | 
|  | vfio_migration_set_device_state(vbasedev, VFIO_DEVICE_STATE_RUNNING); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Some device state transitions require resetting the device if they fail. | 
|  | * This function sets the device in new_state and resets the device if that | 
|  | * fails. Reset is done by using ERROR as the recover state. | 
|  | */ | 
|  | static int | 
|  | vfio_migration_set_state_or_reset(VFIODevice *vbasedev, | 
|  | enum vfio_device_mig_state new_state, | 
|  | Error **errp) | 
|  | { | 
|  | return vfio_migration_set_state(vbasedev, new_state, | 
|  | VFIO_DEVICE_STATE_ERROR, errp); | 
|  | } | 
|  |  | 
|  | static int vfio_load_buffer(QEMUFile *f, VFIODevice *vbasedev, | 
|  | uint64_t data_size) | 
|  | { | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  | int ret; | 
|  |  | 
|  | ret = qemu_file_get_to_fd(f, migration->data_fd, data_size); | 
|  | trace_vfio_load_state_device_data(vbasedev->name, data_size, ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int vfio_save_device_config_state(QEMUFile *f, void *opaque, | 
|  | Error **errp) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | int ret; | 
|  |  | 
|  | qemu_put_be64(f, VFIO_MIG_FLAG_DEV_CONFIG_STATE); | 
|  |  | 
|  | if (vbasedev->ops && vbasedev->ops->vfio_save_config) { | 
|  | ret = vbasedev->ops->vfio_save_config(vbasedev, f, errp); | 
|  | if (ret) { | 
|  | return ret; | 
|  | } | 
|  | } | 
|  |  | 
|  | qemu_put_be64(f, VFIO_MIG_FLAG_END_OF_STATE); | 
|  |  | 
|  | trace_vfio_save_device_config_state(vbasedev->name); | 
|  |  | 
|  | ret = qemu_file_get_error(f); | 
|  | if (ret < 0) { | 
|  | error_setg_errno(errp, -ret, "Failed to save state"); | 
|  | } | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int vfio_load_device_config_state(QEMUFile *f, void *opaque) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | uint64_t data; | 
|  |  | 
|  | if (vbasedev->ops && vbasedev->ops->vfio_load_config) { | 
|  | int ret; | 
|  |  | 
|  | ret = vbasedev->ops->vfio_load_config(vbasedev, f); | 
|  | if (ret) { | 
|  | error_report("%s: Failed to load device config space", | 
|  | vbasedev->name); | 
|  | return ret; | 
|  | } | 
|  | } | 
|  |  | 
|  | data = qemu_get_be64(f); | 
|  | if (data != VFIO_MIG_FLAG_END_OF_STATE) { | 
|  | error_report("%s: Failed loading device config space, " | 
|  | "end flag incorrect 0x%"PRIx64, vbasedev->name, data); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | trace_vfio_load_device_config_state(vbasedev->name); | 
|  | return qemu_file_get_error(f); | 
|  | } | 
|  |  | 
|  | static void vfio_migration_cleanup(VFIODevice *vbasedev) | 
|  | { | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  |  | 
|  | close(migration->data_fd); | 
|  | migration->data_fd = -1; | 
|  | } | 
|  |  | 
|  | static int vfio_query_stop_copy_size(VFIODevice *vbasedev, | 
|  | uint64_t *stop_copy_size) | 
|  | { | 
|  | uint64_t buf[DIV_ROUND_UP(sizeof(struct vfio_device_feature) + | 
|  | sizeof(struct vfio_device_feature_mig_data_size), | 
|  | sizeof(uint64_t))] = {}; | 
|  | struct vfio_device_feature *feature = (struct vfio_device_feature *)buf; | 
|  | struct vfio_device_feature_mig_data_size *mig_data_size = | 
|  | (struct vfio_device_feature_mig_data_size *)feature->data; | 
|  |  | 
|  | feature->argsz = sizeof(buf); | 
|  | feature->flags = | 
|  | VFIO_DEVICE_FEATURE_GET | VFIO_DEVICE_FEATURE_MIG_DATA_SIZE; | 
|  |  | 
|  | if (ioctl(vbasedev->fd, VFIO_DEVICE_FEATURE, feature)) { | 
|  | return -errno; | 
|  | } | 
|  |  | 
|  | *stop_copy_size = mig_data_size->stop_copy_length; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int vfio_query_precopy_size(VFIOMigration *migration) | 
|  | { | 
|  | struct vfio_precopy_info precopy = { | 
|  | .argsz = sizeof(precopy), | 
|  | }; | 
|  |  | 
|  | migration->precopy_init_size = 0; | 
|  | migration->precopy_dirty_size = 0; | 
|  |  | 
|  | if (ioctl(migration->data_fd, VFIO_MIG_GET_PRECOPY_INFO, &precopy)) { | 
|  | return -errno; | 
|  | } | 
|  |  | 
|  | migration->precopy_init_size = precopy.initial_bytes; | 
|  | migration->precopy_dirty_size = precopy.dirty_bytes; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Returns the size of saved data on success and -errno on error */ | 
|  | static ssize_t vfio_save_block(QEMUFile *f, VFIOMigration *migration) | 
|  | { | 
|  | ssize_t data_size; | 
|  |  | 
|  | data_size = read(migration->data_fd, migration->data_buffer, | 
|  | migration->data_buffer_size); | 
|  | if (data_size < 0) { | 
|  | /* | 
|  | * Pre-copy emptied all the device state for now. For more information, | 
|  | * please refer to the Linux kernel VFIO uAPI. | 
|  | */ | 
|  | if (errno == ENOMSG) { | 
|  | if (!migration->event_precopy_empty_hit) { | 
|  | trace_vfio_save_block_precopy_empty_hit(migration->vbasedev->name); | 
|  | migration->event_precopy_empty_hit = true; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | return -errno; | 
|  | } | 
|  | if (data_size == 0) { | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Non-empty read: re-arm the trace event */ | 
|  | migration->event_precopy_empty_hit = false; | 
|  |  | 
|  | qemu_put_be64(f, VFIO_MIG_FLAG_DEV_DATA_STATE); | 
|  | qemu_put_be64(f, data_size); | 
|  | qemu_put_buffer(f, migration->data_buffer, data_size); | 
|  | bytes_transferred += data_size; | 
|  |  | 
|  | trace_vfio_save_block(migration->vbasedev->name, data_size); | 
|  |  | 
|  | return qemu_file_get_error(f) ?: data_size; | 
|  | } | 
|  |  | 
|  | static void vfio_update_estimated_pending_data(VFIOMigration *migration, | 
|  | uint64_t data_size) | 
|  | { | 
|  | if (!data_size) { | 
|  | /* | 
|  | * Pre-copy emptied all the device state for now, update estimated sizes | 
|  | * accordingly. | 
|  | */ | 
|  | migration->precopy_init_size = 0; | 
|  | migration->precopy_dirty_size = 0; | 
|  |  | 
|  | return; | 
|  | } | 
|  |  | 
|  | if (migration->precopy_init_size) { | 
|  | uint64_t init_size = MIN(migration->precopy_init_size, data_size); | 
|  |  | 
|  | migration->precopy_init_size -= init_size; | 
|  | data_size -= init_size; | 
|  | } | 
|  |  | 
|  | migration->precopy_dirty_size -= MIN(migration->precopy_dirty_size, | 
|  | data_size); | 
|  | } | 
|  |  | 
|  | static bool vfio_precopy_supported(VFIODevice *vbasedev) | 
|  | { | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  |  | 
|  | return migration->mig_flags & VFIO_MIGRATION_PRE_COPY; | 
|  | } | 
|  |  | 
|  | /* ---------------------------------------------------------------------- */ | 
|  |  | 
|  | static int vfio_save_prepare(void *opaque, Error **errp) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  |  | 
|  | /* | 
|  | * Snapshot doesn't use postcopy nor background snapshot, so allow snapshot | 
|  | * even if they are on. | 
|  | */ | 
|  | if (runstate_check(RUN_STATE_SAVE_VM)) { | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | if (migrate_postcopy_ram()) { | 
|  | error_setg( | 
|  | errp, "%s: VFIO migration is not supported with postcopy migration", | 
|  | vbasedev->name); | 
|  | return -EOPNOTSUPP; | 
|  | } | 
|  |  | 
|  | if (migrate_background_snapshot()) { | 
|  | error_setg( | 
|  | errp, | 
|  | "%s: VFIO migration is not supported with background snapshot", | 
|  | vbasedev->name); | 
|  | return -EOPNOTSUPP; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int vfio_save_setup(QEMUFile *f, void *opaque, Error **errp) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  | uint64_t stop_copy_size = VFIO_MIG_DEFAULT_DATA_BUFFER_SIZE; | 
|  | int ret; | 
|  |  | 
|  | qemu_put_be64(f, VFIO_MIG_FLAG_DEV_SETUP_STATE); | 
|  |  | 
|  | vfio_query_stop_copy_size(vbasedev, &stop_copy_size); | 
|  | migration->data_buffer_size = MIN(VFIO_MIG_DEFAULT_DATA_BUFFER_SIZE, | 
|  | stop_copy_size); | 
|  | migration->data_buffer = g_try_malloc0(migration->data_buffer_size); | 
|  | if (!migration->data_buffer) { | 
|  | error_setg(errp, "%s: Failed to allocate migration data buffer", | 
|  | vbasedev->name); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | migration->event_save_iterate_started = false; | 
|  | migration->event_precopy_empty_hit = false; | 
|  |  | 
|  | if (vfio_precopy_supported(vbasedev)) { | 
|  | switch (migration->device_state) { | 
|  | case VFIO_DEVICE_STATE_RUNNING: | 
|  | ret = vfio_migration_set_state(vbasedev, VFIO_DEVICE_STATE_PRE_COPY, | 
|  | VFIO_DEVICE_STATE_RUNNING, errp); | 
|  | if (ret) { | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | vfio_query_precopy_size(migration); | 
|  |  | 
|  | break; | 
|  | case VFIO_DEVICE_STATE_STOP: | 
|  | /* vfio_save_complete_precopy() will go to STOP_COPY */ | 
|  | break; | 
|  | default: | 
|  | error_setg(errp, "%s: Invalid device state %d", vbasedev->name, | 
|  | migration->device_state); | 
|  | return -EINVAL; | 
|  | } | 
|  | } | 
|  |  | 
|  | trace_vfio_save_setup(vbasedev->name, migration->data_buffer_size); | 
|  |  | 
|  | qemu_put_be64(f, VFIO_MIG_FLAG_END_OF_STATE); | 
|  |  | 
|  | ret = qemu_file_get_error(f); | 
|  | if (ret < 0) { | 
|  | error_setg_errno(errp, -ret, "%s: save setup failed", vbasedev->name); | 
|  | } | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void vfio_save_cleanup(void *opaque) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  | Error *local_err = NULL; | 
|  | int ret; | 
|  |  | 
|  | /* | 
|  | * Changing device state from STOP_COPY to STOP can take time. Do it here, | 
|  | * after migration has completed, so it won't increase downtime. | 
|  | */ | 
|  | if (migration->device_state == VFIO_DEVICE_STATE_STOP_COPY) { | 
|  | ret = vfio_migration_set_state_or_reset(vbasedev, | 
|  | VFIO_DEVICE_STATE_STOP, | 
|  | &local_err); | 
|  | if (ret) { | 
|  | error_report_err(local_err); | 
|  | } | 
|  | } | 
|  |  | 
|  | g_free(migration->data_buffer); | 
|  | migration->data_buffer = NULL; | 
|  | migration->precopy_init_size = 0; | 
|  | migration->precopy_dirty_size = 0; | 
|  | migration->initial_data_sent = false; | 
|  | vfio_migration_cleanup(vbasedev); | 
|  | trace_vfio_save_cleanup(vbasedev->name); | 
|  | } | 
|  |  | 
|  | static void vfio_state_pending_estimate(void *opaque, uint64_t *must_precopy, | 
|  | uint64_t *can_postcopy) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  |  | 
|  | if (!vfio_device_state_is_precopy(vbasedev)) { | 
|  | return; | 
|  | } | 
|  |  | 
|  | *must_precopy += | 
|  | migration->precopy_init_size + migration->precopy_dirty_size; | 
|  |  | 
|  | trace_vfio_state_pending_estimate(vbasedev->name, *must_precopy, | 
|  | *can_postcopy, | 
|  | migration->precopy_init_size, | 
|  | migration->precopy_dirty_size); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Migration size of VFIO devices can be as little as a few KBs or as big as | 
|  | * many GBs. This value should be big enough to cover the worst case. | 
|  | */ | 
|  | #define VFIO_MIG_STOP_COPY_SIZE (100 * GiB) | 
|  |  | 
|  | static void vfio_state_pending_exact(void *opaque, uint64_t *must_precopy, | 
|  | uint64_t *can_postcopy) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  | uint64_t stop_copy_size = VFIO_MIG_STOP_COPY_SIZE; | 
|  |  | 
|  | /* | 
|  | * If getting pending migration size fails, VFIO_MIG_STOP_COPY_SIZE is | 
|  | * reported so downtime limit won't be violated. | 
|  | */ | 
|  | vfio_query_stop_copy_size(vbasedev, &stop_copy_size); | 
|  | *must_precopy += stop_copy_size; | 
|  |  | 
|  | if (vfio_device_state_is_precopy(vbasedev)) { | 
|  | vfio_query_precopy_size(migration); | 
|  | } | 
|  |  | 
|  | trace_vfio_state_pending_exact(vbasedev->name, *must_precopy, *can_postcopy, | 
|  | stop_copy_size, migration->precopy_init_size, | 
|  | migration->precopy_dirty_size); | 
|  | } | 
|  |  | 
|  | static bool vfio_is_active_iterate(void *opaque) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  |  | 
|  | return vfio_device_state_is_precopy(vbasedev); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Note about migration rate limiting: VFIO migration buffer size is currently | 
|  | * limited to 1MB, so there is no need to check if migration rate exceeded (as | 
|  | * in the worst case it will exceed by 1MB). However, if the buffer size is | 
|  | * later changed to a bigger value, migration rate should be enforced here. | 
|  | */ | 
|  | static int vfio_save_iterate(QEMUFile *f, void *opaque) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  | ssize_t data_size; | 
|  |  | 
|  | if (!migration->event_save_iterate_started) { | 
|  | trace_vfio_save_iterate_start(vbasedev->name); | 
|  | migration->event_save_iterate_started = true; | 
|  | } | 
|  |  | 
|  | data_size = vfio_save_block(f, migration); | 
|  | if (data_size < 0) { | 
|  | return data_size; | 
|  | } | 
|  |  | 
|  | vfio_update_estimated_pending_data(migration, data_size); | 
|  |  | 
|  | if (migrate_switchover_ack() && !migration->precopy_init_size && | 
|  | !migration->initial_data_sent) { | 
|  | qemu_put_be64(f, VFIO_MIG_FLAG_DEV_INIT_DATA_SENT); | 
|  | migration->initial_data_sent = true; | 
|  | } else { | 
|  | qemu_put_be64(f, VFIO_MIG_FLAG_END_OF_STATE); | 
|  | } | 
|  |  | 
|  | trace_vfio_save_iterate(vbasedev->name, migration->precopy_init_size, | 
|  | migration->precopy_dirty_size); | 
|  |  | 
|  | return !migration->precopy_init_size && !migration->precopy_dirty_size; | 
|  | } | 
|  |  | 
|  | static int vfio_save_complete_precopy(QEMUFile *f, void *opaque) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | ssize_t data_size; | 
|  | int ret; | 
|  | Error *local_err = NULL; | 
|  |  | 
|  | trace_vfio_save_complete_precopy_start(vbasedev->name); | 
|  |  | 
|  | /* We reach here with device state STOP or STOP_COPY only */ | 
|  | ret = vfio_migration_set_state(vbasedev, VFIO_DEVICE_STATE_STOP_COPY, | 
|  | VFIO_DEVICE_STATE_STOP, &local_err); | 
|  | if (ret) { | 
|  | error_report_err(local_err); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | do { | 
|  | data_size = vfio_save_block(f, vbasedev->migration); | 
|  | if (data_size < 0) { | 
|  | return data_size; | 
|  | } | 
|  | } while (data_size); | 
|  |  | 
|  | qemu_put_be64(f, VFIO_MIG_FLAG_END_OF_STATE); | 
|  | ret = qemu_file_get_error(f); | 
|  |  | 
|  | trace_vfio_save_complete_precopy(vbasedev->name, ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void vfio_save_state(QEMUFile *f, void *opaque) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | Error *local_err = NULL; | 
|  | int ret; | 
|  |  | 
|  | ret = vfio_save_device_config_state(f, opaque, &local_err); | 
|  | if (ret) { | 
|  | error_prepend(&local_err, | 
|  | "vfio: Failed to save device config space of %s - ", | 
|  | vbasedev->name); | 
|  | qemu_file_set_error_obj(f, ret, local_err); | 
|  | } | 
|  | } | 
|  |  | 
|  | static int vfio_load_setup(QEMUFile *f, void *opaque, Error **errp) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  |  | 
|  | return vfio_migration_set_state(vbasedev, VFIO_DEVICE_STATE_RESUMING, | 
|  | vbasedev->migration->device_state, errp); | 
|  | } | 
|  |  | 
|  | static int vfio_load_cleanup(void *opaque) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  |  | 
|  | vfio_migration_cleanup(vbasedev); | 
|  | trace_vfio_load_cleanup(vbasedev->name); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int vfio_load_state(QEMUFile *f, void *opaque, int version_id) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | int ret = 0; | 
|  | uint64_t data; | 
|  |  | 
|  | data = qemu_get_be64(f); | 
|  | while (data != VFIO_MIG_FLAG_END_OF_STATE) { | 
|  |  | 
|  | trace_vfio_load_state(vbasedev->name, data); | 
|  |  | 
|  | switch (data) { | 
|  | case VFIO_MIG_FLAG_DEV_CONFIG_STATE: | 
|  | { | 
|  | return vfio_load_device_config_state(f, opaque); | 
|  | } | 
|  | case VFIO_MIG_FLAG_DEV_SETUP_STATE: | 
|  | { | 
|  | data = qemu_get_be64(f); | 
|  | if (data == VFIO_MIG_FLAG_END_OF_STATE) { | 
|  | return ret; | 
|  | } else { | 
|  | error_report("%s: SETUP STATE: EOS not found 0x%"PRIx64, | 
|  | vbasedev->name, data); | 
|  | return -EINVAL; | 
|  | } | 
|  | break; | 
|  | } | 
|  | case VFIO_MIG_FLAG_DEV_DATA_STATE: | 
|  | { | 
|  | uint64_t data_size = qemu_get_be64(f); | 
|  |  | 
|  | if (data_size) { | 
|  | ret = vfio_load_buffer(f, vbasedev, data_size); | 
|  | if (ret < 0) { | 
|  | return ret; | 
|  | } | 
|  | } | 
|  | break; | 
|  | } | 
|  | case VFIO_MIG_FLAG_DEV_INIT_DATA_SENT: | 
|  | { | 
|  | if (!vfio_precopy_supported(vbasedev) || | 
|  | !migrate_switchover_ack()) { | 
|  | error_report("%s: Received INIT_DATA_SENT but switchover ack " | 
|  | "is not used", vbasedev->name); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | ret = qemu_loadvm_approve_switchover(); | 
|  | if (ret) { | 
|  | error_report( | 
|  | "%s: qemu_loadvm_approve_switchover failed, err=%d (%s)", | 
|  | vbasedev->name, ret, strerror(-ret)); | 
|  | } | 
|  |  | 
|  | return ret; | 
|  | } | 
|  | default: | 
|  | error_report("%s: Unknown tag 0x%"PRIx64, vbasedev->name, data); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | data = qemu_get_be64(f); | 
|  | ret = qemu_file_get_error(f); | 
|  | if (ret) { | 
|  | return ret; | 
|  | } | 
|  | } | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static bool vfio_switchover_ack_needed(void *opaque) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  |  | 
|  | return vfio_precopy_supported(vbasedev); | 
|  | } | 
|  |  | 
|  | static const SaveVMHandlers savevm_vfio_handlers = { | 
|  | .save_prepare = vfio_save_prepare, | 
|  | .save_setup = vfio_save_setup, | 
|  | .save_cleanup = vfio_save_cleanup, | 
|  | .state_pending_estimate = vfio_state_pending_estimate, | 
|  | .state_pending_exact = vfio_state_pending_exact, | 
|  | .is_active_iterate = vfio_is_active_iterate, | 
|  | .save_live_iterate = vfio_save_iterate, | 
|  | .save_live_complete_precopy = vfio_save_complete_precopy, | 
|  | .save_state = vfio_save_state, | 
|  | .load_setup = vfio_load_setup, | 
|  | .load_cleanup = vfio_load_cleanup, | 
|  | .load_state = vfio_load_state, | 
|  | .switchover_ack_needed = vfio_switchover_ack_needed, | 
|  | }; | 
|  |  | 
|  | /* ---------------------------------------------------------------------- */ | 
|  |  | 
|  | static void vfio_vmstate_change_prepare(void *opaque, bool running, | 
|  | RunState state) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  | enum vfio_device_mig_state new_state; | 
|  | Error *local_err = NULL; | 
|  | int ret; | 
|  |  | 
|  | new_state = migration->device_state == VFIO_DEVICE_STATE_PRE_COPY ? | 
|  | VFIO_DEVICE_STATE_PRE_COPY_P2P : | 
|  | VFIO_DEVICE_STATE_RUNNING_P2P; | 
|  |  | 
|  | ret = vfio_migration_set_state_or_reset(vbasedev, new_state, &local_err); | 
|  | if (ret) { | 
|  | /* | 
|  | * Migration should be aborted in this case, but vm_state_notify() | 
|  | * currently does not support reporting failures. | 
|  | */ | 
|  | migration_file_set_error(ret, local_err); | 
|  | } | 
|  |  | 
|  | trace_vfio_vmstate_change_prepare(vbasedev->name, running, | 
|  | RunState_str(state), | 
|  | mig_state_to_str(new_state)); | 
|  | } | 
|  |  | 
|  | static void vfio_vmstate_change(void *opaque, bool running, RunState state) | 
|  | { | 
|  | VFIODevice *vbasedev = opaque; | 
|  | enum vfio_device_mig_state new_state; | 
|  | Error *local_err = NULL; | 
|  | int ret; | 
|  |  | 
|  | if (running) { | 
|  | new_state = VFIO_DEVICE_STATE_RUNNING; | 
|  | } else { | 
|  | new_state = | 
|  | (vfio_device_state_is_precopy(vbasedev) && | 
|  | (state == RUN_STATE_FINISH_MIGRATE || state == RUN_STATE_PAUSED)) ? | 
|  | VFIO_DEVICE_STATE_STOP_COPY : | 
|  | VFIO_DEVICE_STATE_STOP; | 
|  | } | 
|  |  | 
|  | ret = vfio_migration_set_state_or_reset(vbasedev, new_state, &local_err); | 
|  | if (ret) { | 
|  | /* | 
|  | * Migration should be aborted in this case, but vm_state_notify() | 
|  | * currently does not support reporting failures. | 
|  | */ | 
|  | migration_file_set_error(ret, local_err); | 
|  | } | 
|  |  | 
|  | trace_vfio_vmstate_change(vbasedev->name, running, RunState_str(state), | 
|  | mig_state_to_str(new_state)); | 
|  | } | 
|  |  | 
|  | static int vfio_migration_state_notifier(NotifierWithReturn *notifier, | 
|  | MigrationEvent *e, Error **errp) | 
|  | { | 
|  | VFIOMigration *migration = container_of(notifier, VFIOMigration, | 
|  | migration_state); | 
|  | VFIODevice *vbasedev = migration->vbasedev; | 
|  | Error *local_err = NULL; | 
|  | int ret; | 
|  |  | 
|  | trace_vfio_migration_state_notifier(vbasedev->name, e->type); | 
|  |  | 
|  | if (e->type == MIG_EVENT_PRECOPY_FAILED) { | 
|  | /* | 
|  | * MigrationNotifyFunc may not return an error code and an Error | 
|  | * object for MIG_EVENT_PRECOPY_FAILED. Hence, report the error | 
|  | * locally and ignore the errp argument. | 
|  | */ | 
|  | ret = vfio_migration_set_state_or_reset(vbasedev, | 
|  | VFIO_DEVICE_STATE_RUNNING, | 
|  | &local_err); | 
|  | if (ret) { | 
|  | error_report_err(local_err); | 
|  | } | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void vfio_migration_free(VFIODevice *vbasedev) | 
|  | { | 
|  | g_free(vbasedev->migration); | 
|  | vbasedev->migration = NULL; | 
|  | } | 
|  |  | 
|  | static int vfio_migration_query_flags(VFIODevice *vbasedev, uint64_t *mig_flags) | 
|  | { | 
|  | uint64_t buf[DIV_ROUND_UP(sizeof(struct vfio_device_feature) + | 
|  | sizeof(struct vfio_device_feature_migration), | 
|  | sizeof(uint64_t))] = {}; | 
|  | struct vfio_device_feature *feature = (struct vfio_device_feature *)buf; | 
|  | struct vfio_device_feature_migration *mig = | 
|  | (struct vfio_device_feature_migration *)feature->data; | 
|  |  | 
|  | feature->argsz = sizeof(buf); | 
|  | feature->flags = VFIO_DEVICE_FEATURE_GET | VFIO_DEVICE_FEATURE_MIGRATION; | 
|  | if (ioctl(vbasedev->fd, VFIO_DEVICE_FEATURE, feature)) { | 
|  | return -errno; | 
|  | } | 
|  |  | 
|  | *mig_flags = mig->flags; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static bool vfio_dma_logging_supported(VFIODevice *vbasedev) | 
|  | { | 
|  | uint64_t buf[DIV_ROUND_UP(sizeof(struct vfio_device_feature), | 
|  | sizeof(uint64_t))] = {}; | 
|  | struct vfio_device_feature *feature = (struct vfio_device_feature *)buf; | 
|  |  | 
|  | feature->argsz = sizeof(buf); | 
|  | feature->flags = VFIO_DEVICE_FEATURE_PROBE | | 
|  | VFIO_DEVICE_FEATURE_DMA_LOGGING_START; | 
|  |  | 
|  | return !ioctl(vbasedev->fd, VFIO_DEVICE_FEATURE, feature); | 
|  | } | 
|  |  | 
|  | static int vfio_migration_init(VFIODevice *vbasedev) | 
|  | { | 
|  | int ret; | 
|  | Object *obj; | 
|  | VFIOMigration *migration; | 
|  | char id[256] = ""; | 
|  | g_autofree char *path = NULL, *oid = NULL; | 
|  | uint64_t mig_flags = 0; | 
|  | VMChangeStateHandler *prepare_cb; | 
|  |  | 
|  | if (!vbasedev->ops->vfio_get_object) { | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | obj = vbasedev->ops->vfio_get_object(vbasedev); | 
|  | if (!obj) { | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | ret = vfio_migration_query_flags(vbasedev, &mig_flags); | 
|  | if (ret) { | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | /* Basic migration functionality must be supported */ | 
|  | if (!(mig_flags & VFIO_MIGRATION_STOP_COPY)) { | 
|  | return -EOPNOTSUPP; | 
|  | } | 
|  |  | 
|  | vbasedev->migration = g_new0(VFIOMigration, 1); | 
|  | migration = vbasedev->migration; | 
|  | migration->vbasedev = vbasedev; | 
|  | migration->device_state = VFIO_DEVICE_STATE_RUNNING; | 
|  | migration->data_fd = -1; | 
|  | migration->mig_flags = mig_flags; | 
|  |  | 
|  | vbasedev->dirty_pages_supported = vfio_dma_logging_supported(vbasedev); | 
|  |  | 
|  | oid = vmstate_if_get_id(VMSTATE_IF(DEVICE(obj))); | 
|  | if (oid) { | 
|  | path = g_strdup_printf("%s/vfio", oid); | 
|  | } else { | 
|  | path = g_strdup("vfio"); | 
|  | } | 
|  | strpadcpy(id, sizeof(id), path, '\0'); | 
|  |  | 
|  | register_savevm_live(id, VMSTATE_INSTANCE_ID_ANY, 1, &savevm_vfio_handlers, | 
|  | vbasedev); | 
|  |  | 
|  | prepare_cb = migration->mig_flags & VFIO_MIGRATION_P2P ? | 
|  | vfio_vmstate_change_prepare : | 
|  | NULL; | 
|  | migration->vm_state = qdev_add_vm_change_state_handler_full( | 
|  | vbasedev->dev, vfio_vmstate_change, prepare_cb, vbasedev); | 
|  | migration_add_notifier(&migration->migration_state, | 
|  | vfio_migration_state_notifier); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void vfio_migration_deinit(VFIODevice *vbasedev) | 
|  | { | 
|  | VFIOMigration *migration = vbasedev->migration; | 
|  |  | 
|  | migration_remove_notifier(&migration->migration_state); | 
|  | qemu_del_vm_change_state_handler(migration->vm_state); | 
|  | unregister_savevm(VMSTATE_IF(vbasedev->dev), "vfio", vbasedev); | 
|  | vfio_migration_free(vbasedev); | 
|  | vfio_unblock_multiple_devices_migration(); | 
|  | } | 
|  |  | 
|  | static int vfio_block_migration(VFIODevice *vbasedev, Error *err, Error **errp) | 
|  | { | 
|  | if (vbasedev->enable_migration == ON_OFF_AUTO_ON) { | 
|  | error_propagate(errp, err); | 
|  | return -EINVAL; | 
|  | } | 
|  |  | 
|  | vbasedev->migration_blocker = error_copy(err); | 
|  | error_free(err); | 
|  |  | 
|  | return migrate_add_blocker_normal(&vbasedev->migration_blocker, errp); | 
|  | } | 
|  |  | 
|  | /* ---------------------------------------------------------------------- */ | 
|  |  | 
|  | int64_t vfio_mig_bytes_transferred(void) | 
|  | { | 
|  | return bytes_transferred; | 
|  | } | 
|  |  | 
|  | void vfio_reset_bytes_transferred(void) | 
|  | { | 
|  | bytes_transferred = 0; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Return true when either migration initialized or blocker registered. | 
|  | * Currently only return false when adding blocker fails which will | 
|  | * de-register vfio device. | 
|  | */ | 
|  | bool vfio_migration_realize(VFIODevice *vbasedev, Error **errp) | 
|  | { | 
|  | Error *err = NULL; | 
|  | int ret; | 
|  |  | 
|  | if (vbasedev->enable_migration == ON_OFF_AUTO_OFF) { | 
|  | error_setg(&err, "%s: Migration is disabled for VFIO device", | 
|  | vbasedev->name); | 
|  | return !vfio_block_migration(vbasedev, err, errp); | 
|  | } | 
|  |  | 
|  | ret = vfio_migration_init(vbasedev); | 
|  | if (ret) { | 
|  | if (ret == -ENOTTY) { | 
|  | error_setg(&err, "%s: VFIO migration is not supported in kernel", | 
|  | vbasedev->name); | 
|  | } else { | 
|  | error_setg(&err, | 
|  | "%s: Migration couldn't be initialized for VFIO device, " | 
|  | "err: %d (%s)", | 
|  | vbasedev->name, ret, strerror(-ret)); | 
|  | } | 
|  |  | 
|  | return !vfio_block_migration(vbasedev, err, errp); | 
|  | } | 
|  |  | 
|  | if ((!vbasedev->dirty_pages_supported || | 
|  | vbasedev->device_dirty_page_tracking == ON_OFF_AUTO_OFF) && | 
|  | !vbasedev->iommu_dirty_tracking) { | 
|  | if (vbasedev->enable_migration == ON_OFF_AUTO_AUTO) { | 
|  | error_setg(&err, | 
|  | "%s: VFIO device doesn't support device and " | 
|  | "IOMMU dirty tracking", vbasedev->name); | 
|  | goto add_blocker; | 
|  | } | 
|  |  | 
|  | warn_report("%s: VFIO device doesn't support device and " | 
|  | "IOMMU dirty tracking", vbasedev->name); | 
|  | } | 
|  |  | 
|  | ret = vfio_block_multiple_devices_migration(vbasedev, errp); | 
|  | if (ret) { | 
|  | goto out_deinit; | 
|  | } | 
|  |  | 
|  | if (vfio_viommu_preset(vbasedev)) { | 
|  | error_setg(&err, "%s: Migration is currently not supported " | 
|  | "with vIOMMU enabled", vbasedev->name); | 
|  | goto add_blocker; | 
|  | } | 
|  |  | 
|  | trace_vfio_migration_realize(vbasedev->name); | 
|  | return true; | 
|  |  | 
|  | add_blocker: | 
|  | ret = vfio_block_migration(vbasedev, err, errp); | 
|  | out_deinit: | 
|  | if (ret) { | 
|  | vfio_migration_deinit(vbasedev); | 
|  | } | 
|  | return !ret; | 
|  | } | 
|  |  | 
|  | void vfio_migration_exit(VFIODevice *vbasedev) | 
|  | { | 
|  | if (vbasedev->migration) { | 
|  | vfio_migration_deinit(vbasedev); | 
|  | } | 
|  |  | 
|  | migrate_del_blocker(&vbasedev->migration_blocker); | 
|  | } |