| #ifndef QEMU_HW_XEN_COMMON_H |
| #define QEMU_HW_XEN_COMMON_H |
| |
| /* |
| * If we have new enough libxenctrl then we do not want/need these compat |
| * interfaces, despite what the user supplied cflags might say. They |
| * must be undefined before including xenctrl.h |
| */ |
| #undef XC_WANT_COMPAT_EVTCHN_API |
| #undef XC_WANT_COMPAT_GNTTAB_API |
| #undef XC_WANT_COMPAT_MAP_FOREIGN_API |
| |
| #include <xenctrl.h> |
| #include <xenstore.h> |
| #include <xen/io/xenbus.h> |
| |
| #include "hw/hw.h" |
| #include "hw/xen/xen.h" |
| #include "hw/pci/pci.h" |
| #include "qemu/queue.h" |
| #include "hw/xen/trace.h" |
| |
| extern xc_interface *xen_xc; |
| |
| /* |
| * We don't support Xen prior to 4.2.0. |
| */ |
| |
| /* Xen 4.2 through 4.6 */ |
| #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40701 |
| |
| typedef xc_interface xenforeignmemory_handle; |
| typedef xc_evtchn xenevtchn_handle; |
| typedef xc_gnttab xengnttab_handle; |
| |
| #define xenevtchn_open(l, f) xc_evtchn_open(l, f); |
| #define xenevtchn_close(h) xc_evtchn_close(h) |
| #define xenevtchn_fd(h) xc_evtchn_fd(h) |
| #define xenevtchn_pending(h) xc_evtchn_pending(h) |
| #define xenevtchn_notify(h, p) xc_evtchn_notify(h, p) |
| #define xenevtchn_bind_interdomain(h, d, p) xc_evtchn_bind_interdomain(h, d, p) |
| #define xenevtchn_unmask(h, p) xc_evtchn_unmask(h, p) |
| #define xenevtchn_unbind(h, p) xc_evtchn_unbind(h, p) |
| |
| #define xengnttab_open(l, f) xc_gnttab_open(l, f) |
| #define xengnttab_close(h) xc_gnttab_close(h) |
| #define xengnttab_set_max_grants(h, n) xc_gnttab_set_max_grants(h, n) |
| #define xengnttab_map_grant_ref(h, d, r, p) xc_gnttab_map_grant_ref(h, d, r, p) |
| #define xengnttab_unmap(h, a, n) xc_gnttab_munmap(h, a, n) |
| #define xengnttab_map_grant_refs(h, c, d, r, p) \ |
| xc_gnttab_map_grant_refs(h, c, d, r, p) |
| #define xengnttab_map_domain_grant_refs(h, c, d, r, p) \ |
| xc_gnttab_map_domain_grant_refs(h, c, d, r, p) |
| |
| #define xenforeignmemory_open(l, f) xen_xc |
| #define xenforeignmemory_close(h) |
| |
| static inline void *xenforeignmemory_map(xc_interface *h, uint32_t dom, |
| int prot, size_t pages, |
| const xen_pfn_t arr[/*pages*/], |
| int err[/*pages*/]) |
| { |
| if (err) |
| return xc_map_foreign_bulk(h, dom, prot, arr, err, pages); |
| else |
| return xc_map_foreign_pages(h, dom, prot, arr, pages); |
| } |
| |
| #define xenforeignmemory_unmap(h, p, s) munmap(p, s * XC_PAGE_SIZE) |
| |
| #else /* CONFIG_XEN_CTRL_INTERFACE_VERSION >= 40701 */ |
| |
| #include <xenevtchn.h> |
| #include <xengnttab.h> |
| #include <xenforeignmemory.h> |
| |
| #endif |
| |
| extern xenforeignmemory_handle *xen_fmem; |
| |
| #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40900 |
| |
| typedef xc_interface xendevicemodel_handle; |
| |
| static inline xendevicemodel_handle *xendevicemodel_open( |
| struct xentoollog_logger *logger, unsigned int open_flags) |
| { |
| return xen_xc; |
| } |
| |
| #if CONFIG_XEN_CTRL_INTERFACE_VERSION >= 40500 |
| |
| static inline int xendevicemodel_create_ioreq_server( |
| xendevicemodel_handle *dmod, domid_t domid, int handle_bufioreq, |
| ioservid_t *id) |
| { |
| return xc_hvm_create_ioreq_server(dmod, domid, handle_bufioreq, |
| id); |
| } |
| |
| static inline int xendevicemodel_get_ioreq_server_info( |
| xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, |
| xen_pfn_t *ioreq_pfn, xen_pfn_t *bufioreq_pfn, |
| evtchn_port_t *bufioreq_port) |
| { |
| return xc_hvm_get_ioreq_server_info(dmod, domid, id, ioreq_pfn, |
| bufioreq_pfn, bufioreq_port); |
| } |
| |
| static inline int xendevicemodel_map_io_range_to_ioreq_server( |
| xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, int is_mmio, |
| uint64_t start, uint64_t end) |
| { |
| return xc_hvm_map_io_range_to_ioreq_server(dmod, domid, id, is_mmio, |
| start, end); |
| } |
| |
| static inline int xendevicemodel_unmap_io_range_from_ioreq_server( |
| xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, int is_mmio, |
| uint64_t start, uint64_t end) |
| { |
| return xc_hvm_unmap_io_range_from_ioreq_server(dmod, domid, id, is_mmio, |
| start, end); |
| } |
| |
| static inline int xendevicemodel_map_pcidev_to_ioreq_server( |
| xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, |
| uint16_t segment, uint8_t bus, uint8_t device, uint8_t function) |
| { |
| return xc_hvm_map_pcidev_to_ioreq_server(dmod, domid, id, segment, |
| bus, device, function); |
| } |
| |
| static inline int xendevicemodel_unmap_pcidev_from_ioreq_server( |
| xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, |
| uint16_t segment, uint8_t bus, uint8_t device, uint8_t function) |
| { |
| return xc_hvm_unmap_pcidev_from_ioreq_server(dmod, domid, id, segment, |
| bus, device, function); |
| } |
| |
| static inline int xendevicemodel_destroy_ioreq_server( |
| xendevicemodel_handle *dmod, domid_t domid, ioservid_t id) |
| { |
| return xc_hvm_destroy_ioreq_server(dmod, domid, id); |
| } |
| |
| static inline int xendevicemodel_set_ioreq_server_state( |
| xendevicemodel_handle *dmod, domid_t domid, ioservid_t id, int enabled) |
| { |
| return xc_hvm_set_ioreq_server_state(dmod, domid, id, enabled); |
| } |
| |
| #endif /* CONFIG_XEN_CTRL_INTERFACE_VERSION >= 40500 */ |
| |
| static inline int xendevicemodel_set_pci_intx_level( |
| xendevicemodel_handle *dmod, domid_t domid, uint16_t segment, |
| uint8_t bus, uint8_t device, uint8_t intx, unsigned int level) |
| { |
| return xc_hvm_set_pci_intx_level(dmod, domid, segment, bus, device, |
| intx, level); |
| } |
| |
| static inline int xendevicemodel_set_isa_irq_level( |
| xendevicemodel_handle *dmod, domid_t domid, uint8_t irq, |
| unsigned int level) |
| { |
| return xc_hvm_set_isa_irq_level(dmod, domid, irq, level); |
| } |
| |
| static inline int xendevicemodel_set_pci_link_route( |
| xendevicemodel_handle *dmod, domid_t domid, uint8_t link, uint8_t irq) |
| { |
| return xc_hvm_set_pci_link_route(dmod, domid, link, irq); |
| } |
| |
| static inline int xendevicemodel_inject_msi( |
| xendevicemodel_handle *dmod, domid_t domid, uint64_t msi_addr, |
| uint32_t msi_data) |
| { |
| return xc_hvm_inject_msi(dmod, domid, msi_addr, msi_data); |
| } |
| |
| static inline int xendevicemodel_track_dirty_vram( |
| xendevicemodel_handle *dmod, domid_t domid, uint64_t first_pfn, |
| uint32_t nr, unsigned long *dirty_bitmap) |
| { |
| return xc_hvm_track_dirty_vram(dmod, domid, first_pfn, nr, |
| dirty_bitmap); |
| } |
| |
| static inline int xendevicemodel_modified_memory( |
| xendevicemodel_handle *dmod, domid_t domid, uint64_t first_pfn, |
| uint32_t nr) |
| { |
| return xc_hvm_modified_memory(dmod, domid, first_pfn, nr); |
| } |
| |
| static inline int xendevicemodel_set_mem_type( |
| xendevicemodel_handle *dmod, domid_t domid, hvmmem_type_t mem_type, |
| uint64_t first_pfn, uint32_t nr) |
| { |
| return xc_hvm_set_mem_type(dmod, domid, mem_type, first_pfn, nr); |
| } |
| |
| static inline int xendevicemodel_restrict( |
| xendevicemodel_handle *dmod, domid_t domid) |
| { |
| errno = ENOTTY; |
| return -1; |
| } |
| |
| static inline int xenforeignmemory_restrict( |
| xenforeignmemory_handle *fmem, domid_t domid) |
| { |
| errno = ENOTTY; |
| return -1; |
| } |
| |
| #else /* CONFIG_XEN_CTRL_INTERFACE_VERSION >= 40900 */ |
| |
| #undef XC_WANT_COMPAT_DEVICEMODEL_API |
| #include <xendevicemodel.h> |
| |
| #endif |
| |
| extern xendevicemodel_handle *xen_dmod; |
| |
| static inline int xen_set_mem_type(domid_t domid, hvmmem_type_t type, |
| uint64_t first_pfn, uint32_t nr) |
| { |
| return xendevicemodel_set_mem_type(xen_dmod, domid, type, first_pfn, |
| nr); |
| } |
| |
| static inline int xen_set_pci_intx_level(domid_t domid, uint16_t segment, |
| uint8_t bus, uint8_t device, |
| uint8_t intx, unsigned int level) |
| { |
| return xendevicemodel_set_pci_intx_level(xen_dmod, domid, segment, bus, |
| device, intx, level); |
| } |
| |
| static inline int xen_set_pci_link_route(domid_t domid, uint8_t link, |
| uint8_t irq) |
| { |
| return xendevicemodel_set_pci_link_route(xen_dmod, domid, link, irq); |
| } |
| |
| static inline int xen_inject_msi(domid_t domid, uint64_t msi_addr, |
| uint32_t msi_data) |
| { |
| return xendevicemodel_inject_msi(xen_dmod, domid, msi_addr, msi_data); |
| } |
| |
| static inline int xen_set_isa_irq_level(domid_t domid, uint8_t irq, |
| unsigned int level) |
| { |
| return xendevicemodel_set_isa_irq_level(xen_dmod, domid, irq, level); |
| } |
| |
| static inline int xen_track_dirty_vram(domid_t domid, uint64_t first_pfn, |
| uint32_t nr, unsigned long *bitmap) |
| { |
| return xendevicemodel_track_dirty_vram(xen_dmod, domid, first_pfn, nr, |
| bitmap); |
| } |
| |
| static inline int xen_modified_memory(domid_t domid, uint64_t first_pfn, |
| uint32_t nr) |
| { |
| return xendevicemodel_modified_memory(xen_dmod, domid, first_pfn, nr); |
| } |
| |
| static inline int xen_restrict(domid_t domid) |
| { |
| int rc; |
| |
| /* Attempt to restrict devicemodel operations */ |
| rc = xendevicemodel_restrict(xen_dmod, domid); |
| trace_xen_domid_restrict(rc ? errno : 0); |
| |
| if (rc < 0) { |
| /* |
| * If errno is ENOTTY then restriction is not implemented so |
| * there's no point in trying to restrict other types of |
| * operation, but it should not be treated as a failure. |
| */ |
| if (errno == ENOTTY) { |
| return 0; |
| } |
| |
| return rc; |
| } |
| |
| /* Restrict foreignmemory operations */ |
| rc = xenforeignmemory_restrict(xen_fmem, domid); |
| trace_xen_domid_restrict(rc ? errno : 0); |
| |
| return rc; |
| } |
| |
| void destroy_hvm_domain(bool reboot); |
| |
| /* shutdown/destroy current domain because of an error */ |
| void xen_shutdown_fatal_error(const char *fmt, ...) GCC_FMT_ATTR(1, 2); |
| |
| #ifdef HVM_PARAM_VMPORT_REGS_PFN |
| static inline int xen_get_vmport_regs_pfn(xc_interface *xc, domid_t dom, |
| xen_pfn_t *vmport_regs_pfn) |
| { |
| int rc; |
| uint64_t value; |
| rc = xc_hvm_param_get(xc, dom, HVM_PARAM_VMPORT_REGS_PFN, &value); |
| if (rc >= 0) { |
| *vmport_regs_pfn = (xen_pfn_t) value; |
| } |
| return rc; |
| } |
| #else |
| static inline int xen_get_vmport_regs_pfn(xc_interface *xc, domid_t dom, |
| xen_pfn_t *vmport_regs_pfn) |
| { |
| return -ENOSYS; |
| } |
| #endif |
| |
| /* Xen before 4.6 */ |
| #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40600 |
| |
| #ifndef HVM_IOREQSRV_BUFIOREQ_ATOMIC |
| #define HVM_IOREQSRV_BUFIOREQ_ATOMIC 2 |
| #endif |
| |
| #endif |
| |
| static inline int xen_get_default_ioreq_server_info(domid_t dom, |
| xen_pfn_t *ioreq_pfn, |
| xen_pfn_t *bufioreq_pfn, |
| evtchn_port_t |
| *bufioreq_evtchn) |
| { |
| unsigned long param; |
| int rc; |
| |
| rc = xc_get_hvm_param(xen_xc, dom, HVM_PARAM_IOREQ_PFN, ¶m); |
| if (rc < 0) { |
| fprintf(stderr, "failed to get HVM_PARAM_IOREQ_PFN\n"); |
| return -1; |
| } |
| |
| *ioreq_pfn = param; |
| |
| rc = xc_get_hvm_param(xen_xc, dom, HVM_PARAM_BUFIOREQ_PFN, ¶m); |
| if (rc < 0) { |
| fprintf(stderr, "failed to get HVM_PARAM_BUFIOREQ_PFN\n"); |
| return -1; |
| } |
| |
| *bufioreq_pfn = param; |
| |
| rc = xc_get_hvm_param(xen_xc, dom, HVM_PARAM_BUFIOREQ_EVTCHN, |
| ¶m); |
| if (rc < 0) { |
| fprintf(stderr, "failed to get HVM_PARAM_BUFIOREQ_EVTCHN\n"); |
| return -1; |
| } |
| |
| *bufioreq_evtchn = param; |
| |
| return 0; |
| } |
| |
| /* Xen before 4.5 */ |
| #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40500 |
| |
| #ifndef HVM_PARAM_BUFIOREQ_EVTCHN |
| #define HVM_PARAM_BUFIOREQ_EVTCHN 26 |
| #endif |
| |
| #define IOREQ_TYPE_PCI_CONFIG 2 |
| |
| typedef uint16_t ioservid_t; |
| |
| static inline void xen_map_memory_section(domid_t dom, |
| ioservid_t ioservid, |
| MemoryRegionSection *section) |
| { |
| } |
| |
| static inline void xen_unmap_memory_section(domid_t dom, |
| ioservid_t ioservid, |
| MemoryRegionSection *section) |
| { |
| } |
| |
| static inline void xen_map_io_section(domid_t dom, |
| ioservid_t ioservid, |
| MemoryRegionSection *section) |
| { |
| } |
| |
| static inline void xen_unmap_io_section(domid_t dom, |
| ioservid_t ioservid, |
| MemoryRegionSection *section) |
| { |
| } |
| |
| static inline void xen_map_pcidev(domid_t dom, |
| ioservid_t ioservid, |
| PCIDevice *pci_dev) |
| { |
| } |
| |
| static inline void xen_unmap_pcidev(domid_t dom, |
| ioservid_t ioservid, |
| PCIDevice *pci_dev) |
| { |
| } |
| |
| static inline void xen_create_ioreq_server(domid_t dom, |
| ioservid_t *ioservid) |
| { |
| } |
| |
| static inline void xen_destroy_ioreq_server(domid_t dom, |
| ioservid_t ioservid) |
| { |
| } |
| |
| static inline int xen_get_ioreq_server_info(domid_t dom, |
| ioservid_t ioservid, |
| xen_pfn_t *ioreq_pfn, |
| xen_pfn_t *bufioreq_pfn, |
| evtchn_port_t *bufioreq_evtchn) |
| { |
| return xen_get_default_ioreq_server_info(dom, ioreq_pfn, |
| bufioreq_pfn, |
| bufioreq_evtchn); |
| } |
| |
| static inline int xen_set_ioreq_server_state(domid_t dom, |
| ioservid_t ioservid, |
| bool enable) |
| { |
| return 0; |
| } |
| |
| /* Xen 4.5 */ |
| #else |
| |
| static bool use_default_ioreq_server; |
| |
| static inline void xen_map_memory_section(domid_t dom, |
| ioservid_t ioservid, |
| MemoryRegionSection *section) |
| { |
| hwaddr start_addr = section->offset_within_address_space; |
| ram_addr_t size = int128_get64(section->size); |
| hwaddr end_addr = start_addr + size - 1; |
| |
| if (use_default_ioreq_server) { |
| return; |
| } |
| |
| trace_xen_map_mmio_range(ioservid, start_addr, end_addr); |
| xendevicemodel_map_io_range_to_ioreq_server(xen_dmod, dom, ioservid, 1, |
| start_addr, end_addr); |
| } |
| |
| static inline void xen_unmap_memory_section(domid_t dom, |
| ioservid_t ioservid, |
| MemoryRegionSection *section) |
| { |
| hwaddr start_addr = section->offset_within_address_space; |
| ram_addr_t size = int128_get64(section->size); |
| hwaddr end_addr = start_addr + size - 1; |
| |
| if (use_default_ioreq_server) { |
| return; |
| } |
| |
| trace_xen_unmap_mmio_range(ioservid, start_addr, end_addr); |
| xendevicemodel_unmap_io_range_from_ioreq_server(xen_dmod, dom, ioservid, |
| 1, start_addr, end_addr); |
| } |
| |
| static inline void xen_map_io_section(domid_t dom, |
| ioservid_t ioservid, |
| MemoryRegionSection *section) |
| { |
| hwaddr start_addr = section->offset_within_address_space; |
| ram_addr_t size = int128_get64(section->size); |
| hwaddr end_addr = start_addr + size - 1; |
| |
| if (use_default_ioreq_server) { |
| return; |
| } |
| |
| trace_xen_map_portio_range(ioservid, start_addr, end_addr); |
| xendevicemodel_map_io_range_to_ioreq_server(xen_dmod, dom, ioservid, 0, |
| start_addr, end_addr); |
| } |
| |
| static inline void xen_unmap_io_section(domid_t dom, |
| ioservid_t ioservid, |
| MemoryRegionSection *section) |
| { |
| hwaddr start_addr = section->offset_within_address_space; |
| ram_addr_t size = int128_get64(section->size); |
| hwaddr end_addr = start_addr + size - 1; |
| |
| if (use_default_ioreq_server) { |
| return; |
| } |
| |
| trace_xen_unmap_portio_range(ioservid, start_addr, end_addr); |
| xendevicemodel_unmap_io_range_from_ioreq_server(xen_dmod, dom, ioservid, |
| 0, start_addr, end_addr); |
| } |
| |
| static inline void xen_map_pcidev(domid_t dom, |
| ioservid_t ioservid, |
| PCIDevice *pci_dev) |
| { |
| if (use_default_ioreq_server) { |
| return; |
| } |
| |
| trace_xen_map_pcidev(ioservid, pci_bus_num(pci_dev->bus), |
| PCI_SLOT(pci_dev->devfn), PCI_FUNC(pci_dev->devfn)); |
| xendevicemodel_map_pcidev_to_ioreq_server(xen_dmod, dom, ioservid, 0, |
| pci_bus_num(pci_dev->bus), |
| PCI_SLOT(pci_dev->devfn), |
| PCI_FUNC(pci_dev->devfn)); |
| } |
| |
| static inline void xen_unmap_pcidev(domid_t dom, |
| ioservid_t ioservid, |
| PCIDevice *pci_dev) |
| { |
| if (use_default_ioreq_server) { |
| return; |
| } |
| |
| trace_xen_unmap_pcidev(ioservid, pci_bus_num(pci_dev->bus), |
| PCI_SLOT(pci_dev->devfn), PCI_FUNC(pci_dev->devfn)); |
| xendevicemodel_unmap_pcidev_from_ioreq_server(xen_dmod, dom, ioservid, 0, |
| pci_bus_num(pci_dev->bus), |
| PCI_SLOT(pci_dev->devfn), |
| PCI_FUNC(pci_dev->devfn)); |
| } |
| |
| static inline void xen_create_ioreq_server(domid_t dom, |
| ioservid_t *ioservid) |
| { |
| int rc = xendevicemodel_create_ioreq_server(xen_dmod, dom, |
| HVM_IOREQSRV_BUFIOREQ_ATOMIC, |
| ioservid); |
| |
| if (rc == 0) { |
| trace_xen_ioreq_server_create(*ioservid); |
| return; |
| } |
| |
| *ioservid = 0; |
| use_default_ioreq_server = true; |
| trace_xen_default_ioreq_server(); |
| } |
| |
| static inline void xen_destroy_ioreq_server(domid_t dom, |
| ioservid_t ioservid) |
| { |
| if (use_default_ioreq_server) { |
| return; |
| } |
| |
| trace_xen_ioreq_server_destroy(ioservid); |
| xendevicemodel_destroy_ioreq_server(xen_dmod, dom, ioservid); |
| } |
| |
| static inline int xen_get_ioreq_server_info(domid_t dom, |
| ioservid_t ioservid, |
| xen_pfn_t *ioreq_pfn, |
| xen_pfn_t *bufioreq_pfn, |
| evtchn_port_t *bufioreq_evtchn) |
| { |
| if (use_default_ioreq_server) { |
| return xen_get_default_ioreq_server_info(dom, ioreq_pfn, |
| bufioreq_pfn, |
| bufioreq_evtchn); |
| } |
| |
| return xendevicemodel_get_ioreq_server_info(xen_dmod, dom, ioservid, |
| ioreq_pfn, bufioreq_pfn, |
| bufioreq_evtchn); |
| } |
| |
| static inline int xen_set_ioreq_server_state(domid_t dom, |
| ioservid_t ioservid, |
| bool enable) |
| { |
| if (use_default_ioreq_server) { |
| return 0; |
| } |
| |
| trace_xen_ioreq_server_state(ioservid, enable); |
| return xendevicemodel_set_ioreq_server_state(xen_dmod, dom, ioservid, |
| enable); |
| } |
| |
| #endif |
| |
| #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40600 |
| static inline int xen_xc_domain_add_to_physmap(xc_interface *xch, uint32_t domid, |
| unsigned int space, |
| unsigned long idx, |
| xen_pfn_t gpfn) |
| { |
| return xc_domain_add_to_physmap(xch, domid, space, idx, gpfn); |
| } |
| #else |
| static inline int xen_xc_domain_add_to_physmap(xc_interface *xch, uint32_t domid, |
| unsigned int space, |
| unsigned long idx, |
| xen_pfn_t gpfn) |
| { |
| /* In Xen 4.6 rc is -1 and errno contains the error value. */ |
| int rc = xc_domain_add_to_physmap(xch, domid, space, idx, gpfn); |
| if (rc == -1) |
| return errno; |
| return rc; |
| } |
| #endif |
| |
| #ifdef CONFIG_XEN_PV_DOMAIN_BUILD |
| #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40700 |
| static inline int xen_domain_create(xc_interface *xc, uint32_t ssidref, |
| xen_domain_handle_t handle, uint32_t flags, |
| uint32_t *pdomid) |
| { |
| return xc_domain_create(xc, ssidref, handle, flags, pdomid); |
| } |
| #else |
| static inline int xen_domain_create(xc_interface *xc, uint32_t ssidref, |
| xen_domain_handle_t handle, uint32_t flags, |
| uint32_t *pdomid) |
| { |
| return xc_domain_create(xc, ssidref, handle, flags, pdomid, NULL); |
| } |
| #endif |
| #endif |
| |
| /* Xen before 4.8 */ |
| |
| #if CONFIG_XEN_CTRL_INTERFACE_VERSION < 40800 |
| |
| |
| typedef void *xengnttab_grant_copy_segment_t; |
| |
| static inline int xengnttab_grant_copy(xengnttab_handle *xgt, uint32_t count, |
| xengnttab_grant_copy_segment_t *segs) |
| { |
| return -ENOSYS; |
| } |
| #endif |
| |
| #endif /* QEMU_HW_XEN_COMMON_H */ |