| /* SPDX-License-Identifier: MIT */ |
| /****************************************************************************** |
| * memory.h |
| * |
| * Memory reservation and information. |
| * |
| * Copyright (c) 2005, Keir Fraser <keir@xensource.com> |
| */ |
| |
| #ifndef __XEN_PUBLIC_MEMORY_H__ |
| #define __XEN_PUBLIC_MEMORY_H__ |
| |
| FILE_LICENCE ( MIT ); |
| |
| #include "xen.h" |
| #include "physdev.h" |
| |
| /* |
| * Increase or decrease the specified domain's memory reservation. Returns the |
| * number of extents successfully allocated or freed. |
| * arg == addr of struct xen_memory_reservation. |
| */ |
| #define XENMEM_increase_reservation 0 |
| #define XENMEM_decrease_reservation 1 |
| #define XENMEM_populate_physmap 6 |
| |
| #if __XEN_INTERFACE_VERSION__ >= 0x00030209 |
| /* |
| * Maximum # bits addressable by the user of the allocated region (e.g., I/O |
| * devices often have a 32-bit limitation even in 64-bit systems). If zero |
| * then the user has no addressing restriction. This field is not used by |
| * XENMEM_decrease_reservation. |
| */ |
| #define XENMEMF_address_bits(x) (x) |
| #define XENMEMF_get_address_bits(x) ((x) & 0xffu) |
| /* NUMA node to allocate from. */ |
| #define XENMEMF_node(x) (((x) + 1) << 8) |
| #define XENMEMF_get_node(x) ((((x) >> 8) - 1) & 0xffu) |
| /* Flag to populate physmap with populate-on-demand entries */ |
| #define XENMEMF_populate_on_demand (1<<16) |
| /* Flag to request allocation only from the node specified */ |
| #define XENMEMF_exact_node_request (1<<17) |
| #define XENMEMF_exact_node(n) (XENMEMF_node(n) | XENMEMF_exact_node_request) |
| /* Flag to indicate the node specified is virtual node */ |
| #define XENMEMF_vnode (1<<18) |
| #endif |
| |
| struct xen_memory_reservation { |
| |
| /* |
| * XENMEM_increase_reservation: |
| * OUT: MFN (*not* GMFN) bases of extents that were allocated |
| * XENMEM_decrease_reservation: |
| * IN: GMFN bases of extents to free |
| * XENMEM_populate_physmap: |
| * IN: GPFN bases of extents to populate with memory |
| * OUT: GMFN bases of extents that were allocated |
| * (NB. This command also updates the mach_to_phys translation table) |
| * XENMEM_claim_pages: |
| * IN: must be zero |
| */ |
| XEN_GUEST_HANDLE(xen_pfn_t) extent_start; |
| |
| /* Number of extents, and size/alignment of each (2^extent_order pages). */ |
| xen_ulong_t nr_extents; |
| unsigned int extent_order; |
| |
| #if __XEN_INTERFACE_VERSION__ >= 0x00030209 |
| /* XENMEMF flags. */ |
| unsigned int mem_flags; |
| #else |
| unsigned int address_bits; |
| #endif |
| |
| /* |
| * Domain whose reservation is being changed. |
| * Unprivileged domains can specify only DOMID_SELF. |
| */ |
| domid_t domid; |
| }; |
| typedef struct xen_memory_reservation xen_memory_reservation_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_memory_reservation_t); |
| |
| /* |
| * An atomic exchange of memory pages. If return code is zero then |
| * @out.extent_list provides GMFNs of the newly-allocated memory. |
| * Returns zero on complete success, otherwise a negative error code. |
| * On complete success then always @nr_exchanged == @in.nr_extents. |
| * On partial success @nr_exchanged indicates how much work was done. |
| * |
| * Note that only PV guests can use this operation. |
| */ |
| #define XENMEM_exchange 11 |
| struct xen_memory_exchange { |
| /* |
| * [IN] Details of memory extents to be exchanged (GMFN bases). |
| * Note that @in.address_bits is ignored and unused. |
| */ |
| struct xen_memory_reservation in; |
| |
| /* |
| * [IN/OUT] Details of new memory extents. |
| * We require that: |
| * 1. @in.domid == @out.domid |
| * 2. @in.nr_extents << @in.extent_order == |
| * @out.nr_extents << @out.extent_order |
| * 3. @in.extent_start and @out.extent_start lists must not overlap |
| * 4. @out.extent_start lists GPFN bases to be populated |
| * 5. @out.extent_start is overwritten with allocated GMFN bases |
| */ |
| struct xen_memory_reservation out; |
| |
| /* |
| * [OUT] Number of input extents that were successfully exchanged: |
| * 1. The first @nr_exchanged input extents were successfully |
| * deallocated. |
| * 2. The corresponding first entries in the output extent list correctly |
| * indicate the GMFNs that were successfully exchanged. |
| * 3. All other input and output extents are untouched. |
| * 4. If not all input exents are exchanged then the return code of this |
| * command will be non-zero. |
| * 5. THIS FIELD MUST BE INITIALISED TO ZERO BY THE CALLER! |
| */ |
| xen_ulong_t nr_exchanged; |
| }; |
| typedef struct xen_memory_exchange xen_memory_exchange_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_memory_exchange_t); |
| |
| /* |
| * Returns the maximum machine frame number of mapped RAM in this system. |
| * This command always succeeds (it never returns an error code). |
| * arg == NULL. |
| */ |
| #define XENMEM_maximum_ram_page 2 |
| |
| struct xen_memory_domain { |
| /* [IN] Domain information is being queried for. */ |
| domid_t domid; |
| }; |
| |
| /* |
| * Returns the current or maximum memory reservation, in pages, of the |
| * specified domain (may be DOMID_SELF). Returns -ve errcode on failure. |
| * arg == addr of struct xen_memory_domain. |
| */ |
| #define XENMEM_current_reservation 3 |
| #define XENMEM_maximum_reservation 4 |
| |
| /* |
| * Returns the maximum GFN in use by the specified domain (may be DOMID_SELF). |
| * Returns -ve errcode on failure. |
| * arg == addr of struct xen_memory_domain. |
| */ |
| #define XENMEM_maximum_gpfn 14 |
| |
| /* |
| * Returns a list of MFN bases of 2MB extents comprising the machine_to_phys |
| * mapping table. Architectures which do not have a m2p table do not implement |
| * this command. |
| * arg == addr of xen_machphys_mfn_list_t. |
| */ |
| #define XENMEM_machphys_mfn_list 5 |
| struct xen_machphys_mfn_list { |
| /* |
| * Size of the 'extent_start' array. Fewer entries will be filled if the |
| * machphys table is smaller than max_extents * 2MB. |
| */ |
| unsigned int max_extents; |
| |
| /* |
| * Pointer to buffer to fill with list of extent starts. If there are |
| * any large discontiguities in the machine address space, 2MB gaps in |
| * the machphys table will be represented by an MFN base of zero. |
| */ |
| XEN_GUEST_HANDLE(xen_pfn_t) extent_start; |
| |
| /* |
| * Number of extents written to the above array. This will be smaller |
| * than 'max_extents' if the machphys table is smaller than max_e * 2MB. |
| */ |
| unsigned int nr_extents; |
| }; |
| typedef struct xen_machphys_mfn_list xen_machphys_mfn_list_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_machphys_mfn_list_t); |
| |
| /* |
| * For a compat caller, this is identical to XENMEM_machphys_mfn_list. |
| * |
| * For a non compat caller, this functions similarly to |
| * XENMEM_machphys_mfn_list, but returns the mfns making up the compatibility |
| * m2p table. |
| */ |
| #define XENMEM_machphys_compat_mfn_list 25 |
| |
| /* |
| * Returns the location in virtual address space of the machine_to_phys |
| * mapping table. Architectures which do not have a m2p table, or which do not |
| * map it by default into guest address space, do not implement this command. |
| * arg == addr of xen_machphys_mapping_t. |
| */ |
| #define XENMEM_machphys_mapping 12 |
| struct xen_machphys_mapping { |
| xen_ulong_t v_start, v_end; /* Start and end virtual addresses. */ |
| xen_ulong_t max_mfn; /* Maximum MFN that can be looked up. */ |
| }; |
| typedef struct xen_machphys_mapping xen_machphys_mapping_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_machphys_mapping_t); |
| |
| /* Source mapping space. */ |
| /* ` enum phys_map_space { */ |
| #define XENMAPSPACE_shared_info 0 /* shared info page */ |
| #define XENMAPSPACE_grant_table 1 /* grant table page */ |
| #define XENMAPSPACE_gmfn 2 /* GMFN */ |
| #define XENMAPSPACE_gmfn_range 3 /* GMFN range, XENMEM_add_to_physmap only. */ |
| #define XENMAPSPACE_gmfn_foreign 4 /* GMFN from another dom, |
| * XENMEM_add_to_physmap_batch only. */ |
| #define XENMAPSPACE_dev_mmio 5 /* device mmio region |
| ARM only; the region is mapped in |
| Stage-2 using the Normal Memory |
| Inner/Outer Write-Back Cacheable |
| memory attribute. */ |
| /* ` } */ |
| |
| /* |
| * Sets the GPFN at which a particular page appears in the specified guest's |
| * physical address space (translated guests only). |
| * arg == addr of xen_add_to_physmap_t. |
| */ |
| #define XENMEM_add_to_physmap 7 |
| struct xen_add_to_physmap { |
| /* Which domain to change the mapping for. */ |
| domid_t domid; |
| |
| /* Number of pages to go through for gmfn_range */ |
| uint16_t size; |
| |
| unsigned int space; /* => enum phys_map_space */ |
| |
| #define XENMAPIDX_grant_table_status 0x80000000 |
| |
| /* Index into space being mapped. */ |
| xen_ulong_t idx; |
| |
| /* GPFN in domid where the source mapping page should appear. */ |
| xen_pfn_t gpfn; |
| }; |
| typedef struct xen_add_to_physmap xen_add_to_physmap_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_t); |
| |
| /* A batched version of add_to_physmap. */ |
| #define XENMEM_add_to_physmap_batch 23 |
| struct xen_add_to_physmap_batch { |
| /* IN */ |
| /* Which domain to change the mapping for. */ |
| domid_t domid; |
| uint16_t space; /* => enum phys_map_space */ |
| |
| /* Number of pages to go through */ |
| uint16_t size; |
| |
| #if __XEN_INTERFACE_VERSION__ < 0x00040700 |
| domid_t foreign_domid; /* IFF gmfn_foreign. Should be 0 for other spaces. */ |
| #else |
| union xen_add_to_physmap_batch_extra { |
| domid_t foreign_domid; /* gmfn_foreign */ |
| uint16_t res0; /* All the other spaces. Should be 0 */ |
| } u; |
| #endif |
| |
| /* Indexes into space being mapped. */ |
| XEN_GUEST_HANDLE(xen_ulong_t) idxs; |
| |
| /* GPFN in domid where the source mapping page should appear. */ |
| XEN_GUEST_HANDLE(xen_pfn_t) gpfns; |
| |
| /* OUT */ |
| |
| /* Per index error code. */ |
| XEN_GUEST_HANDLE(int) errs; |
| }; |
| typedef struct xen_add_to_physmap_batch xen_add_to_physmap_batch_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_batch_t); |
| |
| #if __XEN_INTERFACE_VERSION__ < 0x00040400 |
| #define XENMEM_add_to_physmap_range XENMEM_add_to_physmap_batch |
| #define xen_add_to_physmap_range xen_add_to_physmap_batch |
| typedef struct xen_add_to_physmap_batch xen_add_to_physmap_range_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_range_t); |
| #endif |
| |
| /* |
| * Unmaps the page appearing at a particular GPFN from the specified guest's |
| * physical address space (translated guests only). |
| * arg == addr of xen_remove_from_physmap_t. |
| */ |
| #define XENMEM_remove_from_physmap 15 |
| struct xen_remove_from_physmap { |
| /* Which domain to change the mapping for. */ |
| domid_t domid; |
| |
| /* GPFN of the current mapping of the page. */ |
| xen_pfn_t gpfn; |
| }; |
| typedef struct xen_remove_from_physmap xen_remove_from_physmap_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_remove_from_physmap_t); |
| |
| /*** REMOVED ***/ |
| /*#define XENMEM_translate_gpfn_list 8*/ |
| |
| /* |
| * Returns the pseudo-physical memory map as it was when the domain |
| * was started (specified by XENMEM_set_memory_map). |
| * arg == addr of xen_memory_map_t. |
| */ |
| #define XENMEM_memory_map 9 |
| struct xen_memory_map { |
| /* |
| * On call the number of entries which can be stored in buffer. On |
| * return the number of entries which have been stored in |
| * buffer. |
| */ |
| unsigned int nr_entries; |
| |
| /* |
| * Entries in the buffer are in the same format as returned by the |
| * BIOS INT 0x15 EAX=0xE820 call. |
| */ |
| XEN_GUEST_HANDLE(void) buffer; |
| }; |
| typedef struct xen_memory_map xen_memory_map_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_memory_map_t); |
| |
| /* |
| * Returns the real physical memory map. Passes the same structure as |
| * XENMEM_memory_map. |
| * Specifying buffer as NULL will return the number of entries required |
| * to store the complete memory map. |
| * arg == addr of xen_memory_map_t. |
| */ |
| #define XENMEM_machine_memory_map 10 |
| |
| /* |
| * Set the pseudo-physical memory map of a domain, as returned by |
| * XENMEM_memory_map. |
| * arg == addr of xen_foreign_memory_map_t. |
| */ |
| #define XENMEM_set_memory_map 13 |
| struct xen_foreign_memory_map { |
| domid_t domid; |
| struct xen_memory_map map; |
| }; |
| typedef struct xen_foreign_memory_map xen_foreign_memory_map_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_foreign_memory_map_t); |
| |
| #define XENMEM_set_pod_target 16 |
| #define XENMEM_get_pod_target 17 |
| struct xen_pod_target { |
| /* IN */ |
| uint64_t target_pages; |
| /* OUT */ |
| uint64_t tot_pages; |
| uint64_t pod_cache_pages; |
| uint64_t pod_entries; |
| /* IN */ |
| domid_t domid; |
| }; |
| typedef struct xen_pod_target xen_pod_target_t; |
| |
| #if defined(__XEN__) || defined(__XEN_TOOLS__) |
| |
| #ifndef uint64_aligned_t |
| #define uint64_aligned_t uint64_t |
| #endif |
| |
| /* |
| * Get the number of MFNs saved through memory sharing. |
| * The call never fails. |
| */ |
| #define XENMEM_get_sharing_freed_pages 18 |
| #define XENMEM_get_sharing_shared_pages 19 |
| |
| #define XENMEM_paging_op 20 |
| #define XENMEM_paging_op_nominate 0 |
| #define XENMEM_paging_op_evict 1 |
| #define XENMEM_paging_op_prep 2 |
| |
| struct xen_mem_paging_op { |
| uint8_t op; /* XENMEM_paging_op_* */ |
| domid_t domain; |
| |
| /* IN: (XENMEM_paging_op_prep) buffer to immediately fill page from */ |
| XEN_GUEST_HANDLE_64(const_uint8) buffer; |
| /* IN: gfn of page being operated on */ |
| uint64_aligned_t gfn; |
| }; |
| typedef struct xen_mem_paging_op xen_mem_paging_op_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_mem_paging_op_t); |
| |
| #define XENMEM_access_op 21 |
| #define XENMEM_access_op_set_access 0 |
| #define XENMEM_access_op_get_access 1 |
| /* |
| * XENMEM_access_op_enable_emulate and XENMEM_access_op_disable_emulate are |
| * currently unused, but since they have been in use please do not reuse them. |
| * |
| * #define XENMEM_access_op_enable_emulate 2 |
| * #define XENMEM_access_op_disable_emulate 3 |
| */ |
| #define XENMEM_access_op_set_access_multi 4 |
| |
| typedef enum { |
| XENMEM_access_n, |
| XENMEM_access_r, |
| XENMEM_access_w, |
| XENMEM_access_rw, |
| XENMEM_access_x, |
| XENMEM_access_rx, |
| XENMEM_access_wx, |
| XENMEM_access_rwx, |
| /* |
| * Page starts off as r-x, but automatically |
| * change to r-w on a write |
| */ |
| XENMEM_access_rx2rw, |
| /* |
| * Log access: starts off as n, automatically |
| * goes to rwx, generating an event without |
| * pausing the vcpu |
| */ |
| XENMEM_access_n2rwx, |
| /* Take the domain default */ |
| XENMEM_access_default |
| } xenmem_access_t; |
| |
| struct xen_mem_access_op { |
| /* XENMEM_access_op_* */ |
| uint8_t op; |
| /* xenmem_access_t */ |
| uint8_t access; |
| domid_t domid; |
| /* |
| * Number of pages for set op (or size of pfn_list for |
| * XENMEM_access_op_set_access_multi) |
| * Ignored on setting default access and other ops |
| */ |
| uint32_t nr; |
| /* |
| * First pfn for set op |
| * pfn for get op |
| * ~0ull is used to set and get the default access for pages |
| */ |
| uint64_aligned_t pfn; |
| /* |
| * List of pfns to set access for |
| * Used only with XENMEM_access_op_set_access_multi |
| */ |
| XEN_GUEST_HANDLE(const_uint64) pfn_list; |
| /* |
| * Corresponding list of access settings for pfn_list |
| * Used only with XENMEM_access_op_set_access_multi |
| */ |
| XEN_GUEST_HANDLE(const_uint8) access_list; |
| }; |
| typedef struct xen_mem_access_op xen_mem_access_op_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_mem_access_op_t); |
| |
| #define XENMEM_sharing_op 22 |
| #define XENMEM_sharing_op_nominate_gfn 0 |
| #define XENMEM_sharing_op_nominate_gref 1 |
| #define XENMEM_sharing_op_share 2 |
| #define XENMEM_sharing_op_debug_gfn 3 |
| #define XENMEM_sharing_op_debug_mfn 4 |
| #define XENMEM_sharing_op_debug_gref 5 |
| #define XENMEM_sharing_op_add_physmap 6 |
| #define XENMEM_sharing_op_audit 7 |
| #define XENMEM_sharing_op_range_share 8 |
| #define XENMEM_sharing_op_fork 9 |
| #define XENMEM_sharing_op_fork_reset 10 |
| |
| #define XENMEM_SHARING_OP_S_HANDLE_INVALID (-10) |
| #define XENMEM_SHARING_OP_C_HANDLE_INVALID (-9) |
| |
| /* The following allows sharing of grant refs. This is useful |
| * for sharing utilities sitting as "filters" in IO backends |
| * (e.g. memshr + blktap(2)). The IO backend is only exposed |
| * to grant references, and this allows sharing of the grefs */ |
| #define XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG (xen_mk_ullong(1) << 62) |
| |
| #define XENMEM_SHARING_OP_FIELD_MAKE_GREF(field, val) \ |
| (field) = (XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG | (val)) |
| #define XENMEM_SHARING_OP_FIELD_IS_GREF(field) \ |
| ((field) & XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG) |
| #define XENMEM_SHARING_OP_FIELD_GET_GREF(field) \ |
| ((field) & (~XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG)) |
| |
| struct xen_mem_sharing_op { |
| uint8_t op; /* XENMEM_sharing_op_* */ |
| domid_t domain; |
| |
| union { |
| struct mem_sharing_op_nominate { /* OP_NOMINATE_xxx */ |
| union { |
| uint64_aligned_t gfn; /* IN: gfn to nominate */ |
| uint32_t grant_ref; /* IN: grant ref to nominate */ |
| } u; |
| uint64_aligned_t handle; /* OUT: the handle */ |
| } nominate; |
| struct mem_sharing_op_share { /* OP_SHARE/ADD_PHYSMAP */ |
| uint64_aligned_t source_gfn; /* IN: the gfn of the source page */ |
| uint64_aligned_t source_handle; /* IN: handle to the source page */ |
| uint64_aligned_t client_gfn; /* IN: the client gfn */ |
| uint64_aligned_t client_handle; /* IN: handle to the client page */ |
| domid_t client_domain; /* IN: the client domain id */ |
| } share; |
| struct mem_sharing_op_range { /* OP_RANGE_SHARE */ |
| uint64_aligned_t first_gfn; /* IN: the first gfn */ |
| uint64_aligned_t last_gfn; /* IN: the last gfn */ |
| uint64_aligned_t opaque; /* Must be set to 0 */ |
| domid_t client_domain; /* IN: the client domain id */ |
| uint16_t _pad[3]; /* Must be set to 0 */ |
| } range; |
| struct mem_sharing_op_debug { /* OP_DEBUG_xxx */ |
| union { |
| uint64_aligned_t gfn; /* IN: gfn to debug */ |
| uint64_aligned_t mfn; /* IN: mfn to debug */ |
| uint32_t gref; /* IN: gref to debug */ |
| } u; |
| } debug; |
| struct mem_sharing_op_fork { /* OP_FORK{,_RESET} */ |
| domid_t parent_domain; /* IN: parent's domain id */ |
| /* Only makes sense for short-lived forks */ |
| #define XENMEM_FORK_WITH_IOMMU_ALLOWED (1u << 0) |
| /* Only makes sense for short-lived forks */ |
| #define XENMEM_FORK_BLOCK_INTERRUPTS (1u << 1) |
| #define XENMEM_FORK_RESET_STATE (1u << 2) |
| #define XENMEM_FORK_RESET_MEMORY (1u << 3) |
| uint16_t flags; /* IN: optional settings */ |
| uint32_t pad; /* Must be set to 0 */ |
| } fork; |
| } u; |
| }; |
| typedef struct xen_mem_sharing_op xen_mem_sharing_op_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_mem_sharing_op_t); |
| |
| /* |
| * Attempt to stake a claim for a domain on a quantity of pages |
| * of system RAM, but _not_ assign specific pageframes. Only |
| * arithmetic is performed so the hypercall is very fast and need |
| * not be preemptible, thus sidestepping time-of-check-time-of-use |
| * races for memory allocation. Returns 0 if the hypervisor page |
| * allocator has atomically and successfully claimed the requested |
| * number of pages, else non-zero. |
| * |
| * Any domain may have only one active claim. When sufficient memory |
| * has been allocated to resolve the claim, the claim silently expires. |
| * Claiming zero pages effectively resets any outstanding claim and |
| * is always successful. |
| * |
| * Note that a valid claim may be staked even after memory has been |
| * allocated for a domain. In this case, the claim is not incremental, |
| * i.e. if the domain's total page count is 3, and a claim is staked |
| * for 10, only 7 additional pages are claimed. |
| * |
| * Caller must be privileged or the hypercall fails. |
| */ |
| #define XENMEM_claim_pages 24 |
| |
| /* |
| * XENMEM_claim_pages flags - the are no flags at this time. |
| * The zero value is appropriate. |
| */ |
| |
| /* |
| * With some legacy devices, certain guest-physical addresses cannot safely |
| * be used for other purposes, e.g. to map guest RAM. This hypercall |
| * enumerates those regions so the toolstack can avoid using them. |
| */ |
| #define XENMEM_reserved_device_memory_map 27 |
| struct xen_reserved_device_memory { |
| xen_pfn_t start_pfn; |
| xen_ulong_t nr_pages; |
| }; |
| typedef struct xen_reserved_device_memory xen_reserved_device_memory_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_reserved_device_memory_t); |
| |
| struct xen_reserved_device_memory_map { |
| #define XENMEM_RDM_ALL 1 /* Request all regions (ignore dev union). */ |
| /* IN */ |
| uint32_t flags; |
| /* |
| * IN/OUT |
| * |
| * Gets set to the required number of entries when too low, |
| * signaled by error code -ERANGE. |
| */ |
| unsigned int nr_entries; |
| /* OUT */ |
| XEN_GUEST_HANDLE(xen_reserved_device_memory_t) buffer; |
| /* IN */ |
| union { |
| physdev_pci_device_t pci; |
| } dev; |
| }; |
| typedef struct xen_reserved_device_memory_map xen_reserved_device_memory_map_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_reserved_device_memory_map_t); |
| |
| #endif /* defined(__XEN__) || defined(__XEN_TOOLS__) */ |
| |
| /* |
| * Get the pages for a particular guest resource, so that they can be |
| * mapped directly by a tools domain. |
| */ |
| #define XENMEM_acquire_resource 28 |
| struct xen_mem_acquire_resource { |
| /* IN - The domain whose resource is to be mapped */ |
| domid_t domid; |
| /* IN - the type of resource */ |
| uint16_t type; |
| |
| #define XENMEM_resource_ioreq_server 0 |
| #define XENMEM_resource_grant_table 1 |
| #define XENMEM_resource_vmtrace_buf 2 |
| |
| /* |
| * IN - a type-specific resource identifier, which must be zero |
| * unless stated otherwise. |
| * |
| * type == XENMEM_resource_ioreq_server -> id == ioreq server id |
| * type == XENMEM_resource_grant_table -> id defined below |
| */ |
| uint32_t id; |
| |
| #define XENMEM_resource_grant_table_id_shared 0 |
| #define XENMEM_resource_grant_table_id_status 1 |
| |
| /* |
| * IN/OUT |
| * |
| * As an IN parameter number of frames of the resource to be mapped. |
| * This value may be updated over the course of the operation. |
| * |
| * When frame_list is NULL and nr_frames is 0, this is interpreted as a |
| * request for the size of the resource, which shall be returned in the |
| * nr_frames field. |
| * |
| * The size of a resource will never be zero, but a nonzero result doesn't |
| * guarantee that a subsequent mapping request will be successful. There |
| * are further type/id specific constraints which may change between the |
| * two calls. |
| */ |
| uint32_t nr_frames; |
| /* |
| * Padding field, must be zero on input. |
| * In a previous version this was an output field with the lowest bit |
| * named XENMEM_rsrc_acq_caller_owned. Future versions of this interface |
| * will not reuse this bit as an output with the field being zero on |
| * input. |
| */ |
| uint32_t pad; |
| /* |
| * IN - the index of the initial frame to be mapped. This parameter |
| * is ignored if nr_frames is 0. This value may be updated |
| * over the course of the operation. |
| */ |
| uint64_t frame; |
| |
| #define XENMEM_resource_ioreq_server_frame_bufioreq 0 |
| #define XENMEM_resource_ioreq_server_frame_ioreq(n) (1 + (n)) |
| |
| /* |
| * IN/OUT - If the tools domain is PV then, upon return, frame_list |
| * will be populated with the MFNs of the resource. |
| * If the tools domain is HVM then it is expected that, on |
| * entry, frame_list will be populated with a list of GFNs |
| * that will be mapped to the MFNs of the resource. |
| * If -EIO is returned then the frame_list has only been |
| * partially mapped and it is up to the caller to unmap all |
| * the GFNs. |
| * This parameter may be NULL if nr_frames is 0. This |
| * value may be updated over the course of the operation. |
| */ |
| XEN_GUEST_HANDLE(xen_pfn_t) frame_list; |
| }; |
| typedef struct xen_mem_acquire_resource xen_mem_acquire_resource_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_mem_acquire_resource_t); |
| |
| /* |
| * XENMEM_get_vnumainfo used by guest to get |
| * vNUMA topology from hypervisor. |
| */ |
| #define XENMEM_get_vnumainfo 26 |
| |
| /* vNUMA node memory ranges */ |
| struct xen_vmemrange { |
| uint64_t start, end; |
| unsigned int flags; |
| unsigned int nid; |
| }; |
| typedef struct xen_vmemrange xen_vmemrange_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_vmemrange_t); |
| |
| /* |
| * vNUMA topology specifies vNUMA node number, distance table, |
| * memory ranges and vcpu mapping provided for guests. |
| * XENMEM_get_vnumainfo hypercall expects to see from guest |
| * nr_vnodes, nr_vmemranges and nr_vcpus to indicate available memory. |
| * After filling guests structures, nr_vnodes, nr_vmemranges and nr_vcpus |
| * copied back to guest. Domain returns expected values of nr_vnodes, |
| * nr_vmemranges and nr_vcpus to guest if the values where incorrect. |
| */ |
| struct xen_vnuma_topology_info { |
| /* IN */ |
| domid_t domid; |
| uint16_t pad; |
| /* IN/OUT */ |
| unsigned int nr_vnodes; |
| unsigned int nr_vcpus; |
| unsigned int nr_vmemranges; |
| /* OUT */ |
| union { |
| XEN_GUEST_HANDLE(uint) h; |
| uint64_t pad; |
| } vdistance; |
| union { |
| XEN_GUEST_HANDLE(uint) h; |
| uint64_t pad; |
| } vcpu_to_vnode; |
| union { |
| XEN_GUEST_HANDLE(xen_vmemrange_t) h; |
| uint64_t pad; |
| } vmemrange; |
| }; |
| typedef struct xen_vnuma_topology_info xen_vnuma_topology_info_t; |
| DEFINE_XEN_GUEST_HANDLE(xen_vnuma_topology_info_t); |
| |
| /* Next available subop number is 29 */ |
| |
| #endif /* __XEN_PUBLIC_MEMORY_H__ */ |
| |
| /* |
| * Local variables: |
| * mode: C |
| * c-file-style: "BSD" |
| * c-basic-offset: 4 |
| * tab-width: 4 |
| * indent-tabs-mode: nil |
| * End: |
| */ |