qemu/include/hw/xen/interface/memory.h
<<
>>
Prefs
   1/******************************************************************************
   2 * memory.h
   3 *
   4 * Memory reservation and information.
   5 *
   6 * Permission is hereby granted, free of charge, to any person obtaining a copy
   7 * of this software and associated documentation files (the "Software"), to
   8 * deal in the Software without restriction, including without limitation the
   9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
  10 * sell copies of the Software, and to permit persons to whom the Software is
  11 * furnished to do so, subject to the following conditions:
  12 *
  13 * The above copyright notice and this permission notice shall be included in
  14 * all copies or substantial portions of the Software.
  15 *
  16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
  17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
  18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
  19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
  20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
  21 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
  22 * DEALINGS IN THE SOFTWARE.
  23 *
  24 * Copyright (c) 2005, Keir Fraser <keir@xensource.com>
  25 */
  26
  27#ifndef __XEN_PUBLIC_MEMORY_H__
  28#define __XEN_PUBLIC_MEMORY_H__
  29
  30#include "xen.h"
  31#include "physdev.h"
  32
  33/*
  34 * Increase or decrease the specified domain's memory reservation. Returns the
  35 * number of extents successfully allocated or freed.
  36 * arg == addr of struct xen_memory_reservation.
  37 */
  38#define XENMEM_increase_reservation 0
  39#define XENMEM_decrease_reservation 1
  40#define XENMEM_populate_physmap     6
  41
  42#if __XEN_INTERFACE_VERSION__ >= 0x00030209
  43/*
  44 * Maximum # bits addressable by the user of the allocated region (e.g., I/O
  45 * devices often have a 32-bit limitation even in 64-bit systems). If zero
  46 * then the user has no addressing restriction. This field is not used by
  47 * XENMEM_decrease_reservation.
  48 */
  49#define XENMEMF_address_bits(x)     (x)
  50#define XENMEMF_get_address_bits(x) ((x) & 0xffu)
  51/* NUMA node to allocate from. */
  52#define XENMEMF_node(x)     (((x) + 1) << 8)
  53#define XENMEMF_get_node(x) ((((x) >> 8) - 1) & 0xffu)
  54/* Flag to populate physmap with populate-on-demand entries */
  55#define XENMEMF_populate_on_demand (1<<16)
  56/* Flag to request allocation only from the node specified */
  57#define XENMEMF_exact_node_request  (1<<17)
  58#define XENMEMF_exact_node(n) (XENMEMF_node(n) | XENMEMF_exact_node_request)
  59/* Flag to indicate the node specified is virtual node */
  60#define XENMEMF_vnode  (1<<18)
  61#endif
  62
  63struct xen_memory_reservation {
  64
  65    /*
  66     * XENMEM_increase_reservation:
  67     *   OUT: MFN (*not* GMFN) bases of extents that were allocated
  68     * XENMEM_decrease_reservation:
  69     *   IN:  GMFN bases of extents to free
  70     * XENMEM_populate_physmap:
  71     *   IN:  GPFN bases of extents to populate with memory
  72     *   OUT: GMFN bases of extents that were allocated
  73     *   (NB. This command also updates the mach_to_phys translation table)
  74     * XENMEM_claim_pages:
  75     *   IN: must be zero
  76     */
  77    XEN_GUEST_HANDLE(xen_pfn_t) extent_start;
  78
  79    /* Number of extents, and size/alignment of each (2^extent_order pages). */
  80    xen_ulong_t    nr_extents;
  81    unsigned int   extent_order;
  82
  83#if __XEN_INTERFACE_VERSION__ >= 0x00030209
  84    /* XENMEMF flags. */
  85    unsigned int   mem_flags;
  86#else
  87    unsigned int   address_bits;
  88#endif
  89
  90    /*
  91     * Domain whose reservation is being changed.
  92     * Unprivileged domains can specify only DOMID_SELF.
  93     */
  94    domid_t        domid;
  95};
  96typedef struct xen_memory_reservation xen_memory_reservation_t;
  97DEFINE_XEN_GUEST_HANDLE(xen_memory_reservation_t);
  98
  99/*
 100 * An atomic exchange of memory pages. If return code is zero then
 101 * @out.extent_list provides GMFNs of the newly-allocated memory.
 102 * Returns zero on complete success, otherwise a negative error code.
 103 * On complete success then always @nr_exchanged == @in.nr_extents.
 104 * On partial success @nr_exchanged indicates how much work was done.
 105 *
 106 * Note that only PV guests can use this operation.
 107 */
 108#define XENMEM_exchange             11
 109struct xen_memory_exchange {
 110    /*
 111     * [IN] Details of memory extents to be exchanged (GMFN bases).
 112     * Note that @in.address_bits is ignored and unused.
 113     */
 114    struct xen_memory_reservation in;
 115
 116    /*
 117     * [IN/OUT] Details of new memory extents.
 118     * We require that:
 119     *  1. @in.domid == @out.domid
 120     *  2. @in.nr_extents  << @in.extent_order ==
 121     *     @out.nr_extents << @out.extent_order
 122     *  3. @in.extent_start and @out.extent_start lists must not overlap
 123     *  4. @out.extent_start lists GPFN bases to be populated
 124     *  5. @out.extent_start is overwritten with allocated GMFN bases
 125     */
 126    struct xen_memory_reservation out;
 127
 128    /*
 129     * [OUT] Number of input extents that were successfully exchanged:
 130     *  1. The first @nr_exchanged input extents were successfully
 131     *     deallocated.
 132     *  2. The corresponding first entries in the output extent list correctly
 133     *     indicate the GMFNs that were successfully exchanged.
 134     *  3. All other input and output extents are untouched.
 135     *  4. If not all input exents are exchanged then the return code of this
 136     *     command will be non-zero.
 137     *  5. THIS FIELD MUST BE INITIALISED TO ZERO BY THE CALLER!
 138     */
 139    xen_ulong_t nr_exchanged;
 140};
 141typedef struct xen_memory_exchange xen_memory_exchange_t;
 142DEFINE_XEN_GUEST_HANDLE(xen_memory_exchange_t);
 143
 144/*
 145 * Returns the maximum machine frame number of mapped RAM in this system.
 146 * This command always succeeds (it never returns an error code).
 147 * arg == NULL.
 148 */
 149#define XENMEM_maximum_ram_page     2
 150
 151struct xen_memory_domain {
 152    /* [IN] Domain information is being queried for. */
 153    domid_t domid;
 154};
 155
 156/*
 157 * Returns the current or maximum memory reservation, in pages, of the
 158 * specified domain (may be DOMID_SELF). Returns -ve errcode on failure.
 159 * arg == addr of struct xen_memory_domain.
 160 */
 161#define XENMEM_current_reservation  3
 162#define XENMEM_maximum_reservation  4
 163
 164/*
 165 * Returns the maximum GFN in use by the specified domain (may be DOMID_SELF).
 166 * Returns -ve errcode on failure.
 167 * arg == addr of struct xen_memory_domain.
 168 */
 169#define XENMEM_maximum_gpfn         14
 170
 171/*
 172 * Returns a list of MFN bases of 2MB extents comprising the machine_to_phys
 173 * mapping table. Architectures which do not have a m2p table do not implement
 174 * this command.
 175 * arg == addr of xen_machphys_mfn_list_t.
 176 */
 177#define XENMEM_machphys_mfn_list    5
 178struct xen_machphys_mfn_list {
 179    /*
 180     * Size of the 'extent_start' array. Fewer entries will be filled if the
 181     * machphys table is smaller than max_extents * 2MB.
 182     */
 183    unsigned int max_extents;
 184
 185    /*
 186     * Pointer to buffer to fill with list of extent starts. If there are
 187     * any large discontiguities in the machine address space, 2MB gaps in
 188     * the machphys table will be represented by an MFN base of zero.
 189     */
 190    XEN_GUEST_HANDLE(xen_pfn_t) extent_start;
 191
 192    /*
 193     * Number of extents written to the above array. This will be smaller
 194     * than 'max_extents' if the machphys table is smaller than max_e * 2MB.
 195     */
 196    unsigned int nr_extents;
 197};
 198typedef struct xen_machphys_mfn_list xen_machphys_mfn_list_t;
 199DEFINE_XEN_GUEST_HANDLE(xen_machphys_mfn_list_t);
 200
 201/*
 202 * For a compat caller, this is identical to XENMEM_machphys_mfn_list.
 203 *
 204 * For a non compat caller, this functions similarly to
 205 * XENMEM_machphys_mfn_list, but returns the mfns making up the compatibility
 206 * m2p table.
 207 */
 208#define XENMEM_machphys_compat_mfn_list     25
 209
 210/*
 211 * Returns the location in virtual address space of the machine_to_phys
 212 * mapping table. Architectures which do not have a m2p table, or which do not
 213 * map it by default into guest address space, do not implement this command.
 214 * arg == addr of xen_machphys_mapping_t.
 215 */
 216#define XENMEM_machphys_mapping     12
 217struct xen_machphys_mapping {
 218    xen_ulong_t v_start, v_end; /* Start and end virtual addresses.   */
 219    xen_ulong_t max_mfn;        /* Maximum MFN that can be looked up. */
 220};
 221typedef struct xen_machphys_mapping xen_machphys_mapping_t;
 222DEFINE_XEN_GUEST_HANDLE(xen_machphys_mapping_t);
 223
 224/* Source mapping space. */
 225/* ` enum phys_map_space { */
 226#define XENMAPSPACE_shared_info  0 /* shared info page */
 227#define XENMAPSPACE_grant_table  1 /* grant table page */
 228#define XENMAPSPACE_gmfn         2 /* GMFN */
 229#define XENMAPSPACE_gmfn_range   3 /* GMFN range, XENMEM_add_to_physmap only. */
 230#define XENMAPSPACE_gmfn_foreign 4 /* GMFN from another dom,
 231                                    * XENMEM_add_to_physmap_batch only. */
 232#define XENMAPSPACE_dev_mmio     5 /* device mmio region
 233                                      ARM only; the region is mapped in
 234                                      Stage-2 using the Normal Memory
 235                                      Inner/Outer Write-Back Cacheable
 236                                      memory attribute. */
 237/* ` } */
 238
 239/*
 240 * Sets the GPFN at which a particular page appears in the specified guest's
 241 * physical address space (translated guests only).
 242 * arg == addr of xen_add_to_physmap_t.
 243 */
 244#define XENMEM_add_to_physmap      7
 245struct xen_add_to_physmap {
 246    /* Which domain to change the mapping for. */
 247    domid_t domid;
 248
 249    /* Number of pages to go through for gmfn_range */
 250    uint16_t    size;
 251
 252    unsigned int space; /* => enum phys_map_space */
 253
 254#define XENMAPIDX_grant_table_status 0x80000000
 255
 256    /* Index into space being mapped. */
 257    xen_ulong_t idx;
 258
 259    /* GPFN in domid where the source mapping page should appear. */
 260    xen_pfn_t     gpfn;
 261};
 262typedef struct xen_add_to_physmap xen_add_to_physmap_t;
 263DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_t);
 264
 265/* A batched version of add_to_physmap. */
 266#define XENMEM_add_to_physmap_batch 23
 267struct xen_add_to_physmap_batch {
 268    /* IN */
 269    /* Which domain to change the mapping for. */
 270    domid_t domid;
 271    uint16_t space; /* => enum phys_map_space */
 272
 273    /* Number of pages to go through */
 274    uint16_t size;
 275
 276#if __XEN_INTERFACE_VERSION__ < 0x00040700
 277    domid_t foreign_domid; /* IFF gmfn_foreign. Should be 0 for other spaces. */
 278#else
 279    union xen_add_to_physmap_batch_extra {
 280        domid_t foreign_domid; /* gmfn_foreign */
 281        uint16_t res0;  /* All the other spaces. Should be 0 */
 282    } u;
 283#endif
 284
 285    /* Indexes into space being mapped. */
 286    XEN_GUEST_HANDLE(xen_ulong_t) idxs;
 287
 288    /* GPFN in domid where the source mapping page should appear. */
 289    XEN_GUEST_HANDLE(xen_pfn_t) gpfns;
 290
 291    /* OUT */
 292
 293    /* Per index error code. */
 294    XEN_GUEST_HANDLE(int) errs;
 295};
 296typedef struct xen_add_to_physmap_batch xen_add_to_physmap_batch_t;
 297DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_batch_t);
 298
 299#if __XEN_INTERFACE_VERSION__ < 0x00040400
 300#define XENMEM_add_to_physmap_range XENMEM_add_to_physmap_batch
 301#define xen_add_to_physmap_range xen_add_to_physmap_batch
 302typedef struct xen_add_to_physmap_batch xen_add_to_physmap_range_t;
 303DEFINE_XEN_GUEST_HANDLE(xen_add_to_physmap_range_t);
 304#endif
 305
 306/*
 307 * Unmaps the page appearing at a particular GPFN from the specified guest's
 308 * physical address space (translated guests only).
 309 * arg == addr of xen_remove_from_physmap_t.
 310 */
 311#define XENMEM_remove_from_physmap      15
 312struct xen_remove_from_physmap {
 313    /* Which domain to change the mapping for. */
 314    domid_t domid;
 315
 316    /* GPFN of the current mapping of the page. */
 317    xen_pfn_t     gpfn;
 318};
 319typedef struct xen_remove_from_physmap xen_remove_from_physmap_t;
 320DEFINE_XEN_GUEST_HANDLE(xen_remove_from_physmap_t);
 321
 322/*** REMOVED ***/
 323/*#define XENMEM_translate_gpfn_list  8*/
 324
 325/*
 326 * Returns the pseudo-physical memory map as it was when the domain
 327 * was started (specified by XENMEM_set_memory_map).
 328 * arg == addr of xen_memory_map_t.
 329 */
 330#define XENMEM_memory_map           9
 331struct xen_memory_map {
 332    /*
 333     * On call the number of entries which can be stored in buffer. On
 334     * return the number of entries which have been stored in
 335     * buffer.
 336     */
 337    unsigned int nr_entries;
 338
 339    /*
 340     * Entries in the buffer are in the same format as returned by the
 341     * BIOS INT 0x15 EAX=0xE820 call.
 342     */
 343    XEN_GUEST_HANDLE(void) buffer;
 344};
 345typedef struct xen_memory_map xen_memory_map_t;
 346DEFINE_XEN_GUEST_HANDLE(xen_memory_map_t);
 347
 348/*
 349 * Returns the real physical memory map. Passes the same structure as
 350 * XENMEM_memory_map.
 351 * Specifying buffer as NULL will return the number of entries required
 352 * to store the complete memory map.
 353 * arg == addr of xen_memory_map_t.
 354 */
 355#define XENMEM_machine_memory_map   10
 356
 357/*
 358 * Set the pseudo-physical memory map of a domain, as returned by
 359 * XENMEM_memory_map.
 360 * arg == addr of xen_foreign_memory_map_t.
 361 */
 362#define XENMEM_set_memory_map       13
 363struct xen_foreign_memory_map {
 364    domid_t domid;
 365    struct xen_memory_map map;
 366};
 367typedef struct xen_foreign_memory_map xen_foreign_memory_map_t;
 368DEFINE_XEN_GUEST_HANDLE(xen_foreign_memory_map_t);
 369
 370#define XENMEM_set_pod_target       16
 371#define XENMEM_get_pod_target       17
 372struct xen_pod_target {
 373    /* IN */
 374    uint64_t target_pages;
 375    /* OUT */
 376    uint64_t tot_pages;
 377    uint64_t pod_cache_pages;
 378    uint64_t pod_entries;
 379    /* IN */
 380    domid_t domid;
 381};
 382typedef struct xen_pod_target xen_pod_target_t;
 383
 384#if defined(__XEN__) || defined(__XEN_TOOLS__)
 385
 386#ifndef uint64_aligned_t
 387#define uint64_aligned_t uint64_t
 388#endif
 389
 390/*
 391 * Get the number of MFNs saved through memory sharing.
 392 * The call never fails.
 393 */
 394#define XENMEM_get_sharing_freed_pages    18
 395#define XENMEM_get_sharing_shared_pages   19
 396
 397#define XENMEM_paging_op                    20
 398#define XENMEM_paging_op_nominate           0
 399#define XENMEM_paging_op_evict              1
 400#define XENMEM_paging_op_prep               2
 401
 402struct xen_mem_paging_op {
 403    uint8_t     op;         /* XENMEM_paging_op_* */
 404    domid_t     domain;
 405
 406    /* IN: (XENMEM_paging_op_prep) buffer to immediately fill page from */
 407    XEN_GUEST_HANDLE_64(const_uint8) buffer;
 408    /* IN:  gfn of page being operated on */
 409    uint64_aligned_t    gfn;
 410};
 411typedef struct xen_mem_paging_op xen_mem_paging_op_t;
 412DEFINE_XEN_GUEST_HANDLE(xen_mem_paging_op_t);
 413
 414#define XENMEM_access_op                    21
 415#define XENMEM_access_op_set_access         0
 416#define XENMEM_access_op_get_access         1
 417/*
 418 * XENMEM_access_op_enable_emulate and XENMEM_access_op_disable_emulate are
 419 * currently unused, but since they have been in use please do not reuse them.
 420 *
 421 * #define XENMEM_access_op_enable_emulate     2
 422 * #define XENMEM_access_op_disable_emulate    3
 423 */
 424#define XENMEM_access_op_set_access_multi   4
 425
 426typedef enum {
 427    XENMEM_access_n,
 428    XENMEM_access_r,
 429    XENMEM_access_w,
 430    XENMEM_access_rw,
 431    XENMEM_access_x,
 432    XENMEM_access_rx,
 433    XENMEM_access_wx,
 434    XENMEM_access_rwx,
 435    /*
 436     * Page starts off as r-x, but automatically
 437     * change to r-w on a write
 438     */
 439    XENMEM_access_rx2rw,
 440    /*
 441     * Log access: starts off as n, automatically
 442     * goes to rwx, generating an event without
 443     * pausing the vcpu
 444     */
 445    XENMEM_access_n2rwx,
 446    /* Take the domain default */
 447    XENMEM_access_default
 448} xenmem_access_t;
 449
 450struct xen_mem_access_op {
 451    /* XENMEM_access_op_* */
 452    uint8_t op;
 453    /* xenmem_access_t */
 454    uint8_t access;
 455    domid_t domid;
 456    /*
 457     * Number of pages for set op (or size of pfn_list for
 458     * XENMEM_access_op_set_access_multi)
 459     * Ignored on setting default access and other ops
 460     */
 461    uint32_t nr;
 462    /*
 463     * First pfn for set op
 464     * pfn for get op
 465     * ~0ull is used to set and get the default access for pages
 466     */
 467    uint64_aligned_t pfn;
 468    /*
 469     * List of pfns to set access for
 470     * Used only with XENMEM_access_op_set_access_multi
 471     */
 472    XEN_GUEST_HANDLE(const_uint64) pfn_list;
 473    /*
 474     * Corresponding list of access settings for pfn_list
 475     * Used only with XENMEM_access_op_set_access_multi
 476     */
 477    XEN_GUEST_HANDLE(const_uint8) access_list;
 478};
 479typedef struct xen_mem_access_op xen_mem_access_op_t;
 480DEFINE_XEN_GUEST_HANDLE(xen_mem_access_op_t);
 481
 482#define XENMEM_sharing_op                   22
 483#define XENMEM_sharing_op_nominate_gfn      0
 484#define XENMEM_sharing_op_nominate_gref     1
 485#define XENMEM_sharing_op_share             2
 486#define XENMEM_sharing_op_debug_gfn         3
 487#define XENMEM_sharing_op_debug_mfn         4
 488#define XENMEM_sharing_op_debug_gref        5
 489#define XENMEM_sharing_op_add_physmap       6
 490#define XENMEM_sharing_op_audit             7
 491#define XENMEM_sharing_op_range_share       8
 492#define XENMEM_sharing_op_fork              9
 493#define XENMEM_sharing_op_fork_reset        10
 494
 495#define XENMEM_SHARING_OP_S_HANDLE_INVALID  (-10)
 496#define XENMEM_SHARING_OP_C_HANDLE_INVALID  (-9)
 497
 498/* The following allows sharing of grant refs. This is useful
 499 * for sharing utilities sitting as "filters" in IO backends
 500 * (e.g. memshr + blktap(2)). The IO backend is only exposed
 501 * to grant references, and this allows sharing of the grefs */
 502#define XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG   (xen_mk_ullong(1) << 62)
 503
 504#define XENMEM_SHARING_OP_FIELD_MAKE_GREF(field, val)  \
 505    (field) = (XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG | val)
 506#define XENMEM_SHARING_OP_FIELD_IS_GREF(field)         \
 507    ((field) & XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG)
 508#define XENMEM_SHARING_OP_FIELD_GET_GREF(field)        \
 509    ((field) & (~XENMEM_SHARING_OP_FIELD_IS_GREF_FLAG))
 510
 511struct xen_mem_sharing_op {
 512    uint8_t     op;     /* XENMEM_sharing_op_* */
 513    domid_t     domain;
 514
 515    union {
 516        struct mem_sharing_op_nominate {  /* OP_NOMINATE_xxx           */
 517            union {
 518                uint64_aligned_t gfn;     /* IN: gfn to nominate       */
 519                uint32_t      grant_ref;  /* IN: grant ref to nominate */
 520            } u;
 521            uint64_aligned_t  handle;     /* OUT: the handle           */
 522        } nominate;
 523        struct mem_sharing_op_share {     /* OP_SHARE/ADD_PHYSMAP */
 524            uint64_aligned_t source_gfn;    /* IN: the gfn of the source page */
 525            uint64_aligned_t source_handle; /* IN: handle to the source page */
 526            uint64_aligned_t client_gfn;    /* IN: the client gfn */
 527            uint64_aligned_t client_handle; /* IN: handle to the client page */
 528            domid_t  client_domain; /* IN: the client domain id */
 529        } share;
 530        struct mem_sharing_op_range {         /* OP_RANGE_SHARE */
 531            uint64_aligned_t first_gfn;      /* IN: the first gfn */
 532            uint64_aligned_t last_gfn;       /* IN: the last gfn */
 533            uint64_aligned_t opaque;         /* Must be set to 0 */
 534            domid_t client_domain;           /* IN: the client domain id */
 535            uint16_t _pad[3];                /* Must be set to 0 */
 536        } range;
 537        struct mem_sharing_op_debug {     /* OP_DEBUG_xxx */
 538            union {
 539                uint64_aligned_t gfn;      /* IN: gfn to debug          */
 540                uint64_aligned_t mfn;      /* IN: mfn to debug          */
 541                uint32_t gref;     /* IN: gref to debug         */
 542            } u;
 543        } debug;
 544        struct mem_sharing_op_fork {      /* OP_FORK */
 545            domid_t parent_domain;        /* IN: parent's domain id */
 546/* Only makes sense for short-lived forks */
 547#define XENMEM_FORK_WITH_IOMMU_ALLOWED (1u << 0)
 548/* Only makes sense for short-lived forks */
 549#define XENMEM_FORK_BLOCK_INTERRUPTS   (1u << 1)
 550            uint16_t flags;               /* IN: optional settings */
 551            uint32_t pad;                 /* Must be set to 0 */
 552        } fork;
 553    } u;
 554};
 555typedef struct xen_mem_sharing_op xen_mem_sharing_op_t;
 556DEFINE_XEN_GUEST_HANDLE(xen_mem_sharing_op_t);
 557
 558/*
 559 * Attempt to stake a claim for a domain on a quantity of pages
 560 * of system RAM, but _not_ assign specific pageframes.  Only
 561 * arithmetic is performed so the hypercall is very fast and need
 562 * not be preemptible, thus sidestepping time-of-check-time-of-use
 563 * races for memory allocation.  Returns 0 if the hypervisor page
 564 * allocator has atomically and successfully claimed the requested
 565 * number of pages, else non-zero.
 566 *
 567 * Any domain may have only one active claim.  When sufficient memory
 568 * has been allocated to resolve the claim, the claim silently expires.
 569 * Claiming zero pages effectively resets any outstanding claim and
 570 * is always successful.
 571 *
 572 * Note that a valid claim may be staked even after memory has been
 573 * allocated for a domain.  In this case, the claim is not incremental,
 574 * i.e. if the domain's total page count is 3, and a claim is staked
 575 * for 10, only 7 additional pages are claimed.
 576 *
 577 * Caller must be privileged or the hypercall fails.
 578 */
 579#define XENMEM_claim_pages                  24
 580
 581/*
 582 * XENMEM_claim_pages flags - the are no flags at this time.
 583 * The zero value is appropriate.
 584 */
 585
 586/*
 587 * With some legacy devices, certain guest-physical addresses cannot safely
 588 * be used for other purposes, e.g. to map guest RAM.  This hypercall
 589 * enumerates those regions so the toolstack can avoid using them.
 590 */
 591#define XENMEM_reserved_device_memory_map   27
 592struct xen_reserved_device_memory {
 593    xen_pfn_t start_pfn;
 594    xen_ulong_t nr_pages;
 595};
 596typedef struct xen_reserved_device_memory xen_reserved_device_memory_t;
 597DEFINE_XEN_GUEST_HANDLE(xen_reserved_device_memory_t);
 598
 599struct xen_reserved_device_memory_map {
 600#define XENMEM_RDM_ALL 1 /* Request all regions (ignore dev union). */
 601    /* IN */
 602    uint32_t flags;
 603    /*
 604     * IN/OUT
 605     *
 606     * Gets set to the required number of entries when too low,
 607     * signaled by error code -ERANGE.
 608     */
 609    unsigned int nr_entries;
 610    /* OUT */
 611    XEN_GUEST_HANDLE(xen_reserved_device_memory_t) buffer;
 612    /* IN */
 613    union {
 614        physdev_pci_device_t pci;
 615    } dev;
 616};
 617typedef struct xen_reserved_device_memory_map xen_reserved_device_memory_map_t;
 618DEFINE_XEN_GUEST_HANDLE(xen_reserved_device_memory_map_t);
 619
 620#endif /* defined(__XEN__) || defined(__XEN_TOOLS__) */
 621
 622/*
 623 * Get the pages for a particular guest resource, so that they can be
 624 * mapped directly by a tools domain.
 625 */
 626#define XENMEM_acquire_resource 28
 627struct xen_mem_acquire_resource {
 628    /* IN - The domain whose resource is to be mapped */
 629    domid_t domid;
 630    /* IN - the type of resource */
 631    uint16_t type;
 632
 633#define XENMEM_resource_ioreq_server 0
 634#define XENMEM_resource_grant_table 1
 635#define XENMEM_resource_vmtrace_buf 2
 636
 637    /*
 638     * IN - a type-specific resource identifier, which must be zero
 639     *      unless stated otherwise.
 640     *
 641     * type == XENMEM_resource_ioreq_server -> id == ioreq server id
 642     * type == XENMEM_resource_grant_table -> id defined below
 643     */
 644    uint32_t id;
 645
 646#define XENMEM_resource_grant_table_id_shared 0
 647#define XENMEM_resource_grant_table_id_status 1
 648
 649    /*
 650     * IN/OUT
 651     *
 652     * As an IN parameter number of frames of the resource to be mapped.
 653     * This value may be updated over the course of the operation.
 654     *
 655     * When frame_list is NULL and nr_frames is 0, this is interpreted as a
 656     * request for the size of the resource, which shall be returned in the
 657     * nr_frames field.
 658     *
 659     * The size of a resource will never be zero, but a nonzero result doesn't
 660     * guarantee that a subsequent mapping request will be successful.  There
 661     * are further type/id specific constraints which may change between the
 662     * two calls.
 663     */
 664    uint32_t nr_frames;
 665    uint32_t pad;
 666    /*
 667     * IN - the index of the initial frame to be mapped. This parameter
 668     *      is ignored if nr_frames is 0.  This value may be updated
 669     *      over the course of the operation.
 670     */
 671    uint64_t frame;
 672
 673#define XENMEM_resource_ioreq_server_frame_bufioreq 0
 674#define XENMEM_resource_ioreq_server_frame_ioreq(n) (1 + (n))
 675
 676    /*
 677     * IN/OUT - If the tools domain is PV then, upon return, frame_list
 678     *          will be populated with the MFNs of the resource.
 679     *          If the tools domain is HVM then it is expected that, on
 680     *          entry, frame_list will be populated with a list of GFNs
 681     *          that will be mapped to the MFNs of the resource.
 682     *          If -EIO is returned then the frame_list has only been
 683     *          partially mapped and it is up to the caller to unmap all
 684     *          the GFNs.
 685     *          This parameter may be NULL if nr_frames is 0.  This
 686     *          value may be updated over the course of the operation.
 687     */
 688    XEN_GUEST_HANDLE(xen_pfn_t) frame_list;
 689};
 690typedef struct xen_mem_acquire_resource xen_mem_acquire_resource_t;
 691DEFINE_XEN_GUEST_HANDLE(xen_mem_acquire_resource_t);
 692
 693/*
 694 * XENMEM_get_vnumainfo used by guest to get
 695 * vNUMA topology from hypervisor.
 696 */
 697#define XENMEM_get_vnumainfo                26
 698
 699/* vNUMA node memory ranges */
 700struct xen_vmemrange {
 701    uint64_t start, end;
 702    unsigned int flags;
 703    unsigned int nid;
 704};
 705typedef struct xen_vmemrange xen_vmemrange_t;
 706DEFINE_XEN_GUEST_HANDLE(xen_vmemrange_t);
 707
 708/*
 709 * vNUMA topology specifies vNUMA node number, distance table,
 710 * memory ranges and vcpu mapping provided for guests.
 711 * XENMEM_get_vnumainfo hypercall expects to see from guest
 712 * nr_vnodes, nr_vmemranges and nr_vcpus to indicate available memory.
 713 * After filling guests structures, nr_vnodes, nr_vmemranges and nr_vcpus
 714 * copied back to guest. Domain returns expected values of nr_vnodes,
 715 * nr_vmemranges and nr_vcpus to guest if the values where incorrect.
 716 */
 717struct xen_vnuma_topology_info {
 718    /* IN */
 719    domid_t domid;
 720    uint16_t pad;
 721    /* IN/OUT */
 722    unsigned int nr_vnodes;
 723    unsigned int nr_vcpus;
 724    unsigned int nr_vmemranges;
 725    /* OUT */
 726    union {
 727        XEN_GUEST_HANDLE(uint) h;
 728        uint64_t pad;
 729    } vdistance;
 730    union {
 731        XEN_GUEST_HANDLE(uint) h;
 732        uint64_t pad;
 733    } vcpu_to_vnode;
 734    union {
 735        XEN_GUEST_HANDLE(xen_vmemrange_t) h;
 736        uint64_t pad;
 737    } vmemrange;
 738};
 739typedef struct xen_vnuma_topology_info xen_vnuma_topology_info_t;
 740DEFINE_XEN_GUEST_HANDLE(xen_vnuma_topology_info_t);
 741
 742/* Next available subop number is 29 */
 743
 744#endif /* __XEN_PUBLIC_MEMORY_H__ */
 745
 746/*
 747 * Local variables:
 748 * mode: C
 749 * c-file-style: "BSD"
 750 * c-basic-offset: 4
 751 * tab-width: 4
 752 * indent-tabs-mode: nil
 753 * End:
 754 */
 755