linux/arch/powerpc/include/asm/kvm_host.h
<<
>>
Prefs
   1/*
   2 * This program is free software; you can redistribute it and/or modify
   3 * it under the terms of the GNU General Public License, version 2, as
   4 * published by the Free Software Foundation.
   5 *
   6 * This program is distributed in the hope that it will be useful,
   7 * but WITHOUT ANY WARRANTY; without even the implied warranty of
   8 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
   9 * GNU General Public License for more details.
  10 *
  11 * You should have received a copy of the GNU General Public License
  12 * along with this program; if not, write to the Free Software
  13 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301, USA.
  14 *
  15 * Copyright IBM Corp. 2007
  16 *
  17 * Authors: Hollis Blanchard <hollisb@us.ibm.com>
  18 */
  19
  20#ifndef __POWERPC_KVM_HOST_H__
  21#define __POWERPC_KVM_HOST_H__
  22
  23#include <linux/mutex.h>
  24#include <linux/hrtimer.h>
  25#include <linux/interrupt.h>
  26#include <linux/types.h>
  27#include <linux/kvm_types.h>
  28#include <linux/threads.h>
  29#include <linux/spinlock.h>
  30#include <linux/kvm_para.h>
  31#include <linux/list.h>
  32#include <linux/atomic.h>
  33#include <asm/kvm_asm.h>
  34#include <asm/processor.h>
  35#include <asm/page.h>
  36
  37#define KVM_MAX_VCPUS           NR_CPUS
  38#define KVM_MAX_VCORES          NR_CPUS
  39#define KVM_MEMORY_SLOTS 32
  40/* memory slots that does not exposed to userspace */
  41#define KVM_PRIVATE_MEM_SLOTS 4
  42#define KVM_MEM_SLOTS_NUM (KVM_MEMORY_SLOTS + KVM_PRIVATE_MEM_SLOTS)
  43
  44#ifdef CONFIG_KVM_MMIO
  45#define KVM_COALESCED_MMIO_PAGE_OFFSET 1
  46#endif
  47
  48#ifdef CONFIG_KVM_BOOK3S_64_HV
  49#include <linux/mmu_notifier.h>
  50
  51#define KVM_ARCH_WANT_MMU_NOTIFIER
  52
  53struct kvm;
  54extern int kvm_unmap_hva(struct kvm *kvm, unsigned long hva);
  55extern int kvm_age_hva(struct kvm *kvm, unsigned long hva);
  56extern int kvm_test_age_hva(struct kvm *kvm, unsigned long hva);
  57extern void kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte);
  58
  59#endif
  60
  61/* We don't currently support large pages. */
  62#define KVM_HPAGE_GFN_SHIFT(x)  0
  63#define KVM_NR_PAGE_SIZES       1
  64#define KVM_PAGES_PER_HPAGE(x)  (1UL<<31)
  65
  66#define HPTEG_CACHE_NUM                 (1 << 15)
  67#define HPTEG_HASH_BITS_PTE             13
  68#define HPTEG_HASH_BITS_PTE_LONG        12
  69#define HPTEG_HASH_BITS_VPTE            13
  70#define HPTEG_HASH_BITS_VPTE_LONG       5
  71#define HPTEG_HASH_NUM_PTE              (1 << HPTEG_HASH_BITS_PTE)
  72#define HPTEG_HASH_NUM_PTE_LONG         (1 << HPTEG_HASH_BITS_PTE_LONG)
  73#define HPTEG_HASH_NUM_VPTE             (1 << HPTEG_HASH_BITS_VPTE)
  74#define HPTEG_HASH_NUM_VPTE_LONG        (1 << HPTEG_HASH_BITS_VPTE_LONG)
  75
  76/* Physical Address Mask - allowed range of real mode RAM access */
  77#define KVM_PAM                 0x0fffffffffffffffULL
  78
  79struct kvm;
  80struct kvm_run;
  81struct kvm_vcpu;
  82
  83struct lppaca;
  84struct slb_shadow;
  85struct dtl;
  86
  87struct kvm_vm_stat {
  88        u32 remote_tlb_flush;
  89};
  90
  91struct kvm_vcpu_stat {
  92        u32 sum_exits;
  93        u32 mmio_exits;
  94        u32 dcr_exits;
  95        u32 signal_exits;
  96        u32 light_exits;
  97        /* Account for special types of light exits: */
  98        u32 itlb_real_miss_exits;
  99        u32 itlb_virt_miss_exits;
 100        u32 dtlb_real_miss_exits;
 101        u32 dtlb_virt_miss_exits;
 102        u32 syscall_exits;
 103        u32 isi_exits;
 104        u32 dsi_exits;
 105        u32 emulated_inst_exits;
 106        u32 dec_exits;
 107        u32 ext_intr_exits;
 108        u32 halt_wakeup;
 109#ifdef CONFIG_PPC_BOOK3S
 110        u32 pf_storage;
 111        u32 pf_instruc;
 112        u32 sp_storage;
 113        u32 sp_instruc;
 114        u32 queue_intr;
 115        u32 ld;
 116        u32 ld_slow;
 117        u32 st;
 118        u32 st_slow;
 119#endif
 120};
 121
 122enum kvm_exit_types {
 123        MMIO_EXITS,
 124        DCR_EXITS,
 125        SIGNAL_EXITS,
 126        ITLB_REAL_MISS_EXITS,
 127        ITLB_VIRT_MISS_EXITS,
 128        DTLB_REAL_MISS_EXITS,
 129        DTLB_VIRT_MISS_EXITS,
 130        SYSCALL_EXITS,
 131        ISI_EXITS,
 132        DSI_EXITS,
 133        EMULATED_INST_EXITS,
 134        EMULATED_MTMSRWE_EXITS,
 135        EMULATED_WRTEE_EXITS,
 136        EMULATED_MTSPR_EXITS,
 137        EMULATED_MFSPR_EXITS,
 138        EMULATED_MTMSR_EXITS,
 139        EMULATED_MFMSR_EXITS,
 140        EMULATED_TLBSX_EXITS,
 141        EMULATED_TLBWE_EXITS,
 142        EMULATED_RFI_EXITS,
 143        DEC_EXITS,
 144        EXT_INTR_EXITS,
 145        HALT_WAKEUP,
 146        USR_PR_INST,
 147        FP_UNAVAIL,
 148        DEBUG_EXITS,
 149        TIMEINGUEST,
 150        __NUMBER_OF_KVM_EXIT_TYPES
 151};
 152
 153/* allow access to big endian 32bit upper/lower parts and 64bit var */
 154struct kvmppc_exit_timing {
 155        union {
 156                u64 tv64;
 157                struct {
 158                        u32 tbu, tbl;
 159                } tv32;
 160        };
 161};
 162
 163struct kvmppc_pginfo {
 164        unsigned long pfn;
 165        atomic_t refcnt;
 166};
 167
 168struct kvmppc_spapr_tce_table {
 169        struct list_head list;
 170        struct kvm *kvm;
 171        u64 liobn;
 172        u32 window_size;
 173        struct page *pages[0];
 174};
 175
 176struct kvmppc_linear_info {
 177        void            *base_virt;
 178        unsigned long    base_pfn;
 179        unsigned long    npages;
 180        struct list_head list;
 181        atomic_t         use_count;
 182        int              type;
 183};
 184
 185/*
 186 * The reverse mapping array has one entry for each HPTE,
 187 * which stores the guest's view of the second word of the HPTE
 188 * (including the guest physical address of the mapping),
 189 * plus forward and backward pointers in a doubly-linked ring
 190 * of HPTEs that map the same host page.  The pointers in this
 191 * ring are 32-bit HPTE indexes, to save space.
 192 */
 193struct revmap_entry {
 194        unsigned long guest_rpte;
 195        unsigned int forw, back;
 196};
 197
 198/*
 199 * We use the top bit of each memslot->rmap entry as a lock bit,
 200 * and bit 32 as a present flag.  The bottom 32 bits are the
 201 * index in the guest HPT of a HPTE that points to the page.
 202 */
 203#define KVMPPC_RMAP_LOCK_BIT    63
 204#define KVMPPC_RMAP_RC_SHIFT    32
 205#define KVMPPC_RMAP_REFERENCED  (HPTE_R_R << KVMPPC_RMAP_RC_SHIFT)
 206#define KVMPPC_RMAP_CHANGED     (HPTE_R_C << KVMPPC_RMAP_RC_SHIFT)
 207#define KVMPPC_RMAP_PRESENT     0x100000000ul
 208#define KVMPPC_RMAP_INDEX       0xfffffffful
 209
 210/* Low-order bits in kvm->arch.slot_phys[][] */
 211#define KVMPPC_PAGE_ORDER_MASK  0x1f
 212#define KVMPPC_PAGE_NO_CACHE    HPTE_R_I        /* 0x20 */
 213#define KVMPPC_PAGE_WRITETHRU   HPTE_R_W        /* 0x40 */
 214#define KVMPPC_GOT_PAGE         0x80
 215
 216struct kvm_arch_memory_slot {
 217};
 218
 219struct kvm_arch {
 220#ifdef CONFIG_KVM_BOOK3S_64_HV
 221        unsigned long hpt_virt;
 222        struct revmap_entry *revmap;
 223        unsigned int lpid;
 224        unsigned int host_lpid;
 225        unsigned long host_lpcr;
 226        unsigned long sdr1;
 227        unsigned long host_sdr1;
 228        int tlbie_lock;
 229        unsigned long lpcr;
 230        unsigned long rmor;
 231        struct kvmppc_linear_info *rma;
 232        unsigned long vrma_slb_v;
 233        int rma_setup_done;
 234        int using_mmu_notifiers;
 235        struct list_head spapr_tce_tables;
 236        spinlock_t slot_phys_lock;
 237        unsigned long *slot_phys[KVM_MEM_SLOTS_NUM];
 238        int slot_npages[KVM_MEM_SLOTS_NUM];
 239        unsigned short last_vcpu[NR_CPUS];
 240        struct kvmppc_vcore *vcores[KVM_MAX_VCORES];
 241        struct kvmppc_linear_info *hpt_li;
 242#endif /* CONFIG_KVM_BOOK3S_64_HV */
 243};
 244
 245/*
 246 * Struct for a virtual core.
 247 * Note: entry_exit_count combines an entry count in the bottom 8 bits
 248 * and an exit count in the next 8 bits.  This is so that we can
 249 * atomically increment the entry count iff the exit count is 0
 250 * without taking the lock.
 251 */
 252struct kvmppc_vcore {
 253        int n_runnable;
 254        int n_busy;
 255        int num_threads;
 256        int entry_exit_count;
 257        int n_woken;
 258        int nap_count;
 259        int napping_threads;
 260        u16 pcpu;
 261        u8 vcore_state;
 262        u8 in_guest;
 263        struct list_head runnable_threads;
 264        spinlock_t lock;
 265        wait_queue_head_t wq;
 266};
 267
 268#define VCORE_ENTRY_COUNT(vc)   ((vc)->entry_exit_count & 0xff)
 269#define VCORE_EXIT_COUNT(vc)    ((vc)->entry_exit_count >> 8)
 270
 271/* Values for vcore_state */
 272#define VCORE_INACTIVE  0
 273#define VCORE_RUNNING   1
 274#define VCORE_EXITING   2
 275#define VCORE_SLEEPING  3
 276
 277struct kvmppc_pte {
 278        ulong eaddr;
 279        u64 vpage;
 280        ulong raddr;
 281        bool may_read           : 1;
 282        bool may_write          : 1;
 283        bool may_execute        : 1;
 284};
 285
 286struct kvmppc_mmu {
 287        /* book3s_64 only */
 288        void (*slbmte)(struct kvm_vcpu *vcpu, u64 rb, u64 rs);
 289        u64  (*slbmfee)(struct kvm_vcpu *vcpu, u64 slb_nr);
 290        u64  (*slbmfev)(struct kvm_vcpu *vcpu, u64 slb_nr);
 291        void (*slbie)(struct kvm_vcpu *vcpu, u64 slb_nr);
 292        void (*slbia)(struct kvm_vcpu *vcpu);
 293        /* book3s */
 294        void (*mtsrin)(struct kvm_vcpu *vcpu, u32 srnum, ulong value);
 295        u32  (*mfsrin)(struct kvm_vcpu *vcpu, u32 srnum);
 296        int  (*xlate)(struct kvm_vcpu *vcpu, gva_t eaddr, struct kvmppc_pte *pte, bool data);
 297        void (*reset_msr)(struct kvm_vcpu *vcpu);
 298        void (*tlbie)(struct kvm_vcpu *vcpu, ulong addr, bool large);
 299        int  (*esid_to_vsid)(struct kvm_vcpu *vcpu, ulong esid, u64 *vsid);
 300        u64  (*ea_to_vp)(struct kvm_vcpu *vcpu, gva_t eaddr, bool data);
 301        bool (*is_dcbz32)(struct kvm_vcpu *vcpu);
 302};
 303
 304struct kvmppc_slb {
 305        u64 esid;
 306        u64 vsid;
 307        u64 orige;
 308        u64 origv;
 309        bool valid      : 1;
 310        bool Ks         : 1;
 311        bool Kp         : 1;
 312        bool nx         : 1;
 313        bool large      : 1;    /* PTEs are 16MB */
 314        bool tb         : 1;    /* 1TB segment */
 315        bool class      : 1;
 316};
 317
 318struct kvm_vcpu_arch {
 319        ulong host_stack;
 320        u32 host_pid;
 321#ifdef CONFIG_PPC_BOOK3S
 322        struct kvmppc_slb slb[64];
 323        int slb_max;            /* 1 + index of last valid entry in slb[] */
 324        int slb_nr;             /* total number of entries in SLB */
 325        struct kvmppc_mmu mmu;
 326#endif
 327
 328        ulong gpr[32];
 329
 330        u64 fpr[32];
 331        u64 fpscr;
 332
 333#ifdef CONFIG_SPE
 334        ulong evr[32];
 335        ulong spefscr;
 336        ulong host_spefscr;
 337        u64 acc;
 338#endif
 339#ifdef CONFIG_ALTIVEC
 340        vector128 vr[32];
 341        vector128 vscr;
 342#endif
 343
 344#ifdef CONFIG_VSX
 345        u64 vsr[64];
 346#endif
 347
 348#ifdef CONFIG_PPC_BOOK3S
 349        /* For Gekko paired singles */
 350        u32 qpr[32];
 351#endif
 352
 353        ulong pc;
 354        ulong ctr;
 355        ulong lr;
 356
 357        ulong xer;
 358        u32 cr;
 359
 360#ifdef CONFIG_PPC_BOOK3S
 361        ulong hflags;
 362        ulong guest_owned_ext;
 363        ulong purr;
 364        ulong spurr;
 365        ulong dscr;
 366        ulong amr;
 367        ulong uamor;
 368        u32 ctrl;
 369        ulong dabr;
 370#endif
 371        u32 vrsave; /* also USPRG0 */
 372        u32 mmucr;
 373        ulong shadow_msr;
 374        ulong csrr0;
 375        ulong csrr1;
 376        ulong dsrr0;
 377        ulong dsrr1;
 378        ulong mcsrr0;
 379        ulong mcsrr1;
 380        ulong mcsr;
 381        u32 dec;
 382        u32 decar;
 383        u32 tbl;
 384        u32 tbu;
 385        u32 tcr;
 386        ulong tsr; /* we need to perform set/clr_bits() which requires ulong */
 387        u32 ivor[64];
 388        ulong ivpr;
 389        u32 pvr;
 390
 391        u32 shadow_pid;
 392        u32 shadow_pid1;
 393        u32 pid;
 394        u32 swap_pid;
 395
 396        u32 ccr0;
 397        u32 ccr1;
 398        u32 dbcr0;
 399        u32 dbcr1;
 400        u32 dbsr;
 401
 402        u64 mmcr[3];
 403        u32 pmc[8];
 404
 405#ifdef CONFIG_KVM_EXIT_TIMING
 406        struct mutex exit_timing_lock;
 407        struct kvmppc_exit_timing timing_exit;
 408        struct kvmppc_exit_timing timing_last_enter;
 409        u32 last_exit_type;
 410        u32 timing_count_type[__NUMBER_OF_KVM_EXIT_TYPES];
 411        u64 timing_sum_duration[__NUMBER_OF_KVM_EXIT_TYPES];
 412        u64 timing_sum_quad_duration[__NUMBER_OF_KVM_EXIT_TYPES];
 413        u64 timing_min_duration[__NUMBER_OF_KVM_EXIT_TYPES];
 414        u64 timing_max_duration[__NUMBER_OF_KVM_EXIT_TYPES];
 415        u64 timing_last_exit;
 416        struct dentry *debugfs_exit_timing;
 417#endif
 418
 419#ifdef CONFIG_PPC_BOOK3S
 420        ulong fault_dar;
 421        u32 fault_dsisr;
 422#endif
 423
 424#ifdef CONFIG_BOOKE
 425        ulong fault_dear;
 426        ulong fault_esr;
 427        ulong queued_dear;
 428        ulong queued_esr;
 429#endif
 430        gpa_t paddr_accessed;
 431
 432        u8 io_gpr; /* GPR used as IO source/target */
 433        u8 mmio_is_bigendian;
 434        u8 mmio_sign_extend;
 435        u8 dcr_needed;
 436        u8 dcr_is_write;
 437        u8 osi_needed;
 438        u8 osi_enabled;
 439        u8 papr_enabled;
 440        u8 sane;
 441        u8 cpu_type;
 442        u8 hcall_needed;
 443
 444        u32 cpr0_cfgaddr; /* holds the last set cpr0_cfgaddr */
 445
 446        struct hrtimer dec_timer;
 447        struct tasklet_struct tasklet;
 448        u64 dec_jiffies;
 449        u64 dec_expires;
 450        unsigned long pending_exceptions;
 451        u16 last_cpu;
 452        u8 ceded;
 453        u8 prodded;
 454        u32 last_inst;
 455
 456        struct lppaca *vpa;
 457        struct slb_shadow *slb_shadow;
 458        struct dtl *dtl;
 459        struct dtl *dtl_end;
 460
 461        wait_queue_head_t *wqp;
 462        struct kvmppc_vcore *vcore;
 463        int ret;
 464        int trap;
 465        int state;
 466        int ptid;
 467        bool timer_running;
 468        wait_queue_head_t cpu_run;
 469
 470        struct kvm_vcpu_arch_shared *shared;
 471        unsigned long magic_page_pa; /* phys addr to map the magic page to */
 472        unsigned long magic_page_ea; /* effect. addr to map the magic page to */
 473
 474#ifdef CONFIG_KVM_BOOK3S_64_HV
 475        struct kvm_vcpu_arch_shared shregs;
 476
 477        unsigned long pgfault_addr;
 478        long pgfault_index;
 479        unsigned long pgfault_hpte[2];
 480
 481        struct list_head run_list;
 482        struct task_struct *run_task;
 483        struct kvm_run *kvm_run;
 484        pgd_t *pgdir;
 485#endif
 486};
 487
 488/* Values for vcpu->arch.state */
 489#define KVMPPC_VCPU_STOPPED             0
 490#define KVMPPC_VCPU_BUSY_IN_HOST        1
 491#define KVMPPC_VCPU_RUNNABLE            2
 492
 493/* Values for vcpu->arch.io_gpr */
 494#define KVM_MMIO_REG_MASK       0x001f
 495#define KVM_MMIO_REG_EXT_MASK   0xffe0
 496#define KVM_MMIO_REG_GPR        0x0000
 497#define KVM_MMIO_REG_FPR        0x0020
 498#define KVM_MMIO_REG_QPR        0x0040
 499#define KVM_MMIO_REG_FQPR       0x0060
 500
 501#endif /* __POWERPC_KVM_HOST_H__ */
 502