linux/arch/mips/kvm/dyntrans.c
<<
>>
Prefs
   1/*
   2 * This file is subject to the terms and conditions of the GNU General Public
   3 * License.  See the file "COPYING" in the main directory of this archive
   4 * for more details.
   5 *
   6 * KVM/MIPS: Binary Patching for privileged instructions, reduces traps.
   7 *
   8 * Copyright (C) 2012  MIPS Technologies, Inc.  All rights reserved.
   9 * Authors: Sanjay Lal <sanjayl@kymasys.com>
  10 */
  11
  12#include <linux/errno.h>
  13#include <linux/err.h>
  14#include <linux/highmem.h>
  15#include <linux/kvm_host.h>
  16#include <linux/uaccess.h>
  17#include <linux/vmalloc.h>
  18#include <linux/fs.h>
  19#include <linux/memblock.h>
  20#include <asm/cacheflush.h>
  21
  22#include "commpage.h"
  23
  24/**
  25 * kvm_mips_trans_replace() - Replace trapping instruction in guest memory.
  26 * @vcpu:       Virtual CPU.
  27 * @opc:        PC of instruction to replace.
  28 * @replace:    Instruction to write
  29 */
  30static int kvm_mips_trans_replace(struct kvm_vcpu *vcpu, u32 *opc,
  31                                  union mips_instruction replace)
  32{
  33        unsigned long vaddr = (unsigned long)opc;
  34        int err;
  35
  36retry:
  37        /* The GVA page table is still active so use the Linux TLB handlers */
  38        kvm_trap_emul_gva_lockless_begin(vcpu);
  39        err = put_user(replace.word, opc);
  40        kvm_trap_emul_gva_lockless_end(vcpu);
  41
  42        if (unlikely(err)) {
  43                /*
  44                 * We write protect clean pages in GVA page table so normal
  45                 * Linux TLB mod handler doesn't silently dirty the page.
  46                 * Its also possible we raced with a GVA invalidation.
  47                 * Try to force the page to become dirty.
  48                 */
  49                err = kvm_trap_emul_gva_fault(vcpu, vaddr, true);
  50                if (unlikely(err)) {
  51                        kvm_info("%s: Address unwriteable: %p\n",
  52                                 __func__, opc);
  53                        return -EFAULT;
  54                }
  55
  56                /*
  57                 * Try again. This will likely trigger a TLB refill, which will
  58                 * fetch the new dirty entry from the GVA page table, which
  59                 * should then succeed.
  60                 */
  61                goto retry;
  62        }
  63        __local_flush_icache_user_range(vaddr, vaddr + 4);
  64
  65        return 0;
  66}
  67
  68int kvm_mips_trans_cache_index(union mips_instruction inst, u32 *opc,
  69                               struct kvm_vcpu *vcpu)
  70{
  71        union mips_instruction nop_inst = { 0 };
  72
  73        /* Replace the CACHE instruction, with a NOP */
  74        return kvm_mips_trans_replace(vcpu, opc, nop_inst);
  75}
  76
  77/*
  78 * Address based CACHE instructions are transformed into synci(s). A little
  79 * heavy for just D-cache invalidates, but avoids an expensive trap
  80 */
  81int kvm_mips_trans_cache_va(union mips_instruction inst, u32 *opc,
  82                            struct kvm_vcpu *vcpu)
  83{
  84        union mips_instruction synci_inst = { 0 };
  85
  86        synci_inst.i_format.opcode = bcond_op;
  87        synci_inst.i_format.rs = inst.i_format.rs;
  88        synci_inst.i_format.rt = synci_op;
  89        if (cpu_has_mips_r6)
  90                synci_inst.i_format.simmediate = inst.spec3_format.simmediate;
  91        else
  92                synci_inst.i_format.simmediate = inst.i_format.simmediate;
  93
  94        return kvm_mips_trans_replace(vcpu, opc, synci_inst);
  95}
  96
  97int kvm_mips_trans_mfc0(union mips_instruction inst, u32 *opc,
  98                        struct kvm_vcpu *vcpu)
  99{
 100        union mips_instruction mfc0_inst = { 0 };
 101        u32 rd, sel;
 102
 103        rd = inst.c0r_format.rd;
 104        sel = inst.c0r_format.sel;
 105
 106        if (rd == MIPS_CP0_ERRCTL && sel == 0) {
 107                mfc0_inst.r_format.opcode = spec_op;
 108                mfc0_inst.r_format.rd = inst.c0r_format.rt;
 109                mfc0_inst.r_format.func = add_op;
 110        } else {
 111                mfc0_inst.i_format.opcode = lw_op;
 112                mfc0_inst.i_format.rt = inst.c0r_format.rt;
 113                mfc0_inst.i_format.simmediate = KVM_GUEST_COMMPAGE_ADDR |
 114                        offsetof(struct kvm_mips_commpage, cop0.reg[rd][sel]);
 115#ifdef CONFIG_CPU_BIG_ENDIAN
 116                if (sizeof(vcpu->arch.cop0->reg[0][0]) == 8)
 117                        mfc0_inst.i_format.simmediate |= 4;
 118#endif
 119        }
 120
 121        return kvm_mips_trans_replace(vcpu, opc, mfc0_inst);
 122}
 123
 124int kvm_mips_trans_mtc0(union mips_instruction inst, u32 *opc,
 125                        struct kvm_vcpu *vcpu)
 126{
 127        union mips_instruction mtc0_inst = { 0 };
 128        u32 rd, sel;
 129
 130        rd = inst.c0r_format.rd;
 131        sel = inst.c0r_format.sel;
 132
 133        mtc0_inst.i_format.opcode = sw_op;
 134        mtc0_inst.i_format.rt = inst.c0r_format.rt;
 135        mtc0_inst.i_format.simmediate = KVM_GUEST_COMMPAGE_ADDR |
 136                offsetof(struct kvm_mips_commpage, cop0.reg[rd][sel]);
 137#ifdef CONFIG_CPU_BIG_ENDIAN
 138        if (sizeof(vcpu->arch.cop0->reg[0][0]) == 8)
 139                mtc0_inst.i_format.simmediate |= 4;
 140#endif
 141
 142        return kvm_mips_trans_replace(vcpu, opc, mtc0_inst);
 143}
 144