linux/arch/ia64/kvm/mmio.c
<<
>>
Prefs
   1/*
   2 * mmio.c: MMIO emulation components.
   3 * Copyright (c) 2004, Intel Corporation.
   4 *  Yaozu Dong (Eddie Dong) (Eddie.dong@intel.com)
   5 *  Kun Tian (Kevin Tian) (Kevin.tian@intel.com)
   6 *
   7 * Copyright (c) 2007 Intel Corporation  KVM support.
   8 * Xuefei Xu (Anthony Xu) (anthony.xu@intel.com)
   9 * Xiantao Zhang  (xiantao.zhang@intel.com)
  10 *
  11 * This program is free software; you can redistribute it and/or modify it
  12 * under the terms and conditions of the GNU General Public License,
  13 * version 2, as published by the Free Software Foundation.
  14 *
  15 * This program is distributed in the hope it will be useful, but WITHOUT
  16 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  17 * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
  18 * more details.
  19 *
  20 * You should have received a copy of the GNU General Public License along with
  21 * this program; if not, write to the Free Software Foundation, Inc., 59 Temple
  22 * Place - Suite 330, Boston, MA 02111-1307 USA.
  23 *
  24 */
  25
  26#include <linux/kvm_host.h>
  27
  28#include "vcpu.h"
  29
  30static void vlsapic_write_xtp(struct kvm_vcpu *v, uint8_t val)
  31{
  32        VLSAPIC_XTP(v) = val;
  33}
  34
  35/*
  36 * LSAPIC OFFSET
  37 */
  38#define PIB_LOW_HALF(ofst)     !(ofst & (1 << 20))
  39#define PIB_OFST_INTA          0x1E0000
  40#define PIB_OFST_XTP           0x1E0008
  41
  42/*
  43 * execute write IPI op.
  44 */
  45static void vlsapic_write_ipi(struct kvm_vcpu *vcpu,
  46                                        uint64_t addr, uint64_t data)
  47{
  48        struct exit_ctl_data *p = &current_vcpu->arch.exit_data;
  49        unsigned long psr;
  50
  51        local_irq_save(psr);
  52
  53        p->exit_reason = EXIT_REASON_IPI;
  54        p->u.ipi_data.addr.val = addr;
  55        p->u.ipi_data.data.val = data;
  56        vmm_transition(current_vcpu);
  57
  58        local_irq_restore(psr);
  59
  60}
  61
  62void lsapic_write(struct kvm_vcpu *v, unsigned long addr,
  63                        unsigned long length, unsigned long val)
  64{
  65        addr &= (PIB_SIZE - 1);
  66
  67        switch (addr) {
  68        case PIB_OFST_INTA:
  69                panic_vm(v, "Undefined write on PIB INTA\n");
  70                break;
  71        case PIB_OFST_XTP:
  72                if (length == 1) {
  73                        vlsapic_write_xtp(v, val);
  74                } else {
  75                        panic_vm(v, "Undefined write on PIB XTP\n");
  76                }
  77                break;
  78        default:
  79                if (PIB_LOW_HALF(addr)) {
  80                        /*Lower half */
  81                        if (length != 8)
  82                                panic_vm(v, "Can't LHF write with size %ld!\n",
  83                                                length);
  84                        else
  85                                vlsapic_write_ipi(v, addr, val);
  86                } else {   /*Upper half */
  87                        panic_vm(v, "IPI-UHF write %lx\n", addr);
  88                }
  89                break;
  90        }
  91}
  92
  93unsigned long lsapic_read(struct kvm_vcpu *v, unsigned long addr,
  94                unsigned long length)
  95{
  96        uint64_t result = 0;
  97
  98        addr &= (PIB_SIZE - 1);
  99
 100        switch (addr) {
 101        case PIB_OFST_INTA:
 102                if (length == 1) /* 1 byte load */
 103                        ; /* There is no i8259, there is no INTA access*/
 104                else
 105                        panic_vm(v, "Undefined read on PIB INTA\n");
 106
 107                break;
 108        case PIB_OFST_XTP:
 109                if (length == 1) {
 110                        result = VLSAPIC_XTP(v);
 111                } else {
 112                        panic_vm(v, "Undefined read on PIB XTP\n");
 113                }
 114                break;
 115        default:
 116                panic_vm(v, "Undefined addr access for lsapic!\n");
 117                break;
 118        }
 119        return result;
 120}
 121
 122static void mmio_access(struct kvm_vcpu *vcpu, u64 src_pa, u64 *dest,
 123                                        u16 s, int ma, int dir)
 124{
 125        unsigned long iot;
 126        struct exit_ctl_data *p = &vcpu->arch.exit_data;
 127        unsigned long psr;
 128
 129        iot = __gpfn_is_io(src_pa >> PAGE_SHIFT);
 130
 131        local_irq_save(psr);
 132
 133        /*Intercept the access for PIB range*/
 134        if (iot == GPFN_PIB) {
 135                if (!dir)
 136                        lsapic_write(vcpu, src_pa, s, *dest);
 137                else
 138                        *dest = lsapic_read(vcpu, src_pa, s);
 139                goto out;
 140        }
 141        p->exit_reason = EXIT_REASON_MMIO_INSTRUCTION;
 142        p->u.ioreq.addr = src_pa;
 143        p->u.ioreq.size = s;
 144        p->u.ioreq.dir = dir;
 145        if (dir == IOREQ_WRITE)
 146                p->u.ioreq.data = *dest;
 147        p->u.ioreq.state = STATE_IOREQ_READY;
 148        vmm_transition(vcpu);
 149
 150        if (p->u.ioreq.state == STATE_IORESP_READY) {
 151                if (dir == IOREQ_READ)
 152                        /* it's necessary to ensure zero extending */
 153                        *dest = p->u.ioreq.data & (~0UL >> (64-(s*8)));
 154        } else
 155                panic_vm(vcpu, "Unhandled mmio access returned!\n");
 156out:
 157        local_irq_restore(psr);
 158        return ;
 159}
 160
 161/*
 162   dir 1: read 0:write
 163   inst_type 0:integer 1:floating point
 164 */
 165#define SL_INTEGER      0       /* store/load interger*/
 166#define SL_FLOATING     1       /* store/load floating*/
 167
 168void emulate_io_inst(struct kvm_vcpu *vcpu, u64 padr, u64 ma)
 169{
 170        struct kvm_pt_regs *regs;
 171        IA64_BUNDLE bundle;
 172        int slot, dir = 0;
 173        int inst_type = -1;
 174        u16 size = 0;
 175        u64 data, slot1a, slot1b, temp, update_reg;
 176        s32 imm;
 177        INST64 inst;
 178
 179        regs = vcpu_regs(vcpu);
 180
 181        if (fetch_code(vcpu, regs->cr_iip, &bundle)) {
 182                /* if fetch code fail, return and try again */
 183                return;
 184        }
 185        slot = ((struct ia64_psr *)&(regs->cr_ipsr))->ri;
 186        if (!slot)
 187                inst.inst = bundle.slot0;
 188        else if (slot == 1) {
 189                slot1a = bundle.slot1a;
 190                slot1b = bundle.slot1b;
 191                inst.inst = slot1a + (slot1b << 18);
 192        } else if (slot == 2)
 193                inst.inst = bundle.slot2;
 194
 195        /* Integer Load/Store */
 196        if (inst.M1.major == 4 && inst.M1.m == 0 && inst.M1.x == 0) {
 197                inst_type = SL_INTEGER;
 198                size = (inst.M1.x6 & 0x3);
 199                if ((inst.M1.x6 >> 2) > 0xb) {
 200                        /*write*/
 201                        dir = IOREQ_WRITE;
 202                        data = vcpu_get_gr(vcpu, inst.M4.r2);
 203                } else if ((inst.M1.x6 >> 2) < 0xb) {
 204                        /*read*/
 205                        dir = IOREQ_READ;
 206                }
 207        } else if (inst.M2.major == 4 && inst.M2.m == 1 && inst.M2.x == 0) {
 208                /* Integer Load + Reg update */
 209                inst_type = SL_INTEGER;
 210                dir = IOREQ_READ;
 211                size = (inst.M2.x6 & 0x3);
 212                temp = vcpu_get_gr(vcpu, inst.M2.r3);
 213                update_reg = vcpu_get_gr(vcpu, inst.M2.r2);
 214                temp += update_reg;
 215                vcpu_set_gr(vcpu, inst.M2.r3, temp, 0);
 216        } else if (inst.M3.major == 5) {
 217                /*Integer Load/Store + Imm update*/
 218                inst_type = SL_INTEGER;
 219                size = (inst.M3.x6&0x3);
 220                if ((inst.M5.x6 >> 2) > 0xb) {
 221                        /*write*/
 222                        dir = IOREQ_WRITE;
 223                        data = vcpu_get_gr(vcpu, inst.M5.r2);
 224                        temp = vcpu_get_gr(vcpu, inst.M5.r3);
 225                        imm = (inst.M5.s << 31) | (inst.M5.i << 30) |
 226                                (inst.M5.imm7 << 23);
 227                        temp += imm >> 23;
 228                        vcpu_set_gr(vcpu, inst.M5.r3, temp, 0);
 229
 230                } else if ((inst.M3.x6 >> 2) < 0xb) {
 231                        /*read*/
 232                        dir = IOREQ_READ;
 233                        temp = vcpu_get_gr(vcpu, inst.M3.r3);
 234                        imm = (inst.M3.s << 31) | (inst.M3.i << 30) |
 235                                (inst.M3.imm7 << 23);
 236                        temp += imm >> 23;
 237                        vcpu_set_gr(vcpu, inst.M3.r3, temp, 0);
 238
 239                }
 240        } else if (inst.M9.major == 6 && inst.M9.x6 == 0x3B
 241                                && inst.M9.m == 0 && inst.M9.x == 0) {
 242                /* Floating-point spill*/
 243                struct ia64_fpreg v;
 244
 245                inst_type = SL_FLOATING;
 246                dir = IOREQ_WRITE;
 247                vcpu_get_fpreg(vcpu, inst.M9.f2, &v);
 248                /* Write high word. FIXME: this is a kludge!  */
 249                v.u.bits[1] &= 0x3ffff;
 250                mmio_access(vcpu, padr + 8, (u64 *)&v.u.bits[1], 8,
 251                            ma, IOREQ_WRITE);
 252                data = v.u.bits[0];
 253                size = 3;
 254        } else if (inst.M10.major == 7 && inst.M10.x6 == 0x3B) {
 255                /* Floating-point spill + Imm update */
 256                struct ia64_fpreg v;
 257
 258                inst_type = SL_FLOATING;
 259                dir = IOREQ_WRITE;
 260                vcpu_get_fpreg(vcpu, inst.M10.f2, &v);
 261                temp = vcpu_get_gr(vcpu, inst.M10.r3);
 262                imm = (inst.M10.s << 31) | (inst.M10.i << 30) |
 263                        (inst.M10.imm7 << 23);
 264                temp += imm >> 23;
 265                vcpu_set_gr(vcpu, inst.M10.r3, temp, 0);
 266
 267                /* Write high word.FIXME: this is a kludge!  */
 268                v.u.bits[1] &= 0x3ffff;
 269                mmio_access(vcpu, padr + 8, (u64 *)&v.u.bits[1],
 270                            8, ma, IOREQ_WRITE);
 271                data = v.u.bits[0];
 272                size = 3;
 273        } else if (inst.M10.major == 7 && inst.M10.x6 == 0x31) {
 274                /* Floating-point stf8 + Imm update */
 275                struct ia64_fpreg v;
 276                inst_type = SL_FLOATING;
 277                dir = IOREQ_WRITE;
 278                size = 3;
 279                vcpu_get_fpreg(vcpu, inst.M10.f2, &v);
 280                data = v.u.bits[0]; /* Significand.  */
 281                temp = vcpu_get_gr(vcpu, inst.M10.r3);
 282                imm = (inst.M10.s << 31) | (inst.M10.i << 30) |
 283                        (inst.M10.imm7 << 23);
 284                temp += imm >> 23;
 285                vcpu_set_gr(vcpu, inst.M10.r3, temp, 0);
 286        } else if (inst.M15.major == 7 && inst.M15.x6 >= 0x2c
 287                        && inst.M15.x6 <= 0x2f) {
 288                temp = vcpu_get_gr(vcpu, inst.M15.r3);
 289                imm = (inst.M15.s << 31) | (inst.M15.i << 30) |
 290                        (inst.M15.imm7 << 23);
 291                temp += imm >> 23;
 292                vcpu_set_gr(vcpu, inst.M15.r3, temp, 0);
 293
 294                vcpu_increment_iip(vcpu);
 295                return;
 296        } else if (inst.M12.major == 6 && inst.M12.m == 1
 297                        && inst.M12.x == 1 && inst.M12.x6 == 1) {
 298                /* Floating-point Load Pair + Imm ldfp8 M12*/
 299                struct ia64_fpreg v;
 300
 301                inst_type = SL_FLOATING;
 302                dir = IOREQ_READ;
 303                size = 8;     /*ldfd*/
 304                mmio_access(vcpu, padr, &data, size, ma, dir);
 305                v.u.bits[0] = data;
 306                v.u.bits[1] = 0x1003E;
 307                vcpu_set_fpreg(vcpu, inst.M12.f1, &v);
 308                padr += 8;
 309                mmio_access(vcpu, padr, &data, size, ma, dir);
 310                v.u.bits[0] = data;
 311                v.u.bits[1] = 0x1003E;
 312                vcpu_set_fpreg(vcpu, inst.M12.f2, &v);
 313                padr += 8;
 314                vcpu_set_gr(vcpu, inst.M12.r3, padr, 0);
 315                vcpu_increment_iip(vcpu);
 316                return;
 317        } else {
 318                inst_type = -1;
 319                panic_vm(vcpu, "Unsupported MMIO access instruction! "
 320                                "Bunld[0]=0x%lx, Bundle[1]=0x%lx\n",
 321                                bundle.i64[0], bundle.i64[1]);
 322        }
 323
 324        size = 1 << size;
 325        if (dir == IOREQ_WRITE) {
 326                mmio_access(vcpu, padr, &data, size, ma, dir);
 327        } else {
 328                mmio_access(vcpu, padr, &data, size, ma, dir);
 329                if (inst_type == SL_INTEGER)
 330                        vcpu_set_gr(vcpu, inst.M1.r1, data, 0);
 331                else
 332                        panic_vm(vcpu, "Unsupported instruction type!\n");
 333
 334        }
 335        vcpu_increment_iip(vcpu);
 336}
 337