/* * This file is subject to the terms and conditions of the GNU General Public * License. See the file "COPYING" in the main directory of this archive * for more details. * * KVM/MIPS: Binary Patching for privileged instructions, reduces traps. * * Copyright (C) 2012 MIPS Technologies, Inc. All rights reserved. * Authors: Sanjay Lal */ #include #include #include #include #include #include #include #include #include #include "commpage.h" /** * kvm_mips_trans_replace() - Replace trapping instruction in guest memory. * @vcpu: Virtual CPU. * @opc: PC of instruction to replace. * @replace: Instruction to write */ static int kvm_mips_trans_replace(struct kvm_vcpu *vcpu, u32 *opc, union mips_instruction replace) { unsigned long vaddr = (unsigned long)opc; int err; retry: /* The GVA page table is still active so use the Linux TLB handlers */ kvm_trap_emul_gva_lockless_begin(vcpu); err = put_user(replace.word, opc); kvm_trap_emul_gva_lockless_end(vcpu); if (unlikely(err)) { /* * We write protect clean pages in GVA page table so normal * Linux TLB mod handler doesn't silently dirty the page. * Its also possible we raced with a GVA invalidation. * Try to force the page to become dirty. */ err = kvm_trap_emul_gva_fault(vcpu, vaddr, true); if (unlikely(err)) { kvm_info("%s: Address unwriteable: %p\n", __func__, opc); return -EFAULT; } /* * Try again. This will likely trigger a TLB refill, which will * fetch the new dirty entry from the GVA page table, which * should then succeed. */ goto retry; } __local_flush_icache_user_range(vaddr, vaddr + 4); return 0; } int kvm_mips_trans_cache_index(union mips_instruction inst, u32 *opc, struct kvm_vcpu *vcpu) { union mips_instruction nop_inst = { 0 }; /* Replace the CACHE instruction, with a NOP */ return kvm_mips_trans_replace(vcpu, opc, nop_inst); } /* * Address based CACHE instructions are transformed into synci(s). A little * heavy for just D-cache invalidates, but avoids an expensive trap */ int kvm_mips_trans_cache_va(union mips_instruction inst, u32 *opc, struct kvm_vcpu *vcpu) { union mips_instruction synci_inst = { 0 }; synci_inst.i_format.opcode = bcond_op; synci_inst.i_format.rs = inst.i_format.rs; synci_inst.i_format.rt = synci_op; if (cpu_has_mips_r6) synci_inst.i_format.simmediate = inst.spec3_format.simmediate; else synci_inst.i_format.simmediate = inst.i_format.simmediate; return kvm_mips_trans_replace(vcpu, opc, synci_inst); } int kvm_mips_trans_mfc0(union mips_instruction inst, u32 *opc, struct kvm_vcpu *vcpu) { union mips_instruction mfc0_inst = { 0 }; u32 rd, sel; rd = inst.c0r_format.rd; sel = inst.c0r_format.sel; if (rd == MIPS_CP0_ERRCTL && sel == 0) { mfc0_inst.r_format.opcode = spec_op; mfc0_inst.r_format.rd = inst.c0r_format.rt; mfc0_inst.r_format.func = add_op; } else { mfc0_inst.i_format.opcode = lw_op; mfc0_inst.i_format.rt = inst.c0r_format.rt; mfc0_inst.i_format.simmediate = KVM_GUEST_COMMPAGE_ADDR | offsetof(struct kvm_mips_commpage, cop0.reg[rd][sel]); #ifdef CONFIG_CPU_BIG_ENDIAN if (sizeof(vcpu->arch.cop0->reg[0][0]) == 8) mfc0_inst.i_format.simmediate |= 4; #endif } return kvm_mips_trans_replace(vcpu, opc, mfc0_inst); } int kvm_mips_trans_mtc0(union mips_instruction inst, u32 *opc, struct kvm_vcpu *vcpu) { union mips_instruction mtc0_inst = { 0 }; u32 rd, sel; rd = inst.c0r_format.rd; sel = inst.c0r_format.sel; mtc0_inst.i_format.opcode = sw_op; mtc0_inst.i_format.rt = inst.c0r_format.rt; mtc0_inst.i_format.simmediate = KVM_GUEST_COMMPAGE_ADDR | offsetof(struct kvm_mips_commpage, cop0.reg[rd][sel]); #ifdef CONFIG_CPU_BIG_ENDIAN if (sizeof(vcpu->arch.cop0->reg[0][0]) == 8) mtc0_inst.i_format.simmediate |= 4; #endif return kvm_mips_trans_replace(vcpu, opc, mtc0_inst); }