From patchwork Wed Mar 10 11:37:47 2010 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Andrea Gelmini X-Patchwork-Id: 84544 Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by demeter.kernel.org (8.14.3/8.14.3) with ESMTP id o2ABcE4O012989 for ; Wed, 10 Mar 2010 11:38:14 GMT Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756393Ab0CJLiK (ORCPT ); Wed, 10 Mar 2010 06:38:10 -0500 Received: from mail-fx0-f219.google.com ([209.85.220.219]:37368 "EHLO mail-fx0-f219.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1755842Ab0CJLiH (ORCPT ); Wed, 10 Mar 2010 06:38:07 -0500 Received: by fxm19 with SMTP id 19so8544448fxm.21 for ; Wed, 10 Mar 2010 03:38:04 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:received:received:sender:from:to:cc:subject :date:message-id:x-mailer:in-reply-to:references; bh=OiOrmWNzl6LW0wSLd7Ro1N+OKK7k9Q6mQxnjzDqEWOM=; b=bDFe5RI2kzSO2XQyE9WtskeKXwXFNVbzaw8Vr6xNNHVVIY2ElU5RcUOKdkjmHsshA7 C9d+8zvrMcBqSgnImqZ72eau4Ss8K58jpntWBjY0CHq14xyqbMzoy5zeSb2AxaUMhYdD 4wIBomVZU4+HhnbOi3DO9eRT0vjVwA6CdPkNM= DomainKey-Signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=sender:from:to:cc:subject:date:message-id:x-mailer:in-reply-to :references; b=w9DxldePO/d997ohuk2/6JL5WEFhhxd51LIAg1rLTLOTOO6SMkZ1FiGfLZ6MUD5tRu 1MRx9qTu9kNIz9i4PnAqJnOFRtLK7ULP5gbqmXTuy9Ny0hP9LwqB1OH9uDO1JfjqUBao nJROhETmjR4kzGPgQzxteJyhCS+PCepyovGzI= Received: by 10.223.73.82 with SMTP id p18mr1370863faj.34.1268221083597; Wed, 10 Mar 2010 03:38:03 -0800 (PST) Received: from localhost.localdomain (ip-125-235.sn2.eutelia.it [83.211.125.235]) by mx.google.com with ESMTPS id 13sm775283fxm.14.2010.03.10.03.38.02 (version=TLSv1/SSLv3 cipher=RC4-MD5); Wed, 10 Mar 2010 03:38:03 -0800 (PST) From: Andrea Gelmini To: mtosatti@redhat.com Cc: avi@redhat.com, kvm@vger.kernel.org, Andrea Gelmini Subject: [PATCH 2/2] arch/x86/kvm/* Checkpatch cleanup Date: Wed, 10 Mar 2010 12:37:47 +0100 Message-Id: <1268221067-23733-3-git-send-email-andrea.gelmini@gelma.net> X-Mailer: git-send-email 1.7.0.2.157.gb7e7f In-Reply-To: <1268221067-23733-1-git-send-email-andrea.gelmini@gelma.net> References: <1268221067-23733-1-git-send-email-andrea.gelmini@gelma.net> Sender: kvm-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org X-Greylist: IP, sender and recipient auto-whitelisted, not delayed by milter-greylist-4.2.3 (demeter.kernel.org [140.211.167.41]); Wed, 10 Mar 2010 11:38:14 +0000 (UTC) diff --git a/arch/x86/kvm/emulate.c b/arch/x86/kvm/emulate.c index 4dade6a..3ebec1e 100644 --- a/arch/x86/kvm/emulate.c +++ b/arch/x86/kvm/emulate.c @@ -449,35 +449,34 @@ static u32 group2_table[] = { /* Raw emulation: instruction has two explicit operands. */ -#define __emulate_2op_nobyte(_op,_src,_dst,_eflags,_wx,_wy,_lx,_ly,_qx,_qy) \ - do { \ - unsigned long _tmp; \ - \ - switch ((_dst).bytes) { \ - case 2: \ - ____emulate_2op(_op,_src,_dst,_eflags,_wx,_wy,"w"); \ - break; \ - case 4: \ - ____emulate_2op(_op,_src,_dst,_eflags,_lx,_ly,"l"); \ - break; \ - case 8: \ - ON64(____emulate_2op(_op,_src,_dst,_eflags,_qx,_qy,"q")); \ - break; \ - } \ +#define __emulate_2op_nobyte(_op, _src, _dst, _eflags, _wx, _wy, _lx, _ly, _qx, _qy) \ + do { \ + unsigned long _tmp; \ + switch ((_dst).bytes) { \ + case 2: \ + ____emulate_2op(_op, _src, _dst, _eflags, _wx, _wy, "w"); \ + break; \ + case 4: \ + ____emulate_2op(_op, _src, _dst, _eflags, _lx, _ly, "l"); \ + break; \ + case 8: \ + ON64(____emulate_2op(_op, _src, _dst, _eflags, _qx, _qy, "q")); \ + break; \ + } \ } while (0) -#define __emulate_2op(_op,_src,_dst,_eflags,_bx,_by,_wx,_wy,_lx,_ly,_qx,_qy) \ - do { \ - unsigned long _tmp; \ - switch ((_dst).bytes) { \ - case 1: \ - ____emulate_2op(_op,_src,_dst,_eflags,_bx,_by,"b"); \ - break; \ - default: \ - __emulate_2op_nobyte(_op, _src, _dst, _eflags, \ - _wx, _wy, _lx, _ly, _qx, _qy); \ - break; \ - } \ +#define __emulate_2op(_op, _src, _dst, _eflags, _bx, _by, _wx, _wy, _lx, _ly, _qx, _qy) \ + do { \ + unsigned long _tmp; \ + switch ((_dst).bytes) { \ + case 1: \ + ____emulate_2op(_op, _src, _dst, _eflags, _bx, _by, "b"); \ + break; \ + default: \ + __emulate_2op_nobyte(_op, _src, _dst, _eflags, \ + _wx, _wy, _lx, _ly, _qx, _qy); \ + break; \ + } \ } while (0) /* Source operand is byte-sized and may be restricted to just %cl. */ @@ -499,17 +498,17 @@ static u32 group2_table[] = { #define __emulate_2op_cl(_op, _cl, _src, _dst, _eflags, _suffix, _type) \ do { \ unsigned long _tmp; \ - _type _clv = (_cl).val; \ - _type _srcv = (_src).val; \ + _type _clv = (_cl).val; \ + _type _srcv = (_src).val; \ _type _dstv = (_dst).val; \ \ __asm__ __volatile__ ( \ _PRE_EFLAGS("0", "5", "2") \ - _op _suffix " %4,%1 \n" \ + _op _suffix " %4,%1\n" \ _POST_EFLAGS("0", "5", "2") \ : "=m" (_eflags), "+r" (_dstv), "=&r" (_tmp) \ : "c" (_clv) , "r" (_srcv), "i" (EFLAGS_MASK) \ - ); \ + ); \ \ (_cl).val = (unsigned long) _clv; \ (_src).val = (unsigned long) _srcv; \ @@ -520,16 +519,16 @@ static u32 group2_table[] = { do { \ switch ((_dst).bytes) { \ case 2: \ - __emulate_2op_cl(_op, _cl, _src, _dst, _eflags, \ - "w", unsigned short); \ + __emulate_2op_cl(_op, _cl, _src, _dst, _eflags, \ + "w", unsigned short); \ break; \ - case 4: \ - __emulate_2op_cl(_op, _cl, _src, _dst, _eflags, \ - "l", unsigned int); \ + case 4: \ + __emulate_2op_cl(_op, _cl, _src, _dst, _eflags, \ + "l", unsigned int); \ break; \ case 8: \ ON64(__emulate_2op_cl(_op, _cl, _src, _dst, _eflags, \ - "q", unsigned long)); \ + "q", unsigned long)); \ break; \ } \ } while (0) @@ -741,7 +740,7 @@ static int test_cc(unsigned int condition, unsigned int flags) } /* Odd condition identifiers (lsb == 1) have inverted sense. */ - return (!!rc ^ (condition & 1)); + return !!rc ^ (condition & 1); } static void decode_register_operand(struct operand *op, @@ -1194,15 +1193,15 @@ done_prefixes: c->dst.bytes = c->op_bytes; c->dst.ptr = &c->regs[VCPU_REGS_RAX]; switch (c->op_bytes) { - case 1: - c->dst.val = *(u8 *)c->dst.ptr; - break; - case 2: - c->dst.val = *(u16 *)c->dst.ptr; - break; - case 4: - c->dst.val = *(u32 *)c->dst.ptr; - break; + case 1: + c->dst.val = *(u8 *)c->dst.ptr; + break; + case 2: + c->dst.val = *(u16 *)c->dst.ptr; + break; + case 4: + c->dst.val = *(u32 *)c->dst.ptr; + break; } c->dst.orig_val = c->dst.val; break; @@ -1260,7 +1259,7 @@ static int emulate_popf(struct x86_emulate_ctxt *ctxt, change_mask = EFLG_CF | EFLG_PF | EFLG_AF | EFLG_ZF | EFLG_SF | EFLG_OF | EFLG_TF | EFLG_DF | EFLG_NT | EFLG_RF | EFLG_AC | EFLG_ID; - switch(ctxt->mode) { + switch (ctxt->mode) { case X86EMUL_MODE_PROT64: case X86EMUL_MODE_PROT32: case X86EMUL_MODE_PROT16: @@ -1853,8 +1852,8 @@ x86_emulate_insn(struct x86_emulate_ctxt *ctxt, struct x86_emulate_ops *ops) * and REPNE. Test if the repeat string operation prefix is * REPE/REPZ or REPNE/REPNZ and if it's the case it tests the * corresponding termination condition according to: - * - if REPE/REPZ and ZF = 0 then done - * - if REPNE/REPNZ and ZF = 1 then done + * - if REPE/REPZ and ZF = 0 then done + * - if REPNE/REPNZ and ZF = 1 then done */ if ((c->b == 0xa6) || (c->b == 0xa7) || (c->b == 0xae) || (c->b == 0xaf)) { @@ -1918,7 +1917,7 @@ special_insn: switch (c->b) { case 0x00 ... 0x05: - add: /* add */ + add: /* add */ emulate_2op_SrcV("add", c->src, c->dst, ctxt->eflags); break; case 0x06: /* push es */ @@ -1930,14 +1929,14 @@ special_insn: goto done; break; case 0x08 ... 0x0d: - or: /* or */ + or: /* or */ emulate_2op_SrcV("or", c->src, c->dst, ctxt->eflags); break; case 0x0e: /* push cs */ emulate_push_sreg(ctxt, VCPU_SREG_CS); break; case 0x10 ... 0x15: - adc: /* adc */ + adc: /* adc */ emulate_2op_SrcV("adc", c->src, c->dst, ctxt->eflags); break; case 0x16: /* push ss */ @@ -1949,7 +1948,7 @@ special_insn: goto done; break; case 0x18 ... 0x1d: - sbb: /* sbb */ + sbb: /* sbb */ emulate_2op_SrcV("sbb", c->src, c->dst, ctxt->eflags); break; case 0x1e: /* push ds */ @@ -1961,19 +1960,19 @@ special_insn: goto done; break; case 0x20 ... 0x25: - and: /* and */ + and: /* and */ emulate_2op_SrcV("and", c->src, c->dst, ctxt->eflags); break; case 0x28 ... 0x2d: - sub: /* sub */ + sub: /* sub */ emulate_2op_SrcV("sub", c->src, c->dst, ctxt->eflags); break; case 0x30 ... 0x35: - xor: /* xor */ + xor: /* xor */ emulate_2op_SrcV("xor", c->src, c->dst, ctxt->eflags); break; case 0x38 ... 0x3d: - cmp: /* cmp */ + cmp: /* cmp */ emulate_2op_SrcV("cmp", c->src, c->dst, ctxt->eflags); break; case 0x40 ... 0x47: /* inc r16/r32 */ @@ -2224,11 +2223,11 @@ special_insn: emulate_2op_SrcV("cmp", c->src, c->dst, ctxt->eflags); register_address_increment(c, &c->regs[VCPU_REGS_RSI], - (ctxt->eflags & EFLG_DF) ? -c->src.bytes - : c->src.bytes); + (ctxt->eflags & EFLG_DF) ? -c->src.bytes + : c->src.bytes); register_address_increment(c, &c->regs[VCPU_REGS_RDI], - (ctxt->eflags & EFLG_DF) ? -c->dst.bytes - : c->dst.bytes); + (ctxt->eflags & EFLG_DF) ? -c->dst.bytes + : c->dst.bytes); break; case 0xaa ... 0xab: /* stos */ @@ -2288,8 +2287,8 @@ special_insn: c->src.val = c->regs[VCPU_REGS_RCX]; emulate_grp2(ctxt); break; - case 0xe4: /* inb */ - case 0xe5: /* in */ + case 0xe4: /* inb */ + case 0xe5: /* in */ port = c->src.val; io_dir_in = 1; goto do_io; @@ -2315,7 +2314,7 @@ special_insn: c->eip = c->src.val; break; case 0xeb: - jmp: /* jmp rel short */ + jmp: /* jmp rel short */ jmp_rel(c, c->src.val); c->dst.type = OP_NONE; /* Disable writeback. */ break; @@ -2475,7 +2474,7 @@ twobyte_insn: goto cannot_emulate; } break; - case 0x05: /* syscall */ + case 0x05: /* syscall */ rc = emulate_syscall(ctxt); if (rc != X86EMUL_CONTINUE) goto done; @@ -2581,7 +2580,7 @@ twobyte_insn: goto done; break; case 0xa3: - bt: /* bt */ + bt: /* bt */ c->dst.type = OP_NONE; /* only subword offset */ c->src.val &= (c->dst.bytes << 3) - 1; @@ -2600,7 +2599,7 @@ twobyte_insn: goto done; break; case 0xab: - bts: /* bts */ + bts: /* bts */ /* only subword offset */ c->src.val &= (c->dst.bytes << 3) - 1; emulate_2op_SrcV_nobyte("bts", c->src, c->dst, ctxt->eflags); @@ -2629,7 +2628,7 @@ twobyte_insn: } break; case 0xb3: - btr: /* btr */ + btr: /* btr */ /* only subword offset */ c->src.val &= (c->dst.bytes << 3) - 1; emulate_2op_SrcV_nobyte("btr", c->src, c->dst, ctxt->eflags); @@ -2652,7 +2651,7 @@ twobyte_insn: } break; case 0xbb: - btc: /* btc */ + btc: /* btc */ /* only subword offset */ c->src.val &= (c->dst.bytes << 3) - 1; emulate_2op_SrcV_nobyte("btc", c->src, c->dst, ctxt->eflags); diff --git a/arch/x86/kvm/i8254.c b/arch/x86/kvm/i8254.c index 294698b..720c91a 100644 --- a/arch/x86/kvm/i8254.c +++ b/arch/x86/kvm/i8254.c @@ -332,17 +332,15 @@ static void pit_load_count(struct kvm *kvm, int channel, u32 val) switch (ps->channels[0].mode) { case 0: case 1: - /* FIXME: enhance mode 4 precision */ + /* FIXME: enhance mode 4 precision */ case 4: - if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY)) { + if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY)) create_pit_timer(ps, val, 0); - } break; case 2: case 3: - if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY)){ + if (!(ps->flags & KVM_PIT_FLAGS_HPET_LEGACY)) create_pit_timer(ps, val, 1); - } break; default: destroy_pit_timer(&ps->pit_timer); diff --git a/arch/x86/kvm/i8254.h b/arch/x86/kvm/i8254.h index 900d6b0..5f20d35 100644 --- a/arch/x86/kvm/i8254.h +++ b/arch/x86/kvm/i8254.h @@ -42,12 +42,12 @@ struct kvm_pit { struct kvm_irq_mask_notifier mask_notifier; }; -#define KVM_PIT_BASE_ADDRESS 0x40 -#define KVM_SPEAKER_BASE_ADDRESS 0x61 -#define KVM_PIT_MEM_LENGTH 4 -#define KVM_PIT_FREQ 1193181 -#define KVM_MAX_PIT_INTR_INTERVAL HZ / 100 -#define KVM_PIT_CHANNEL_MASK 0x3 +#define KVM_PIT_BASE_ADDRESS 0x40 +#define KVM_SPEAKER_BASE_ADDRESS 0x61 +#define KVM_PIT_MEM_LENGTH 4 +#define KVM_PIT_FREQ 1193181 +#define KVM_MAX_PIT_INTR_INTERVAL (HZ / 100) +#define KVM_PIT_CHANNEL_MASK 0x3 void kvm_inject_pit_timer_irqs(struct kvm_vcpu *vcpu); void kvm_pit_load_count(struct kvm *kvm, int channel, u32 val, int hpet_legacy_start); diff --git a/arch/x86/kvm/i8259.c b/arch/x86/kvm/i8259.c index 07771da..17a381d 100644 --- a/arch/x86/kvm/i8259.c +++ b/arch/x86/kvm/i8259.c @@ -259,9 +259,8 @@ void kvm_pic_reset(struct kvm_kpic_state *s) for (irq = 0; irq < PIC_NUM_PINS/2; irq++) { if (vcpu0 && kvm_apic_accept_pic_intr(vcpu0)) - if (irr & (1 << irq) || isr & (1 << irq)) { + if (irr & (1 << irq) || isr & (1 << irq)) pic_clear_isr(s, irq); - } } } diff --git a/arch/x86/kvm/kvm_timer.h b/arch/x86/kvm/kvm_timer.h index 55c7524..d2f948a 100644 --- a/arch/x86/kvm/kvm_timer.h +++ b/arch/x86/kvm/kvm_timer.h @@ -1,7 +1,7 @@ struct kvm_timer { struct hrtimer timer; - s64 period; /* unit: ns */ + s64 period; /* unit: ns */ atomic_t pending; /* accumulated triggered timers */ bool reinject; struct kvm_timer_ops *t_ops; @@ -10,9 +10,7 @@ struct kvm_timer { }; struct kvm_timer_ops { - bool (*is_periodic)(struct kvm_timer *); + bool (*is_periodic)(struct kvm_timer *); }; - enum hrtimer_restart kvm_timer_fn(struct hrtimer *data); - diff --git a/arch/x86/kvm/lapic.c b/arch/x86/kvm/lapic.c index 4b224f9..a7286e3 100644 --- a/arch/x86/kvm/lapic.c +++ b/arch/x86/kvm/lapic.c @@ -810,9 +810,9 @@ static int apic_reg_write(struct kvm_lapic *apic, u32 reg, u32 val) break; case APIC_SELF_IPI: - if (apic_x2apic_mode(apic)) { + if (apic_x2apic_mode(apic)) apic_reg_write(apic, APIC_ICR, 0x40000 | (val & 0xff)); - } else + else ret = 1; break; default: @@ -845,7 +845,7 @@ static int apic_mmio_write(struct kvm_io_device *this, return 0; } - val = *(u32*)data; + val = *(u32 *)data; /* too common printing */ if (offset != APIC_EOI) diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c index 741373e..0ccd3b4 100644 --- a/arch/x86/kvm/mmu.c +++ b/arch/x86/kvm/mmu.c @@ -766,7 +766,7 @@ static int kvm_set_pte_rmapp(struct kvm *kvm, unsigned long *rmapp, __set_spte(spte, shadow_trap_nonpresent_pte); spte = rmap_next(kvm, rmapp, NULL); } else { - new_spte = *spte &~ (PT64_BASE_ADDR_MASK); + new_spte = *spte &~(PT64_BASE_ADDR_MASK); new_spte |= (u64)new_pfn << PAGE_SHIFT; new_spte &= ~PT_WRITABLE_MASK; @@ -1118,7 +1118,7 @@ static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp, int i; if (sp->unsync) - for (i=0; i < pvec->nr; i++) + for (i = 0; i < pvec->nr; i++) if (pvec->page[i].sp == sp) return 0; @@ -1234,7 +1234,7 @@ struct mmu_page_path { #define for_each_sp(pvec, sp, parents, i) \ for (i = mmu_pages_next(&pvec, &parents, -1), \ sp = pvec.page[i].sp; \ - i < pvec.nr && ({ sp = pvec.page[i].sp; 1;}); \ + i < pvec.nr && ({ sp = pvec.page[i].sp; 1; }); \ i = mmu_pages_next(&pvec, &parents, i)) static int mmu_pages_next(struct kvm_mmu_pages *pvec, @@ -1550,8 +1550,7 @@ void kvm_mmu_change_mmu_pages(struct kvm *kvm, unsigned int kvm_nr_mmu_pages) used_pages--; } kvm->arch.n_free_mmu_pages = 0; - } - else + } else kvm->arch.n_free_mmu_pages += kvm_nr_mmu_pages - kvm->arch.n_alloc_mmu_pages; @@ -1906,8 +1905,8 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep, pgprintk("%s: setting spte %llx\n", __func__, *sptep); pgprintk("instantiating %s PTE (%s) at %ld (%llx) addr %p\n", - is_large_pte(*sptep)? "2MB" : "4kB", - *sptep & PT_PRESENT_MASK ?"RW":"R", gfn, + is_large_pte(*sptep) ? "2MB" : "4kB", + *sptep & PT_PRESENT_MASK ? "RW" : "R", gfn, *sptep, sptep); if (!was_rmapped && is_large_pte(*sptep)) ++vcpu->kvm->stat.lpages; @@ -2310,7 +2309,7 @@ static void reset_rsvds_bits_mask(struct kvm_vcpu *vcpu, int level) context->rsvd_bits_mask[0][1] = exb_bit_rsvd | rsvd_bits(maxphyaddr, 62); /* PDE */ context->rsvd_bits_mask[0][0] = exb_bit_rsvd | - rsvd_bits(maxphyaddr, 62); /* PTE */ + rsvd_bits(maxphyaddr, 62); /* PTE */ context->rsvd_bits_mask[1][1] = exb_bit_rsvd | rsvd_bits(maxphyaddr, 62) | rsvd_bits(13, 20); /* large page */ @@ -2520,7 +2519,7 @@ static void mmu_pte_write_new_pte(struct kvm_vcpu *vcpu, if (sp->role.level != PT_PAGE_TABLE_LEVEL) { ++vcpu->kvm->stat.mmu_pde_zapped; return; - } + } ++vcpu->kvm->stat.mmu_pte_updated; if (sp->role.glevels == PT32_ROOT_LEVEL) diff --git a/arch/x86/kvm/mmutrace.h b/arch/x86/kvm/mmutrace.h index 3e4a5c6..e540c04 100644 --- a/arch/x86/kvm/mmutrace.h +++ b/arch/x86/kvm/mmutrace.h @@ -26,8 +26,8 @@ static const char *access_str[] = { \ "---", "--x", "w--", "w-x", "-u-", "-ux", "wu-", "wux" \ }; \ - union kvm_mmu_page_role role; \ - \ + union kvm_mmu_page_role role; \ + \ role.word = __entry->role; \ \ trace_seq_printf(p, "sp gfn %llx %u/%u q%u%s %s%s %spge" \ @@ -67,7 +67,7 @@ TRACE_EVENT( TP_fast_assign( __entry->addr = addr; __entry->pferr = (!!write_fault << 1) | (!!user_fault << 2) - | (!!fetch_fault << 4); + | (!!fetch_fault << 4); ), TP_printk("addr %llx pferr %x %s", __entry->addr, __entry->pferr, diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c index 52f78dd..510b654 100644 --- a/arch/x86/kvm/svm.c +++ b/arch/x86/kvm/svm.c @@ -542,16 +542,16 @@ static void init_vmcb(struct vcpu_svm *svm) svm->vcpu.fpu_active = 1; - control->intercept_cr_read = INTERCEPT_CR0_MASK | + control->intercept_cr_read = INTERCEPT_CR0_MASK | INTERCEPT_CR3_MASK | INTERCEPT_CR4_MASK; - control->intercept_cr_write = INTERCEPT_CR0_MASK | + control->intercept_cr_write = INTERCEPT_CR0_MASK | INTERCEPT_CR3_MASK | INTERCEPT_CR4_MASK | INTERCEPT_CR8_MASK; - control->intercept_dr_read = INTERCEPT_DR0_MASK | + control->intercept_dr_read = INTERCEPT_DR0_MASK | INTERCEPT_DR1_MASK | INTERCEPT_DR2_MASK | INTERCEPT_DR3_MASK | @@ -560,7 +560,7 @@ static void init_vmcb(struct vcpu_svm *svm) INTERCEPT_DR6_MASK | INTERCEPT_DR7_MASK; - control->intercept_dr_write = INTERCEPT_DR0_MASK | + control->intercept_dr_write = INTERCEPT_DR0_MASK | INTERCEPT_DR1_MASK | INTERCEPT_DR2_MASK | INTERCEPT_DR3_MASK | @@ -574,7 +574,7 @@ static void init_vmcb(struct vcpu_svm *svm) (1 << MC_VECTOR); - control->intercept = (1ULL << INTERCEPT_INTR) | + control->intercept = (1ULL << INTERCEPT_INTR) | (1ULL << INTERCEPT_NMI) | (1ULL << INTERCEPT_SMI) | (1ULL << INTERCEPT_SELECTIVE_CR0) | @@ -1228,7 +1228,7 @@ static int db_interception(struct vcpu_svm *svm) } if (svm->vcpu.guest_debug & - (KVM_GUESTDBG_SINGLESTEP | KVM_GUESTDBG_USE_HW_BP)){ + (KVM_GUESTDBG_SINGLESTEP | KVM_GUESTDBG_USE_HW_BP)) { kvm_run->exit_reason = KVM_EXIT_DEBUG; kvm_run->debug.arch.pc = svm->vmcb->save.cs.base + svm->vmcb->save.rip; @@ -1565,9 +1565,8 @@ static int nested_svm_exit_handled(struct vcpu_svm *svm) } } - if (vmexit == NESTED_EXIT_DONE) { + if (vmexit == NESTED_EXIT_DONE) nested_svm_vmexit(svm); - } return vmexit; } @@ -1723,7 +1722,7 @@ static bool nested_svm_vmrun_msrpm(struct vcpu_svm *svm) if (!nested_msrpm) return false; - for (i=0; i< PAGE_SIZE * (1 << MSRPM_ALLOC_ORDER) / 4; i++) + for (i = 0; i < PAGE_SIZE * (1 << MSRPM_ALLOC_ORDER) / 4; i++) svm->nested.msrpm[i] = svm->msrpm[i] | nested_msrpm[i]; svm->vmcb->control.msrpm_base_pa = __pa(svm->nested.msrpm); @@ -2320,15 +2319,15 @@ static int pause_interception(struct vcpu_svm *svm) } static int (*svm_exit_handlers[])(struct vcpu_svm *svm) = { - [SVM_EXIT_READ_CR0] = emulate_on_interception, - [SVM_EXIT_READ_CR3] = emulate_on_interception, - [SVM_EXIT_READ_CR4] = emulate_on_interception, - [SVM_EXIT_READ_CR8] = emulate_on_interception, + [SVM_EXIT_READ_CR0] = emulate_on_interception, + [SVM_EXIT_READ_CR3] = emulate_on_interception, + [SVM_EXIT_READ_CR4] = emulate_on_interception, + [SVM_EXIT_READ_CR8] = emulate_on_interception, [SVM_EXIT_CR0_SEL_WRITE] = emulate_on_interception, - [SVM_EXIT_WRITE_CR0] = emulate_on_interception, - [SVM_EXIT_WRITE_CR3] = emulate_on_interception, - [SVM_EXIT_WRITE_CR4] = emulate_on_interception, - [SVM_EXIT_WRITE_CR8] = cr8_write_interception, + [SVM_EXIT_WRITE_CR0] = emulate_on_interception, + [SVM_EXIT_WRITE_CR3] = emulate_on_interception, + [SVM_EXIT_WRITE_CR4] = emulate_on_interception, + [SVM_EXIT_WRITE_CR8] = cr8_write_interception, [SVM_EXIT_READ_DR0] = emulate_on_interception, [SVM_EXIT_READ_DR1] = emulate_on_interception, [SVM_EXIT_READ_DR2] = emulate_on_interception, @@ -2358,13 +2357,13 @@ static int (*svm_exit_handlers[])(struct vcpu_svm *svm) = { [SVM_EXIT_VINTR] = interrupt_window_interception, /* [SVM_EXIT_CR0_SEL_WRITE] = emulate_on_interception, */ [SVM_EXIT_CPUID] = cpuid_interception, - [SVM_EXIT_IRET] = iret_interception, - [SVM_EXIT_INVD] = emulate_on_interception, + [SVM_EXIT_IRET] = iret_interception, + [SVM_EXIT_INVD] = emulate_on_interception, [SVM_EXIT_PAUSE] = pause_interception, [SVM_EXIT_HLT] = halt_interception, [SVM_EXIT_INVLPG] = invlpg_interception, [SVM_EXIT_INVLPGA] = invlpga_interception, - [SVM_EXIT_IOIO] = io_interception, + [SVM_EXIT_IOIO] = io_interception, [SVM_EXIT_MSR] = msr_interception, [SVM_EXIT_TASK_SWITCH] = task_switch_interception, [SVM_EXIT_SHUTDOWN] = shutdown_interception, @@ -2375,7 +2374,7 @@ static int (*svm_exit_handlers[])(struct vcpu_svm *svm) = { [SVM_EXIT_STGI] = stgi_interception, [SVM_EXIT_CLGI] = clgi_interception, [SVM_EXIT_SKINIT] = skinit_interception, - [SVM_EXIT_WBINVD] = emulate_on_interception, + [SVM_EXIT_WBINVD] = emulate_on_interception, [SVM_EXIT_MONITOR] = invalid_op_interception, [SVM_EXIT_MWAIT] = invalid_op_interception, [SVM_EXIT_NPF] = pf_interception, @@ -2874,24 +2873,24 @@ static void svm_cpuid_update(struct kvm_vcpu *vcpu) } static const struct trace_print_flags svm_exit_reasons_str[] = { - { SVM_EXIT_READ_CR0, "read_cr0" }, - { SVM_EXIT_READ_CR3, "read_cr3" }, - { SVM_EXIT_READ_CR4, "read_cr4" }, - { SVM_EXIT_READ_CR8, "read_cr8" }, - { SVM_EXIT_WRITE_CR0, "write_cr0" }, - { SVM_EXIT_WRITE_CR3, "write_cr3" }, - { SVM_EXIT_WRITE_CR4, "write_cr4" }, - { SVM_EXIT_WRITE_CR8, "write_cr8" }, - { SVM_EXIT_READ_DR0, "read_dr0" }, - { SVM_EXIT_READ_DR1, "read_dr1" }, - { SVM_EXIT_READ_DR2, "read_dr2" }, - { SVM_EXIT_READ_DR3, "read_dr3" }, - { SVM_EXIT_WRITE_DR0, "write_dr0" }, - { SVM_EXIT_WRITE_DR1, "write_dr1" }, - { SVM_EXIT_WRITE_DR2, "write_dr2" }, - { SVM_EXIT_WRITE_DR3, "write_dr3" }, - { SVM_EXIT_WRITE_DR5, "write_dr5" }, - { SVM_EXIT_WRITE_DR7, "write_dr7" }, + { SVM_EXIT_READ_CR0, "read_cr0" }, + { SVM_EXIT_READ_CR3, "read_cr3" }, + { SVM_EXIT_READ_CR4, "read_cr4" }, + { SVM_EXIT_READ_CR8, "read_cr8" }, + { SVM_EXIT_WRITE_CR0, "write_cr0" }, + { SVM_EXIT_WRITE_CR3, "write_cr3" }, + { SVM_EXIT_WRITE_CR4, "write_cr4" }, + { SVM_EXIT_WRITE_CR8, "write_cr8" }, + { SVM_EXIT_READ_DR0, "read_dr0" }, + { SVM_EXIT_READ_DR1, "read_dr1" }, + { SVM_EXIT_READ_DR2, "read_dr2" }, + { SVM_EXIT_READ_DR3, "read_dr3" }, + { SVM_EXIT_WRITE_DR0, "write_dr0" }, + { SVM_EXIT_WRITE_DR1, "write_dr1" }, + { SVM_EXIT_WRITE_DR2, "write_dr2" }, + { SVM_EXIT_WRITE_DR3, "write_dr3" }, + { SVM_EXIT_WRITE_DR5, "write_dr5" }, + { SVM_EXIT_WRITE_DR7, "write_dr7" }, { SVM_EXIT_EXCP_BASE + DB_VECTOR, "DB excp" }, { SVM_EXIT_EXCP_BASE + BP_VECTOR, "BP excp" }, { SVM_EXIT_EXCP_BASE + UD_VECTOR, "UD excp" }, diff --git a/arch/x86/kvm/trace.h b/arch/x86/kvm/trace.h index 6ad30a2..22838e6 100644 --- a/arch/x86/kvm/trace.h +++ b/arch/x86/kvm/trace.h @@ -35,7 +35,7 @@ TRACE_EVENT(kvm_hypercall, TP_ARGS(nr, a0, a1, a2, a3), TP_STRUCT__entry( - __field( unsigned long, nr ) + __field( unsigned long, nr ) __field( unsigned long, a0 ) __field( unsigned long, a1 ) __field( unsigned long, a2 ) @@ -96,9 +96,9 @@ TRACE_EVENT(kvm_pio, TP_ARGS(rw, port, size, count), TP_STRUCT__entry( - __field( unsigned int, rw ) - __field( unsigned int, port ) - __field( unsigned int, size ) + __field( unsigned int, rw ) + __field( unsigned int, port ) + __field( unsigned int, size ) __field( unsigned int, count ) ), @@ -543,8 +543,8 @@ TRACE_EVENT(kvm_skinit, ), TP_fast_assign( - __entry->rip = rip; - __entry->slb = slb; + __entry->rip = rip; + __entry->slb = slb; ), TP_printk("rip: 0x%016llx slb: 0x%08x\n", diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c index 14873b9..b53a89b 100644 --- a/arch/x86/kvm/vmx.c +++ b/arch/x86/kvm/vmx.c @@ -108,25 +108,25 @@ struct shared_msr_entry { }; struct vcpu_vmx { - struct kvm_vcpu vcpu; - struct list_head local_vcpus_link; - unsigned long host_rsp; - int launched; - u8 fail; - u32 idt_vectoring_info; - struct shared_msr_entry *guest_msrs; - int nmsrs; - int save_nmsrs; + struct kvm_vcpu vcpu; + struct list_head local_vcpus_link; + unsigned long host_rsp; + int launched; + u8 fail; + u32 idt_vectoring_info; + struct shared_msr_entry *guest_msrs; + int nmsrs; + int save_nmsrs; #ifdef CONFIG_X86_64 - u64 msr_host_kernel_gs_base; - u64 msr_guest_kernel_gs_base; + u64 msr_host_kernel_gs_base; + u64 msr_guest_kernel_gs_base; #endif - struct vmcs *vmcs; + struct vmcs *vmcs; struct { - int loaded; - u16 fs_sel, gs_sel, ldt_sel; - int gs_ldt_reload_needed; - int fs_reload_needed; + int loaded; + u16 fs_sel, gs_sel, ldt_sel; + int gs_ldt_reload_needed; + int fs_reload_needed; } host_state; struct { int vm86_active; @@ -143,8 +143,8 @@ struct vcpu_vmx { unsigned rip; } irq; } rmode; - int vpid; - bool emulation_required; + int vpid; + bool emulation_required; /* Support for vnmi-less CPUs */ int soft_vnmi_blocked; @@ -194,9 +194,9 @@ static struct vmx_capability { #define VMX_SEGMENT_FIELD(seg) \ [VCPU_SREG_##seg] = { \ .selector = GUEST_##seg##_SELECTOR, \ - .base = GUEST_##seg##_BASE, \ - .limit = GUEST_##seg##_LIMIT, \ - .ar_bytes = GUEST_##seg##_AR_BYTES, \ + .base = GUEST_##seg##_BASE, \ + .limit = GUEST_##seg##_LIMIT, \ + .ar_bytes = GUEST_##seg##_AR_BYTES, \ } static struct kvm_vmx_segment_field { @@ -1625,7 +1625,7 @@ static void enter_lmode(struct kvm_vcpu *vcpu) guest_tr_ar = vmcs_read32(GUEST_TR_AR_BYTES); if ((guest_tr_ar & AR_TYPE_MASK) != AR_TYPE_BUSY_64_TSS) { - printk(KERN_DEBUG "%s: tss fixup for long mode. \n", + printk(KERN_DEBUG "%s: tss fixup for long mode.\n", __func__); vmcs_write32(GUEST_TR_AR_BYTES, (guest_tr_ar & ~AR_TYPE_MASK) diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c index e46282a..27f2e1d 100644 --- a/arch/x86/kvm/x86.c +++ b/arch/x86/kvm/x86.c @@ -3742,7 +3742,7 @@ static int kvmclock_cpufreq_notifier(struct notifier_block *nb, unsigned long va } static struct notifier_block kvmclock_cpufreq_notifier_block = { - .notifier_call = kvmclock_cpufreq_notifier + .notifier_call = kvmclock_cpufreq_notifier }; static void kvm_timer_init(void) @@ -4146,9 +4146,9 @@ EXPORT_SYMBOL_GPL(kvm_emulate_cpuid); */ static int dm_request_for_irq_injection(struct kvm_vcpu *vcpu) { - return (!irqchip_in_kernel(vcpu->kvm) && !kvm_cpu_has_interrupt(vcpu) && + return !irqchip_in_kernel(vcpu->kvm) && !kvm_cpu_has_interrupt(vcpu) && vcpu->run->request_interrupt_window && - kvm_arch_interrupt_allowed(vcpu)); + kvm_arch_interrupt_allowed(vcpu); } static void post_kvm_run_save(struct kvm_vcpu *vcpu) @@ -4412,9 +4412,8 @@ static int __vcpu_run(struct kvm_vcpu *vcpu) srcu_read_unlock(&kvm->srcu, vcpu->srcu_idx); kvm_vcpu_block(vcpu); vcpu->srcu_idx = srcu_read_lock(&kvm->srcu); - if (test_and_clear_bit(KVM_REQ_UNHALT, &vcpu->requests)) - { - switch(vcpu->arch.mp_state) { + if (test_and_clear_bit(KVM_REQ_UNHALT, &vcpu->requests)) { + switch (vcpu->arch.mp_state) { case KVM_MP_STATE_HALTED: vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE; @@ -4703,8 +4702,7 @@ static void get_segment_descriptor_dtable(struct kvm_vcpu *vcpu, else dtable->limit = kvm_seg.limit; dtable->base = kvm_seg.base; - } - else + } else kvm_x86_ops->get_gdt(vcpu, dtable); } @@ -4871,7 +4869,7 @@ int kvm_load_segment_descriptor(struct kvm_vcpu *vcpu, u16 selector, int seg) } /* CS(RPL) <- CPL */ selector = (selector & 0xfffc) | cpl; - break; + break; case VCPU_SREG_TR: if (kvm_seg.s || (kvm_seg.type != 1 && kvm_seg.type != 9)) goto exception; @@ -5173,7 +5171,7 @@ int kvm_task_switch(struct kvm_vcpu *vcpu, u16 tss_selector, int reason) } if (reason == TASK_SWITCH_IRET || reason == TASK_SWITCH_JMP) { - cseg_desc.type &= ~(1 << 1); //clear the B flag + cseg_desc.type &= ~(1 << 1); /* clear the B flag */ save_guest_segment_descriptor(vcpu, old_tss_sel, &cseg_desc); } diff --git a/arch/x86/kvm/x86.h b/arch/x86/kvm/x86.h index 2d10163..5b95808 100644 --- a/arch/x86/kvm/x86.h +++ b/arch/x86/kvm/x86.h @@ -34,7 +34,7 @@ static inline bool kvm_exception_is_soft(unsigned int nr) } struct kvm_cpuid_entry2 *kvm_find_cpuid_entry(struct kvm_vcpu *vcpu, - u32 function, u32 index); + u32 function, u32 index); static inline bool is_protmode(struct kvm_vcpu *vcpu) {