Blame view

kernel/linux-imx6_3.14.28/arch/x86/kvm/x86.h 3.35 KB
6b13f685e   김민수   BSP 최초 추가
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
  #ifndef ARCH_X86_KVM_X86_H
  #define ARCH_X86_KVM_X86_H
  
  #include <linux/kvm_host.h>
  #include "kvm_cache_regs.h"
  
  static inline void kvm_clear_exception_queue(struct kvm_vcpu *vcpu)
  {
  	vcpu->arch.exception.pending = false;
  }
  
  static inline void kvm_queue_interrupt(struct kvm_vcpu *vcpu, u8 vector,
  	bool soft)
  {
  	vcpu->arch.interrupt.pending = true;
  	vcpu->arch.interrupt.soft = soft;
  	vcpu->arch.interrupt.nr = vector;
  }
  
  static inline void kvm_clear_interrupt_queue(struct kvm_vcpu *vcpu)
  {
  	vcpu->arch.interrupt.pending = false;
  }
  
  static inline bool kvm_event_needs_reinjection(struct kvm_vcpu *vcpu)
  {
  	return vcpu->arch.exception.pending || vcpu->arch.interrupt.pending ||
  		vcpu->arch.nmi_injected;
  }
  
  static inline bool kvm_exception_is_soft(unsigned int nr)
  {
  	return (nr == BP_VECTOR) || (nr == OF_VECTOR);
  }
  
  static inline bool is_protmode(struct kvm_vcpu *vcpu)
  {
  	return kvm_read_cr0_bits(vcpu, X86_CR0_PE);
  }
  
  static inline int is_long_mode(struct kvm_vcpu *vcpu)
  {
  #ifdef CONFIG_X86_64
  	return vcpu->arch.efer & EFER_LMA;
  #else
  	return 0;
  #endif
  }
  
  static inline bool mmu_is_nested(struct kvm_vcpu *vcpu)
  {
  	return vcpu->arch.walk_mmu == &vcpu->arch.nested_mmu;
  }
  
  static inline int is_pae(struct kvm_vcpu *vcpu)
  {
  	return kvm_read_cr4_bits(vcpu, X86_CR4_PAE);
  }
  
  static inline int is_pse(struct kvm_vcpu *vcpu)
  {
  	return kvm_read_cr4_bits(vcpu, X86_CR4_PSE);
  }
  
  static inline int is_paging(struct kvm_vcpu *vcpu)
  {
  	return likely(kvm_read_cr0_bits(vcpu, X86_CR0_PG));
  }
  
  static inline u32 bit(int bitno)
  {
  	return 1 << (bitno & 31);
  }
  
  static inline void vcpu_cache_mmio_info(struct kvm_vcpu *vcpu,
  					gva_t gva, gfn_t gfn, unsigned access)
  {
  	vcpu->arch.mmio_gva = gva & PAGE_MASK;
  	vcpu->arch.access = access;
  	vcpu->arch.mmio_gfn = gfn;
  	vcpu->arch.mmio_gen = kvm_memslots(vcpu->kvm)->generation;
  }
  
  static inline bool vcpu_match_mmio_gen(struct kvm_vcpu *vcpu)
  {
  	return vcpu->arch.mmio_gen == kvm_memslots(vcpu->kvm)->generation;
  }
  
  /*
   * Clear the mmio cache info for the given gva. If gva is MMIO_GVA_ANY, we
   * clear all mmio cache info.
   */
  #define MMIO_GVA_ANY (~(gva_t)0)
  
  static inline void vcpu_clear_mmio_info(struct kvm_vcpu *vcpu, gva_t gva)
  {
  	if (gva != MMIO_GVA_ANY && vcpu->arch.mmio_gva != (gva & PAGE_MASK))
  		return;
  
  	vcpu->arch.mmio_gva = 0;
  }
  
  static inline bool vcpu_match_mmio_gva(struct kvm_vcpu *vcpu, unsigned long gva)
  {
  	if (vcpu_match_mmio_gen(vcpu) && vcpu->arch.mmio_gva &&
  	      vcpu->arch.mmio_gva == (gva & PAGE_MASK))
  		return true;
  
  	return false;
  }
  
  static inline bool vcpu_match_mmio_gpa(struct kvm_vcpu *vcpu, gpa_t gpa)
  {
  	if (vcpu_match_mmio_gen(vcpu) && vcpu->arch.mmio_gfn &&
  	      vcpu->arch.mmio_gfn == gpa >> PAGE_SHIFT)
  		return true;
  
  	return false;
  }
  
  void kvm_before_handle_nmi(struct kvm_vcpu *vcpu);
  void kvm_after_handle_nmi(struct kvm_vcpu *vcpu);
  int kvm_inject_realmode_interrupt(struct kvm_vcpu *vcpu, int irq, int inc_eip);
  
  void kvm_write_tsc(struct kvm_vcpu *vcpu, struct msr_data *msr);
  
  int kvm_read_guest_virt(struct x86_emulate_ctxt *ctxt,
  	gva_t addr, void *val, unsigned int bytes,
  	struct x86_exception *exception);
  
  int kvm_write_guest_virt_system(struct x86_emulate_ctxt *ctxt,
  	gva_t addr, void *val, unsigned int bytes,
  	struct x86_exception *exception);
  
  #define KVM_SUPPORTED_XCR0	(XSTATE_FP | XSTATE_SSE | XSTATE_YMM)
  extern u64 host_xcr0;
  
  extern unsigned int min_timer_period_us;
  
  extern struct static_key kvm_no_apic_vcpu;
  #endif