xref: /OK3568_Linux_fs/kernel/arch/x86/kvm/lapic.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun 
3*4882a593Smuzhiyun /*
4*4882a593Smuzhiyun  * Local APIC virtualization
5*4882a593Smuzhiyun  *
6*4882a593Smuzhiyun  * Copyright (C) 2006 Qumranet, Inc.
7*4882a593Smuzhiyun  * Copyright (C) 2007 Novell
8*4882a593Smuzhiyun  * Copyright (C) 2007 Intel
9*4882a593Smuzhiyun  * Copyright 2009 Red Hat, Inc. and/or its affiliates.
10*4882a593Smuzhiyun  *
11*4882a593Smuzhiyun  * Authors:
12*4882a593Smuzhiyun  *   Dor Laor <dor.laor@qumranet.com>
13*4882a593Smuzhiyun  *   Gregory Haskins <ghaskins@novell.com>
14*4882a593Smuzhiyun  *   Yaozu (Eddie) Dong <eddie.dong@intel.com>
15*4882a593Smuzhiyun  *
16*4882a593Smuzhiyun  * Based on Xen 3.1 code, Copyright (c) 2004, Intel Corporation.
17*4882a593Smuzhiyun  */
18*4882a593Smuzhiyun 
19*4882a593Smuzhiyun #include <linux/kvm_host.h>
20*4882a593Smuzhiyun #include <linux/kvm.h>
21*4882a593Smuzhiyun #include <linux/mm.h>
22*4882a593Smuzhiyun #include <linux/highmem.h>
23*4882a593Smuzhiyun #include <linux/smp.h>
24*4882a593Smuzhiyun #include <linux/hrtimer.h>
25*4882a593Smuzhiyun #include <linux/io.h>
26*4882a593Smuzhiyun #include <linux/export.h>
27*4882a593Smuzhiyun #include <linux/math64.h>
28*4882a593Smuzhiyun #include <linux/slab.h>
29*4882a593Smuzhiyun #include <asm/processor.h>
30*4882a593Smuzhiyun #include <asm/msr.h>
31*4882a593Smuzhiyun #include <asm/page.h>
32*4882a593Smuzhiyun #include <asm/current.h>
33*4882a593Smuzhiyun #include <asm/apicdef.h>
34*4882a593Smuzhiyun #include <asm/delay.h>
35*4882a593Smuzhiyun #include <linux/atomic.h>
36*4882a593Smuzhiyun #include <linux/jump_label.h>
37*4882a593Smuzhiyun #include "kvm_cache_regs.h"
38*4882a593Smuzhiyun #include "irq.h"
39*4882a593Smuzhiyun #include "ioapic.h"
40*4882a593Smuzhiyun #include "trace.h"
41*4882a593Smuzhiyun #include "x86.h"
42*4882a593Smuzhiyun #include "cpuid.h"
43*4882a593Smuzhiyun #include "hyperv.h"
44*4882a593Smuzhiyun 
45*4882a593Smuzhiyun #ifndef CONFIG_X86_64
46*4882a593Smuzhiyun #define mod_64(x, y) ((x) - (y) * div64_u64(x, y))
47*4882a593Smuzhiyun #else
48*4882a593Smuzhiyun #define mod_64(x, y) ((x) % (y))
49*4882a593Smuzhiyun #endif
50*4882a593Smuzhiyun 
51*4882a593Smuzhiyun #define PRId64 "d"
52*4882a593Smuzhiyun #define PRIx64 "llx"
53*4882a593Smuzhiyun #define PRIu64 "u"
54*4882a593Smuzhiyun #define PRIo64 "o"
55*4882a593Smuzhiyun 
56*4882a593Smuzhiyun /* 14 is the version for Xeon and Pentium 8.4.8*/
57*4882a593Smuzhiyun #define APIC_VERSION			(0x14UL | ((KVM_APIC_LVT_NUM - 1) << 16))
58*4882a593Smuzhiyun #define LAPIC_MMIO_LENGTH		(1 << 12)
59*4882a593Smuzhiyun /* followed define is not in apicdef.h */
60*4882a593Smuzhiyun #define MAX_APIC_VECTOR			256
61*4882a593Smuzhiyun #define APIC_VECTORS_PER_REG		32
62*4882a593Smuzhiyun 
63*4882a593Smuzhiyun static bool lapic_timer_advance_dynamic __read_mostly;
64*4882a593Smuzhiyun #define LAPIC_TIMER_ADVANCE_ADJUST_MIN	100	/* clock cycles */
65*4882a593Smuzhiyun #define LAPIC_TIMER_ADVANCE_ADJUST_MAX	10000	/* clock cycles */
66*4882a593Smuzhiyun #define LAPIC_TIMER_ADVANCE_NS_INIT	1000
67*4882a593Smuzhiyun #define LAPIC_TIMER_ADVANCE_NS_MAX     5000
68*4882a593Smuzhiyun /* step-by-step approximation to mitigate fluctuation */
69*4882a593Smuzhiyun #define LAPIC_TIMER_ADVANCE_ADJUST_STEP 8
70*4882a593Smuzhiyun 
apic_test_vector(int vec,void * bitmap)71*4882a593Smuzhiyun static inline int apic_test_vector(int vec, void *bitmap)
72*4882a593Smuzhiyun {
73*4882a593Smuzhiyun 	return test_bit(VEC_POS(vec), (bitmap) + REG_POS(vec));
74*4882a593Smuzhiyun }
75*4882a593Smuzhiyun 
kvm_apic_pending_eoi(struct kvm_vcpu * vcpu,int vector)76*4882a593Smuzhiyun bool kvm_apic_pending_eoi(struct kvm_vcpu *vcpu, int vector)
77*4882a593Smuzhiyun {
78*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun 	return apic_test_vector(vector, apic->regs + APIC_ISR) ||
81*4882a593Smuzhiyun 		apic_test_vector(vector, apic->regs + APIC_IRR);
82*4882a593Smuzhiyun }
83*4882a593Smuzhiyun 
__apic_test_and_set_vector(int vec,void * bitmap)84*4882a593Smuzhiyun static inline int __apic_test_and_set_vector(int vec, void *bitmap)
85*4882a593Smuzhiyun {
86*4882a593Smuzhiyun 	return __test_and_set_bit(VEC_POS(vec), (bitmap) + REG_POS(vec));
87*4882a593Smuzhiyun }
88*4882a593Smuzhiyun 
__apic_test_and_clear_vector(int vec,void * bitmap)89*4882a593Smuzhiyun static inline int __apic_test_and_clear_vector(int vec, void *bitmap)
90*4882a593Smuzhiyun {
91*4882a593Smuzhiyun 	return __test_and_clear_bit(VEC_POS(vec), (bitmap) + REG_POS(vec));
92*4882a593Smuzhiyun }
93*4882a593Smuzhiyun 
94*4882a593Smuzhiyun struct static_key_deferred apic_hw_disabled __read_mostly;
95*4882a593Smuzhiyun struct static_key_deferred apic_sw_disabled __read_mostly;
96*4882a593Smuzhiyun 
apic_enabled(struct kvm_lapic * apic)97*4882a593Smuzhiyun static inline int apic_enabled(struct kvm_lapic *apic)
98*4882a593Smuzhiyun {
99*4882a593Smuzhiyun 	return kvm_apic_sw_enabled(apic) &&	kvm_apic_hw_enabled(apic);
100*4882a593Smuzhiyun }
101*4882a593Smuzhiyun 
102*4882a593Smuzhiyun #define LVT_MASK	\
103*4882a593Smuzhiyun 	(APIC_LVT_MASKED | APIC_SEND_PENDING | APIC_VECTOR_MASK)
104*4882a593Smuzhiyun 
105*4882a593Smuzhiyun #define LINT_MASK	\
106*4882a593Smuzhiyun 	(LVT_MASK | APIC_MODE_MASK | APIC_INPUT_POLARITY | \
107*4882a593Smuzhiyun 	 APIC_LVT_REMOTE_IRR | APIC_LVT_LEVEL_TRIGGER)
108*4882a593Smuzhiyun 
kvm_x2apic_id(struct kvm_lapic * apic)109*4882a593Smuzhiyun static inline u32 kvm_x2apic_id(struct kvm_lapic *apic)
110*4882a593Smuzhiyun {
111*4882a593Smuzhiyun 	return apic->vcpu->vcpu_id;
112*4882a593Smuzhiyun }
113*4882a593Smuzhiyun 
kvm_can_post_timer_interrupt(struct kvm_vcpu * vcpu)114*4882a593Smuzhiyun static bool kvm_can_post_timer_interrupt(struct kvm_vcpu *vcpu)
115*4882a593Smuzhiyun {
116*4882a593Smuzhiyun 	return pi_inject_timer && kvm_vcpu_apicv_active(vcpu) &&
117*4882a593Smuzhiyun 		(kvm_mwait_in_guest(vcpu->kvm) || kvm_hlt_in_guest(vcpu->kvm));
118*4882a593Smuzhiyun }
119*4882a593Smuzhiyun 
kvm_can_use_hv_timer(struct kvm_vcpu * vcpu)120*4882a593Smuzhiyun bool kvm_can_use_hv_timer(struct kvm_vcpu *vcpu)
121*4882a593Smuzhiyun {
122*4882a593Smuzhiyun 	return kvm_x86_ops.set_hv_timer
123*4882a593Smuzhiyun 	       && !(kvm_mwait_in_guest(vcpu->kvm) ||
124*4882a593Smuzhiyun 		    kvm_can_post_timer_interrupt(vcpu));
125*4882a593Smuzhiyun }
126*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_can_use_hv_timer);
127*4882a593Smuzhiyun 
kvm_use_posted_timer_interrupt(struct kvm_vcpu * vcpu)128*4882a593Smuzhiyun static bool kvm_use_posted_timer_interrupt(struct kvm_vcpu *vcpu)
129*4882a593Smuzhiyun {
130*4882a593Smuzhiyun 	return kvm_can_post_timer_interrupt(vcpu) && vcpu->mode == IN_GUEST_MODE;
131*4882a593Smuzhiyun }
132*4882a593Smuzhiyun 
kvm_apic_map_get_logical_dest(struct kvm_apic_map * map,u32 dest_id,struct kvm_lapic *** cluster,u16 * mask)133*4882a593Smuzhiyun static inline bool kvm_apic_map_get_logical_dest(struct kvm_apic_map *map,
134*4882a593Smuzhiyun 		u32 dest_id, struct kvm_lapic ***cluster, u16 *mask) {
135*4882a593Smuzhiyun 	switch (map->mode) {
136*4882a593Smuzhiyun 	case KVM_APIC_MODE_X2APIC: {
137*4882a593Smuzhiyun 		u32 offset = (dest_id >> 16) * 16;
138*4882a593Smuzhiyun 		u32 max_apic_id = map->max_apic_id;
139*4882a593Smuzhiyun 
140*4882a593Smuzhiyun 		if (offset <= max_apic_id) {
141*4882a593Smuzhiyun 			u8 cluster_size = min(max_apic_id - offset + 1, 16U);
142*4882a593Smuzhiyun 
143*4882a593Smuzhiyun 			offset = array_index_nospec(offset, map->max_apic_id + 1);
144*4882a593Smuzhiyun 			*cluster = &map->phys_map[offset];
145*4882a593Smuzhiyun 			*mask = dest_id & (0xffff >> (16 - cluster_size));
146*4882a593Smuzhiyun 		} else {
147*4882a593Smuzhiyun 			*mask = 0;
148*4882a593Smuzhiyun 		}
149*4882a593Smuzhiyun 
150*4882a593Smuzhiyun 		return true;
151*4882a593Smuzhiyun 		}
152*4882a593Smuzhiyun 	case KVM_APIC_MODE_XAPIC_FLAT:
153*4882a593Smuzhiyun 		*cluster = map->xapic_flat_map;
154*4882a593Smuzhiyun 		*mask = dest_id & 0xff;
155*4882a593Smuzhiyun 		return true;
156*4882a593Smuzhiyun 	case KVM_APIC_MODE_XAPIC_CLUSTER:
157*4882a593Smuzhiyun 		*cluster = map->xapic_cluster_map[(dest_id >> 4) & 0xf];
158*4882a593Smuzhiyun 		*mask = dest_id & 0xf;
159*4882a593Smuzhiyun 		return true;
160*4882a593Smuzhiyun 	default:
161*4882a593Smuzhiyun 		/* Not optimized. */
162*4882a593Smuzhiyun 		return false;
163*4882a593Smuzhiyun 	}
164*4882a593Smuzhiyun }
165*4882a593Smuzhiyun 
kvm_apic_map_free(struct rcu_head * rcu)166*4882a593Smuzhiyun static void kvm_apic_map_free(struct rcu_head *rcu)
167*4882a593Smuzhiyun {
168*4882a593Smuzhiyun 	struct kvm_apic_map *map = container_of(rcu, struct kvm_apic_map, rcu);
169*4882a593Smuzhiyun 
170*4882a593Smuzhiyun 	kvfree(map);
171*4882a593Smuzhiyun }
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun /*
174*4882a593Smuzhiyun  * CLEAN -> DIRTY and UPDATE_IN_PROGRESS -> DIRTY changes happen without a lock.
175*4882a593Smuzhiyun  *
176*4882a593Smuzhiyun  * DIRTY -> UPDATE_IN_PROGRESS and UPDATE_IN_PROGRESS -> CLEAN happen with
177*4882a593Smuzhiyun  * apic_map_lock_held.
178*4882a593Smuzhiyun  */
179*4882a593Smuzhiyun enum {
180*4882a593Smuzhiyun 	CLEAN,
181*4882a593Smuzhiyun 	UPDATE_IN_PROGRESS,
182*4882a593Smuzhiyun 	DIRTY
183*4882a593Smuzhiyun };
184*4882a593Smuzhiyun 
kvm_recalculate_apic_map(struct kvm * kvm)185*4882a593Smuzhiyun void kvm_recalculate_apic_map(struct kvm *kvm)
186*4882a593Smuzhiyun {
187*4882a593Smuzhiyun 	struct kvm_apic_map *new, *old = NULL;
188*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu;
189*4882a593Smuzhiyun 	int i;
190*4882a593Smuzhiyun 	u32 max_id = 255; /* enough space for any xAPIC ID */
191*4882a593Smuzhiyun 
192*4882a593Smuzhiyun 	/* Read kvm->arch.apic_map_dirty before kvm->arch.apic_map.  */
193*4882a593Smuzhiyun 	if (atomic_read_acquire(&kvm->arch.apic_map_dirty) == CLEAN)
194*4882a593Smuzhiyun 		return;
195*4882a593Smuzhiyun 
196*4882a593Smuzhiyun 	mutex_lock(&kvm->arch.apic_map_lock);
197*4882a593Smuzhiyun 	/*
198*4882a593Smuzhiyun 	 * Read kvm->arch.apic_map_dirty before kvm->arch.apic_map
199*4882a593Smuzhiyun 	 * (if clean) or the APIC registers (if dirty).
200*4882a593Smuzhiyun 	 */
201*4882a593Smuzhiyun 	if (atomic_cmpxchg_acquire(&kvm->arch.apic_map_dirty,
202*4882a593Smuzhiyun 				   DIRTY, UPDATE_IN_PROGRESS) == CLEAN) {
203*4882a593Smuzhiyun 		/* Someone else has updated the map. */
204*4882a593Smuzhiyun 		mutex_unlock(&kvm->arch.apic_map_lock);
205*4882a593Smuzhiyun 		return;
206*4882a593Smuzhiyun 	}
207*4882a593Smuzhiyun 
208*4882a593Smuzhiyun 	kvm_for_each_vcpu(i, vcpu, kvm)
209*4882a593Smuzhiyun 		if (kvm_apic_present(vcpu))
210*4882a593Smuzhiyun 			max_id = max(max_id, kvm_x2apic_id(vcpu->arch.apic));
211*4882a593Smuzhiyun 
212*4882a593Smuzhiyun 	new = kvzalloc(sizeof(struct kvm_apic_map) +
213*4882a593Smuzhiyun 	                   sizeof(struct kvm_lapic *) * ((u64)max_id + 1),
214*4882a593Smuzhiyun 			   GFP_KERNEL_ACCOUNT);
215*4882a593Smuzhiyun 
216*4882a593Smuzhiyun 	if (!new)
217*4882a593Smuzhiyun 		goto out;
218*4882a593Smuzhiyun 
219*4882a593Smuzhiyun 	new->max_apic_id = max_id;
220*4882a593Smuzhiyun 
221*4882a593Smuzhiyun 	kvm_for_each_vcpu(i, vcpu, kvm) {
222*4882a593Smuzhiyun 		struct kvm_lapic *apic = vcpu->arch.apic;
223*4882a593Smuzhiyun 		struct kvm_lapic **cluster;
224*4882a593Smuzhiyun 		u16 mask;
225*4882a593Smuzhiyun 		u32 ldr;
226*4882a593Smuzhiyun 		u8 xapic_id;
227*4882a593Smuzhiyun 		u32 x2apic_id;
228*4882a593Smuzhiyun 
229*4882a593Smuzhiyun 		if (!kvm_apic_present(vcpu))
230*4882a593Smuzhiyun 			continue;
231*4882a593Smuzhiyun 
232*4882a593Smuzhiyun 		xapic_id = kvm_xapic_id(apic);
233*4882a593Smuzhiyun 		x2apic_id = kvm_x2apic_id(apic);
234*4882a593Smuzhiyun 
235*4882a593Smuzhiyun 		/* Hotplug hack: see kvm_apic_match_physical_addr(), ... */
236*4882a593Smuzhiyun 		if ((apic_x2apic_mode(apic) || x2apic_id > 0xff) &&
237*4882a593Smuzhiyun 				x2apic_id <= new->max_apic_id)
238*4882a593Smuzhiyun 			new->phys_map[x2apic_id] = apic;
239*4882a593Smuzhiyun 		/*
240*4882a593Smuzhiyun 		 * ... xAPIC ID of VCPUs with APIC ID > 0xff will wrap-around,
241*4882a593Smuzhiyun 		 * prevent them from masking VCPUs with APIC ID <= 0xff.
242*4882a593Smuzhiyun 		 */
243*4882a593Smuzhiyun 		if (!apic_x2apic_mode(apic) && !new->phys_map[xapic_id])
244*4882a593Smuzhiyun 			new->phys_map[xapic_id] = apic;
245*4882a593Smuzhiyun 
246*4882a593Smuzhiyun 		if (!kvm_apic_sw_enabled(apic))
247*4882a593Smuzhiyun 			continue;
248*4882a593Smuzhiyun 
249*4882a593Smuzhiyun 		ldr = kvm_lapic_get_reg(apic, APIC_LDR);
250*4882a593Smuzhiyun 
251*4882a593Smuzhiyun 		if (apic_x2apic_mode(apic)) {
252*4882a593Smuzhiyun 			new->mode |= KVM_APIC_MODE_X2APIC;
253*4882a593Smuzhiyun 		} else if (ldr) {
254*4882a593Smuzhiyun 			ldr = GET_APIC_LOGICAL_ID(ldr);
255*4882a593Smuzhiyun 			if (kvm_lapic_get_reg(apic, APIC_DFR) == APIC_DFR_FLAT)
256*4882a593Smuzhiyun 				new->mode |= KVM_APIC_MODE_XAPIC_FLAT;
257*4882a593Smuzhiyun 			else
258*4882a593Smuzhiyun 				new->mode |= KVM_APIC_MODE_XAPIC_CLUSTER;
259*4882a593Smuzhiyun 		}
260*4882a593Smuzhiyun 
261*4882a593Smuzhiyun 		if (!kvm_apic_map_get_logical_dest(new, ldr, &cluster, &mask))
262*4882a593Smuzhiyun 			continue;
263*4882a593Smuzhiyun 
264*4882a593Smuzhiyun 		if (mask)
265*4882a593Smuzhiyun 			cluster[ffs(mask) - 1] = apic;
266*4882a593Smuzhiyun 	}
267*4882a593Smuzhiyun out:
268*4882a593Smuzhiyun 	old = rcu_dereference_protected(kvm->arch.apic_map,
269*4882a593Smuzhiyun 			lockdep_is_held(&kvm->arch.apic_map_lock));
270*4882a593Smuzhiyun 	rcu_assign_pointer(kvm->arch.apic_map, new);
271*4882a593Smuzhiyun 	/*
272*4882a593Smuzhiyun 	 * Write kvm->arch.apic_map before clearing apic->apic_map_dirty.
273*4882a593Smuzhiyun 	 * If another update has come in, leave it DIRTY.
274*4882a593Smuzhiyun 	 */
275*4882a593Smuzhiyun 	atomic_cmpxchg_release(&kvm->arch.apic_map_dirty,
276*4882a593Smuzhiyun 			       UPDATE_IN_PROGRESS, CLEAN);
277*4882a593Smuzhiyun 	mutex_unlock(&kvm->arch.apic_map_lock);
278*4882a593Smuzhiyun 
279*4882a593Smuzhiyun 	if (old)
280*4882a593Smuzhiyun 		call_rcu(&old->rcu, kvm_apic_map_free);
281*4882a593Smuzhiyun 
282*4882a593Smuzhiyun 	kvm_make_scan_ioapic_request(kvm);
283*4882a593Smuzhiyun }
284*4882a593Smuzhiyun 
apic_set_spiv(struct kvm_lapic * apic,u32 val)285*4882a593Smuzhiyun static inline void apic_set_spiv(struct kvm_lapic *apic, u32 val)
286*4882a593Smuzhiyun {
287*4882a593Smuzhiyun 	bool enabled = val & APIC_SPIV_APIC_ENABLED;
288*4882a593Smuzhiyun 
289*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_SPIV, val);
290*4882a593Smuzhiyun 
291*4882a593Smuzhiyun 	if (enabled != apic->sw_enabled) {
292*4882a593Smuzhiyun 		apic->sw_enabled = enabled;
293*4882a593Smuzhiyun 		if (enabled)
294*4882a593Smuzhiyun 			static_key_slow_dec_deferred(&apic_sw_disabled);
295*4882a593Smuzhiyun 		else
296*4882a593Smuzhiyun 			static_key_slow_inc(&apic_sw_disabled.key);
297*4882a593Smuzhiyun 
298*4882a593Smuzhiyun 		atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
299*4882a593Smuzhiyun 	}
300*4882a593Smuzhiyun 
301*4882a593Smuzhiyun 	/* Check if there are APF page ready requests pending */
302*4882a593Smuzhiyun 	if (enabled)
303*4882a593Smuzhiyun 		kvm_make_request(KVM_REQ_APF_READY, apic->vcpu);
304*4882a593Smuzhiyun }
305*4882a593Smuzhiyun 
kvm_apic_set_xapic_id(struct kvm_lapic * apic,u8 id)306*4882a593Smuzhiyun static inline void kvm_apic_set_xapic_id(struct kvm_lapic *apic, u8 id)
307*4882a593Smuzhiyun {
308*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_ID, id << 24);
309*4882a593Smuzhiyun 	atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
310*4882a593Smuzhiyun }
311*4882a593Smuzhiyun 
kvm_apic_set_ldr(struct kvm_lapic * apic,u32 id)312*4882a593Smuzhiyun static inline void kvm_apic_set_ldr(struct kvm_lapic *apic, u32 id)
313*4882a593Smuzhiyun {
314*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_LDR, id);
315*4882a593Smuzhiyun 	atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
316*4882a593Smuzhiyun }
317*4882a593Smuzhiyun 
kvm_apic_set_dfr(struct kvm_lapic * apic,u32 val)318*4882a593Smuzhiyun static inline void kvm_apic_set_dfr(struct kvm_lapic *apic, u32 val)
319*4882a593Smuzhiyun {
320*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_DFR, val);
321*4882a593Smuzhiyun 	atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
322*4882a593Smuzhiyun }
323*4882a593Smuzhiyun 
kvm_apic_calc_x2apic_ldr(u32 id)324*4882a593Smuzhiyun static inline u32 kvm_apic_calc_x2apic_ldr(u32 id)
325*4882a593Smuzhiyun {
326*4882a593Smuzhiyun 	return ((id >> 4) << 16) | (1 << (id & 0xf));
327*4882a593Smuzhiyun }
328*4882a593Smuzhiyun 
kvm_apic_set_x2apic_id(struct kvm_lapic * apic,u32 id)329*4882a593Smuzhiyun static inline void kvm_apic_set_x2apic_id(struct kvm_lapic *apic, u32 id)
330*4882a593Smuzhiyun {
331*4882a593Smuzhiyun 	u32 ldr = kvm_apic_calc_x2apic_ldr(id);
332*4882a593Smuzhiyun 
333*4882a593Smuzhiyun 	WARN_ON_ONCE(id != apic->vcpu->vcpu_id);
334*4882a593Smuzhiyun 
335*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_ID, id);
336*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_LDR, ldr);
337*4882a593Smuzhiyun 	atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
338*4882a593Smuzhiyun }
339*4882a593Smuzhiyun 
apic_lvt_enabled(struct kvm_lapic * apic,int lvt_type)340*4882a593Smuzhiyun static inline int apic_lvt_enabled(struct kvm_lapic *apic, int lvt_type)
341*4882a593Smuzhiyun {
342*4882a593Smuzhiyun 	return !(kvm_lapic_get_reg(apic, lvt_type) & APIC_LVT_MASKED);
343*4882a593Smuzhiyun }
344*4882a593Smuzhiyun 
apic_lvtt_oneshot(struct kvm_lapic * apic)345*4882a593Smuzhiyun static inline int apic_lvtt_oneshot(struct kvm_lapic *apic)
346*4882a593Smuzhiyun {
347*4882a593Smuzhiyun 	return apic->lapic_timer.timer_mode == APIC_LVT_TIMER_ONESHOT;
348*4882a593Smuzhiyun }
349*4882a593Smuzhiyun 
apic_lvtt_period(struct kvm_lapic * apic)350*4882a593Smuzhiyun static inline int apic_lvtt_period(struct kvm_lapic *apic)
351*4882a593Smuzhiyun {
352*4882a593Smuzhiyun 	return apic->lapic_timer.timer_mode == APIC_LVT_TIMER_PERIODIC;
353*4882a593Smuzhiyun }
354*4882a593Smuzhiyun 
apic_lvtt_tscdeadline(struct kvm_lapic * apic)355*4882a593Smuzhiyun static inline int apic_lvtt_tscdeadline(struct kvm_lapic *apic)
356*4882a593Smuzhiyun {
357*4882a593Smuzhiyun 	return apic->lapic_timer.timer_mode == APIC_LVT_TIMER_TSCDEADLINE;
358*4882a593Smuzhiyun }
359*4882a593Smuzhiyun 
apic_lvt_nmi_mode(u32 lvt_val)360*4882a593Smuzhiyun static inline int apic_lvt_nmi_mode(u32 lvt_val)
361*4882a593Smuzhiyun {
362*4882a593Smuzhiyun 	return (lvt_val & (APIC_MODE_MASK | APIC_LVT_MASKED)) == APIC_DM_NMI;
363*4882a593Smuzhiyun }
364*4882a593Smuzhiyun 
kvm_apic_set_version(struct kvm_vcpu * vcpu)365*4882a593Smuzhiyun void kvm_apic_set_version(struct kvm_vcpu *vcpu)
366*4882a593Smuzhiyun {
367*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
368*4882a593Smuzhiyun 	u32 v = APIC_VERSION;
369*4882a593Smuzhiyun 
370*4882a593Smuzhiyun 	if (!lapic_in_kernel(vcpu))
371*4882a593Smuzhiyun 		return;
372*4882a593Smuzhiyun 
373*4882a593Smuzhiyun 	/*
374*4882a593Smuzhiyun 	 * KVM emulates 82093AA datasheet (with in-kernel IOAPIC implementation)
375*4882a593Smuzhiyun 	 * which doesn't have EOI register; Some buggy OSes (e.g. Windows with
376*4882a593Smuzhiyun 	 * Hyper-V role) disable EOI broadcast in lapic not checking for IOAPIC
377*4882a593Smuzhiyun 	 * version first and level-triggered interrupts never get EOIed in
378*4882a593Smuzhiyun 	 * IOAPIC.
379*4882a593Smuzhiyun 	 */
380*4882a593Smuzhiyun 	if (guest_cpuid_has(vcpu, X86_FEATURE_X2APIC) &&
381*4882a593Smuzhiyun 	    !ioapic_in_kernel(vcpu->kvm))
382*4882a593Smuzhiyun 		v |= APIC_LVR_DIRECTED_EOI;
383*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_LVR, v);
384*4882a593Smuzhiyun }
385*4882a593Smuzhiyun 
386*4882a593Smuzhiyun static const unsigned int apic_lvt_mask[KVM_APIC_LVT_NUM] = {
387*4882a593Smuzhiyun 	LVT_MASK ,      /* part LVTT mask, timer mode mask added at runtime */
388*4882a593Smuzhiyun 	LVT_MASK | APIC_MODE_MASK,	/* LVTTHMR */
389*4882a593Smuzhiyun 	LVT_MASK | APIC_MODE_MASK,	/* LVTPC */
390*4882a593Smuzhiyun 	LINT_MASK, LINT_MASK,	/* LVT0-1 */
391*4882a593Smuzhiyun 	LVT_MASK		/* LVTERR */
392*4882a593Smuzhiyun };
393*4882a593Smuzhiyun 
find_highest_vector(void * bitmap)394*4882a593Smuzhiyun static int find_highest_vector(void *bitmap)
395*4882a593Smuzhiyun {
396*4882a593Smuzhiyun 	int vec;
397*4882a593Smuzhiyun 	u32 *reg;
398*4882a593Smuzhiyun 
399*4882a593Smuzhiyun 	for (vec = MAX_APIC_VECTOR - APIC_VECTORS_PER_REG;
400*4882a593Smuzhiyun 	     vec >= 0; vec -= APIC_VECTORS_PER_REG) {
401*4882a593Smuzhiyun 		reg = bitmap + REG_POS(vec);
402*4882a593Smuzhiyun 		if (*reg)
403*4882a593Smuzhiyun 			return __fls(*reg) + vec;
404*4882a593Smuzhiyun 	}
405*4882a593Smuzhiyun 
406*4882a593Smuzhiyun 	return -1;
407*4882a593Smuzhiyun }
408*4882a593Smuzhiyun 
count_vectors(void * bitmap)409*4882a593Smuzhiyun static u8 count_vectors(void *bitmap)
410*4882a593Smuzhiyun {
411*4882a593Smuzhiyun 	int vec;
412*4882a593Smuzhiyun 	u32 *reg;
413*4882a593Smuzhiyun 	u8 count = 0;
414*4882a593Smuzhiyun 
415*4882a593Smuzhiyun 	for (vec = 0; vec < MAX_APIC_VECTOR; vec += APIC_VECTORS_PER_REG) {
416*4882a593Smuzhiyun 		reg = bitmap + REG_POS(vec);
417*4882a593Smuzhiyun 		count += hweight32(*reg);
418*4882a593Smuzhiyun 	}
419*4882a593Smuzhiyun 
420*4882a593Smuzhiyun 	return count;
421*4882a593Smuzhiyun }
422*4882a593Smuzhiyun 
__kvm_apic_update_irr(u32 * pir,void * regs,int * max_irr)423*4882a593Smuzhiyun bool __kvm_apic_update_irr(u32 *pir, void *regs, int *max_irr)
424*4882a593Smuzhiyun {
425*4882a593Smuzhiyun 	u32 i, vec;
426*4882a593Smuzhiyun 	u32 pir_val, irr_val, prev_irr_val;
427*4882a593Smuzhiyun 	int max_updated_irr;
428*4882a593Smuzhiyun 
429*4882a593Smuzhiyun 	max_updated_irr = -1;
430*4882a593Smuzhiyun 	*max_irr = -1;
431*4882a593Smuzhiyun 
432*4882a593Smuzhiyun 	for (i = vec = 0; i <= 7; i++, vec += 32) {
433*4882a593Smuzhiyun 		pir_val = READ_ONCE(pir[i]);
434*4882a593Smuzhiyun 		irr_val = *((u32 *)(regs + APIC_IRR + i * 0x10));
435*4882a593Smuzhiyun 		if (pir_val) {
436*4882a593Smuzhiyun 			prev_irr_val = irr_val;
437*4882a593Smuzhiyun 			irr_val |= xchg(&pir[i], 0);
438*4882a593Smuzhiyun 			*((u32 *)(regs + APIC_IRR + i * 0x10)) = irr_val;
439*4882a593Smuzhiyun 			if (prev_irr_val != irr_val) {
440*4882a593Smuzhiyun 				max_updated_irr =
441*4882a593Smuzhiyun 					__fls(irr_val ^ prev_irr_val) + vec;
442*4882a593Smuzhiyun 			}
443*4882a593Smuzhiyun 		}
444*4882a593Smuzhiyun 		if (irr_val)
445*4882a593Smuzhiyun 			*max_irr = __fls(irr_val) + vec;
446*4882a593Smuzhiyun 	}
447*4882a593Smuzhiyun 
448*4882a593Smuzhiyun 	return ((max_updated_irr != -1) &&
449*4882a593Smuzhiyun 		(max_updated_irr == *max_irr));
450*4882a593Smuzhiyun }
451*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(__kvm_apic_update_irr);
452*4882a593Smuzhiyun 
kvm_apic_update_irr(struct kvm_vcpu * vcpu,u32 * pir,int * max_irr)453*4882a593Smuzhiyun bool kvm_apic_update_irr(struct kvm_vcpu *vcpu, u32 *pir, int *max_irr)
454*4882a593Smuzhiyun {
455*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
456*4882a593Smuzhiyun 
457*4882a593Smuzhiyun 	return __kvm_apic_update_irr(pir, apic->regs, max_irr);
458*4882a593Smuzhiyun }
459*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_apic_update_irr);
460*4882a593Smuzhiyun 
apic_search_irr(struct kvm_lapic * apic)461*4882a593Smuzhiyun static inline int apic_search_irr(struct kvm_lapic *apic)
462*4882a593Smuzhiyun {
463*4882a593Smuzhiyun 	return find_highest_vector(apic->regs + APIC_IRR);
464*4882a593Smuzhiyun }
465*4882a593Smuzhiyun 
apic_find_highest_irr(struct kvm_lapic * apic)466*4882a593Smuzhiyun static inline int apic_find_highest_irr(struct kvm_lapic *apic)
467*4882a593Smuzhiyun {
468*4882a593Smuzhiyun 	int result;
469*4882a593Smuzhiyun 
470*4882a593Smuzhiyun 	/*
471*4882a593Smuzhiyun 	 * Note that irr_pending is just a hint. It will be always
472*4882a593Smuzhiyun 	 * true with virtual interrupt delivery enabled.
473*4882a593Smuzhiyun 	 */
474*4882a593Smuzhiyun 	if (!apic->irr_pending)
475*4882a593Smuzhiyun 		return -1;
476*4882a593Smuzhiyun 
477*4882a593Smuzhiyun 	result = apic_search_irr(apic);
478*4882a593Smuzhiyun 	ASSERT(result == -1 || result >= 16);
479*4882a593Smuzhiyun 
480*4882a593Smuzhiyun 	return result;
481*4882a593Smuzhiyun }
482*4882a593Smuzhiyun 
apic_clear_irr(int vec,struct kvm_lapic * apic)483*4882a593Smuzhiyun static inline void apic_clear_irr(int vec, struct kvm_lapic *apic)
484*4882a593Smuzhiyun {
485*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu;
486*4882a593Smuzhiyun 
487*4882a593Smuzhiyun 	vcpu = apic->vcpu;
488*4882a593Smuzhiyun 
489*4882a593Smuzhiyun 	if (unlikely(vcpu->arch.apicv_active)) {
490*4882a593Smuzhiyun 		/* need to update RVI */
491*4882a593Smuzhiyun 		kvm_lapic_clear_vector(vec, apic->regs + APIC_IRR);
492*4882a593Smuzhiyun 		kvm_x86_ops.hwapic_irr_update(vcpu,
493*4882a593Smuzhiyun 				apic_find_highest_irr(apic));
494*4882a593Smuzhiyun 	} else {
495*4882a593Smuzhiyun 		apic->irr_pending = false;
496*4882a593Smuzhiyun 		kvm_lapic_clear_vector(vec, apic->regs + APIC_IRR);
497*4882a593Smuzhiyun 		if (apic_search_irr(apic) != -1)
498*4882a593Smuzhiyun 			apic->irr_pending = true;
499*4882a593Smuzhiyun 	}
500*4882a593Smuzhiyun }
501*4882a593Smuzhiyun 
kvm_apic_clear_irr(struct kvm_vcpu * vcpu,int vec)502*4882a593Smuzhiyun void kvm_apic_clear_irr(struct kvm_vcpu *vcpu, int vec)
503*4882a593Smuzhiyun {
504*4882a593Smuzhiyun 	apic_clear_irr(vec, vcpu->arch.apic);
505*4882a593Smuzhiyun }
506*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_apic_clear_irr);
507*4882a593Smuzhiyun 
apic_set_isr(int vec,struct kvm_lapic * apic)508*4882a593Smuzhiyun static inline void apic_set_isr(int vec, struct kvm_lapic *apic)
509*4882a593Smuzhiyun {
510*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu;
511*4882a593Smuzhiyun 
512*4882a593Smuzhiyun 	if (__apic_test_and_set_vector(vec, apic->regs + APIC_ISR))
513*4882a593Smuzhiyun 		return;
514*4882a593Smuzhiyun 
515*4882a593Smuzhiyun 	vcpu = apic->vcpu;
516*4882a593Smuzhiyun 
517*4882a593Smuzhiyun 	/*
518*4882a593Smuzhiyun 	 * With APIC virtualization enabled, all caching is disabled
519*4882a593Smuzhiyun 	 * because the processor can modify ISR under the hood.  Instead
520*4882a593Smuzhiyun 	 * just set SVI.
521*4882a593Smuzhiyun 	 */
522*4882a593Smuzhiyun 	if (unlikely(vcpu->arch.apicv_active))
523*4882a593Smuzhiyun 		kvm_x86_ops.hwapic_isr_update(vcpu, vec);
524*4882a593Smuzhiyun 	else {
525*4882a593Smuzhiyun 		++apic->isr_count;
526*4882a593Smuzhiyun 		BUG_ON(apic->isr_count > MAX_APIC_VECTOR);
527*4882a593Smuzhiyun 		/*
528*4882a593Smuzhiyun 		 * ISR (in service register) bit is set when injecting an interrupt.
529*4882a593Smuzhiyun 		 * The highest vector is injected. Thus the latest bit set matches
530*4882a593Smuzhiyun 		 * the highest bit in ISR.
531*4882a593Smuzhiyun 		 */
532*4882a593Smuzhiyun 		apic->highest_isr_cache = vec;
533*4882a593Smuzhiyun 	}
534*4882a593Smuzhiyun }
535*4882a593Smuzhiyun 
apic_find_highest_isr(struct kvm_lapic * apic)536*4882a593Smuzhiyun static inline int apic_find_highest_isr(struct kvm_lapic *apic)
537*4882a593Smuzhiyun {
538*4882a593Smuzhiyun 	int result;
539*4882a593Smuzhiyun 
540*4882a593Smuzhiyun 	/*
541*4882a593Smuzhiyun 	 * Note that isr_count is always 1, and highest_isr_cache
542*4882a593Smuzhiyun 	 * is always -1, with APIC virtualization enabled.
543*4882a593Smuzhiyun 	 */
544*4882a593Smuzhiyun 	if (!apic->isr_count)
545*4882a593Smuzhiyun 		return -1;
546*4882a593Smuzhiyun 	if (likely(apic->highest_isr_cache != -1))
547*4882a593Smuzhiyun 		return apic->highest_isr_cache;
548*4882a593Smuzhiyun 
549*4882a593Smuzhiyun 	result = find_highest_vector(apic->regs + APIC_ISR);
550*4882a593Smuzhiyun 	ASSERT(result == -1 || result >= 16);
551*4882a593Smuzhiyun 
552*4882a593Smuzhiyun 	return result;
553*4882a593Smuzhiyun }
554*4882a593Smuzhiyun 
apic_clear_isr(int vec,struct kvm_lapic * apic)555*4882a593Smuzhiyun static inline void apic_clear_isr(int vec, struct kvm_lapic *apic)
556*4882a593Smuzhiyun {
557*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu;
558*4882a593Smuzhiyun 	if (!__apic_test_and_clear_vector(vec, apic->regs + APIC_ISR))
559*4882a593Smuzhiyun 		return;
560*4882a593Smuzhiyun 
561*4882a593Smuzhiyun 	vcpu = apic->vcpu;
562*4882a593Smuzhiyun 
563*4882a593Smuzhiyun 	/*
564*4882a593Smuzhiyun 	 * We do get here for APIC virtualization enabled if the guest
565*4882a593Smuzhiyun 	 * uses the Hyper-V APIC enlightenment.  In this case we may need
566*4882a593Smuzhiyun 	 * to trigger a new interrupt delivery by writing the SVI field;
567*4882a593Smuzhiyun 	 * on the other hand isr_count and highest_isr_cache are unused
568*4882a593Smuzhiyun 	 * and must be left alone.
569*4882a593Smuzhiyun 	 */
570*4882a593Smuzhiyun 	if (unlikely(vcpu->arch.apicv_active))
571*4882a593Smuzhiyun 		kvm_x86_ops.hwapic_isr_update(vcpu,
572*4882a593Smuzhiyun 					       apic_find_highest_isr(apic));
573*4882a593Smuzhiyun 	else {
574*4882a593Smuzhiyun 		--apic->isr_count;
575*4882a593Smuzhiyun 		BUG_ON(apic->isr_count < 0);
576*4882a593Smuzhiyun 		apic->highest_isr_cache = -1;
577*4882a593Smuzhiyun 	}
578*4882a593Smuzhiyun }
579*4882a593Smuzhiyun 
kvm_lapic_find_highest_irr(struct kvm_vcpu * vcpu)580*4882a593Smuzhiyun int kvm_lapic_find_highest_irr(struct kvm_vcpu *vcpu)
581*4882a593Smuzhiyun {
582*4882a593Smuzhiyun 	/* This may race with setting of irr in __apic_accept_irq() and
583*4882a593Smuzhiyun 	 * value returned may be wrong, but kvm_vcpu_kick() in __apic_accept_irq
584*4882a593Smuzhiyun 	 * will cause vmexit immediately and the value will be recalculated
585*4882a593Smuzhiyun 	 * on the next vmentry.
586*4882a593Smuzhiyun 	 */
587*4882a593Smuzhiyun 	return apic_find_highest_irr(vcpu->arch.apic);
588*4882a593Smuzhiyun }
589*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_lapic_find_highest_irr);
590*4882a593Smuzhiyun 
591*4882a593Smuzhiyun static int __apic_accept_irq(struct kvm_lapic *apic, int delivery_mode,
592*4882a593Smuzhiyun 			     int vector, int level, int trig_mode,
593*4882a593Smuzhiyun 			     struct dest_map *dest_map);
594*4882a593Smuzhiyun 
kvm_apic_set_irq(struct kvm_vcpu * vcpu,struct kvm_lapic_irq * irq,struct dest_map * dest_map)595*4882a593Smuzhiyun int kvm_apic_set_irq(struct kvm_vcpu *vcpu, struct kvm_lapic_irq *irq,
596*4882a593Smuzhiyun 		     struct dest_map *dest_map)
597*4882a593Smuzhiyun {
598*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
599*4882a593Smuzhiyun 
600*4882a593Smuzhiyun 	return __apic_accept_irq(apic, irq->delivery_mode, irq->vector,
601*4882a593Smuzhiyun 			irq->level, irq->trig_mode, dest_map);
602*4882a593Smuzhiyun }
603*4882a593Smuzhiyun 
__pv_send_ipi(unsigned long * ipi_bitmap,struct kvm_apic_map * map,struct kvm_lapic_irq * irq,u32 min)604*4882a593Smuzhiyun static int __pv_send_ipi(unsigned long *ipi_bitmap, struct kvm_apic_map *map,
605*4882a593Smuzhiyun 			 struct kvm_lapic_irq *irq, u32 min)
606*4882a593Smuzhiyun {
607*4882a593Smuzhiyun 	int i, count = 0;
608*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu;
609*4882a593Smuzhiyun 
610*4882a593Smuzhiyun 	if (min > map->max_apic_id)
611*4882a593Smuzhiyun 		return 0;
612*4882a593Smuzhiyun 
613*4882a593Smuzhiyun 	for_each_set_bit(i, ipi_bitmap,
614*4882a593Smuzhiyun 		min((u32)BITS_PER_LONG, (map->max_apic_id - min + 1))) {
615*4882a593Smuzhiyun 		if (map->phys_map[min + i]) {
616*4882a593Smuzhiyun 			vcpu = map->phys_map[min + i]->vcpu;
617*4882a593Smuzhiyun 			count += kvm_apic_set_irq(vcpu, irq, NULL);
618*4882a593Smuzhiyun 		}
619*4882a593Smuzhiyun 	}
620*4882a593Smuzhiyun 
621*4882a593Smuzhiyun 	return count;
622*4882a593Smuzhiyun }
623*4882a593Smuzhiyun 
kvm_pv_send_ipi(struct kvm * kvm,unsigned long ipi_bitmap_low,unsigned long ipi_bitmap_high,u32 min,unsigned long icr,int op_64_bit)624*4882a593Smuzhiyun int kvm_pv_send_ipi(struct kvm *kvm, unsigned long ipi_bitmap_low,
625*4882a593Smuzhiyun 		    unsigned long ipi_bitmap_high, u32 min,
626*4882a593Smuzhiyun 		    unsigned long icr, int op_64_bit)
627*4882a593Smuzhiyun {
628*4882a593Smuzhiyun 	struct kvm_apic_map *map;
629*4882a593Smuzhiyun 	struct kvm_lapic_irq irq = {0};
630*4882a593Smuzhiyun 	int cluster_size = op_64_bit ? 64 : 32;
631*4882a593Smuzhiyun 	int count;
632*4882a593Smuzhiyun 
633*4882a593Smuzhiyun 	if (icr & (APIC_DEST_MASK | APIC_SHORT_MASK))
634*4882a593Smuzhiyun 		return -KVM_EINVAL;
635*4882a593Smuzhiyun 
636*4882a593Smuzhiyun 	irq.vector = icr & APIC_VECTOR_MASK;
637*4882a593Smuzhiyun 	irq.delivery_mode = icr & APIC_MODE_MASK;
638*4882a593Smuzhiyun 	irq.level = (icr & APIC_INT_ASSERT) != 0;
639*4882a593Smuzhiyun 	irq.trig_mode = icr & APIC_INT_LEVELTRIG;
640*4882a593Smuzhiyun 
641*4882a593Smuzhiyun 	rcu_read_lock();
642*4882a593Smuzhiyun 	map = rcu_dereference(kvm->arch.apic_map);
643*4882a593Smuzhiyun 
644*4882a593Smuzhiyun 	count = -EOPNOTSUPP;
645*4882a593Smuzhiyun 	if (likely(map)) {
646*4882a593Smuzhiyun 		count = __pv_send_ipi(&ipi_bitmap_low, map, &irq, min);
647*4882a593Smuzhiyun 		min += cluster_size;
648*4882a593Smuzhiyun 		count += __pv_send_ipi(&ipi_bitmap_high, map, &irq, min);
649*4882a593Smuzhiyun 	}
650*4882a593Smuzhiyun 
651*4882a593Smuzhiyun 	rcu_read_unlock();
652*4882a593Smuzhiyun 	return count;
653*4882a593Smuzhiyun }
654*4882a593Smuzhiyun 
pv_eoi_put_user(struct kvm_vcpu * vcpu,u8 val)655*4882a593Smuzhiyun static int pv_eoi_put_user(struct kvm_vcpu *vcpu, u8 val)
656*4882a593Smuzhiyun {
657*4882a593Smuzhiyun 
658*4882a593Smuzhiyun 	return kvm_write_guest_cached(vcpu->kvm, &vcpu->arch.pv_eoi.data, &val,
659*4882a593Smuzhiyun 				      sizeof(val));
660*4882a593Smuzhiyun }
661*4882a593Smuzhiyun 
pv_eoi_get_user(struct kvm_vcpu * vcpu,u8 * val)662*4882a593Smuzhiyun static int pv_eoi_get_user(struct kvm_vcpu *vcpu, u8 *val)
663*4882a593Smuzhiyun {
664*4882a593Smuzhiyun 
665*4882a593Smuzhiyun 	return kvm_read_guest_cached(vcpu->kvm, &vcpu->arch.pv_eoi.data, val,
666*4882a593Smuzhiyun 				      sizeof(*val));
667*4882a593Smuzhiyun }
668*4882a593Smuzhiyun 
pv_eoi_enabled(struct kvm_vcpu * vcpu)669*4882a593Smuzhiyun static inline bool pv_eoi_enabled(struct kvm_vcpu *vcpu)
670*4882a593Smuzhiyun {
671*4882a593Smuzhiyun 	return vcpu->arch.pv_eoi.msr_val & KVM_MSR_ENABLED;
672*4882a593Smuzhiyun }
673*4882a593Smuzhiyun 
pv_eoi_get_pending(struct kvm_vcpu * vcpu)674*4882a593Smuzhiyun static bool pv_eoi_get_pending(struct kvm_vcpu *vcpu)
675*4882a593Smuzhiyun {
676*4882a593Smuzhiyun 	u8 val;
677*4882a593Smuzhiyun 	if (pv_eoi_get_user(vcpu, &val) < 0) {
678*4882a593Smuzhiyun 		printk(KERN_WARNING "Can't read EOI MSR value: 0x%llx\n",
679*4882a593Smuzhiyun 			   (unsigned long long)vcpu->arch.pv_eoi.msr_val);
680*4882a593Smuzhiyun 		return false;
681*4882a593Smuzhiyun 	}
682*4882a593Smuzhiyun 	return val & 0x1;
683*4882a593Smuzhiyun }
684*4882a593Smuzhiyun 
pv_eoi_set_pending(struct kvm_vcpu * vcpu)685*4882a593Smuzhiyun static void pv_eoi_set_pending(struct kvm_vcpu *vcpu)
686*4882a593Smuzhiyun {
687*4882a593Smuzhiyun 	if (pv_eoi_put_user(vcpu, KVM_PV_EOI_ENABLED) < 0) {
688*4882a593Smuzhiyun 		printk(KERN_WARNING "Can't set EOI MSR value: 0x%llx\n",
689*4882a593Smuzhiyun 			   (unsigned long long)vcpu->arch.pv_eoi.msr_val);
690*4882a593Smuzhiyun 		return;
691*4882a593Smuzhiyun 	}
692*4882a593Smuzhiyun 	__set_bit(KVM_APIC_PV_EOI_PENDING, &vcpu->arch.apic_attention);
693*4882a593Smuzhiyun }
694*4882a593Smuzhiyun 
pv_eoi_clr_pending(struct kvm_vcpu * vcpu)695*4882a593Smuzhiyun static void pv_eoi_clr_pending(struct kvm_vcpu *vcpu)
696*4882a593Smuzhiyun {
697*4882a593Smuzhiyun 	if (pv_eoi_put_user(vcpu, KVM_PV_EOI_DISABLED) < 0) {
698*4882a593Smuzhiyun 		printk(KERN_WARNING "Can't clear EOI MSR value: 0x%llx\n",
699*4882a593Smuzhiyun 			   (unsigned long long)vcpu->arch.pv_eoi.msr_val);
700*4882a593Smuzhiyun 		return;
701*4882a593Smuzhiyun 	}
702*4882a593Smuzhiyun 	__clear_bit(KVM_APIC_PV_EOI_PENDING, &vcpu->arch.apic_attention);
703*4882a593Smuzhiyun }
704*4882a593Smuzhiyun 
apic_has_interrupt_for_ppr(struct kvm_lapic * apic,u32 ppr)705*4882a593Smuzhiyun static int apic_has_interrupt_for_ppr(struct kvm_lapic *apic, u32 ppr)
706*4882a593Smuzhiyun {
707*4882a593Smuzhiyun 	int highest_irr;
708*4882a593Smuzhiyun 	if (apic->vcpu->arch.apicv_active)
709*4882a593Smuzhiyun 		highest_irr = kvm_x86_ops.sync_pir_to_irr(apic->vcpu);
710*4882a593Smuzhiyun 	else
711*4882a593Smuzhiyun 		highest_irr = apic_find_highest_irr(apic);
712*4882a593Smuzhiyun 	if (highest_irr == -1 || (highest_irr & 0xF0) <= ppr)
713*4882a593Smuzhiyun 		return -1;
714*4882a593Smuzhiyun 	return highest_irr;
715*4882a593Smuzhiyun }
716*4882a593Smuzhiyun 
__apic_update_ppr(struct kvm_lapic * apic,u32 * new_ppr)717*4882a593Smuzhiyun static bool __apic_update_ppr(struct kvm_lapic *apic, u32 *new_ppr)
718*4882a593Smuzhiyun {
719*4882a593Smuzhiyun 	u32 tpr, isrv, ppr, old_ppr;
720*4882a593Smuzhiyun 	int isr;
721*4882a593Smuzhiyun 
722*4882a593Smuzhiyun 	old_ppr = kvm_lapic_get_reg(apic, APIC_PROCPRI);
723*4882a593Smuzhiyun 	tpr = kvm_lapic_get_reg(apic, APIC_TASKPRI);
724*4882a593Smuzhiyun 	isr = apic_find_highest_isr(apic);
725*4882a593Smuzhiyun 	isrv = (isr != -1) ? isr : 0;
726*4882a593Smuzhiyun 
727*4882a593Smuzhiyun 	if ((tpr & 0xf0) >= (isrv & 0xf0))
728*4882a593Smuzhiyun 		ppr = tpr & 0xff;
729*4882a593Smuzhiyun 	else
730*4882a593Smuzhiyun 		ppr = isrv & 0xf0;
731*4882a593Smuzhiyun 
732*4882a593Smuzhiyun 	*new_ppr = ppr;
733*4882a593Smuzhiyun 	if (old_ppr != ppr)
734*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_PROCPRI, ppr);
735*4882a593Smuzhiyun 
736*4882a593Smuzhiyun 	return ppr < old_ppr;
737*4882a593Smuzhiyun }
738*4882a593Smuzhiyun 
apic_update_ppr(struct kvm_lapic * apic)739*4882a593Smuzhiyun static void apic_update_ppr(struct kvm_lapic *apic)
740*4882a593Smuzhiyun {
741*4882a593Smuzhiyun 	u32 ppr;
742*4882a593Smuzhiyun 
743*4882a593Smuzhiyun 	if (__apic_update_ppr(apic, &ppr) &&
744*4882a593Smuzhiyun 	    apic_has_interrupt_for_ppr(apic, ppr) != -1)
745*4882a593Smuzhiyun 		kvm_make_request(KVM_REQ_EVENT, apic->vcpu);
746*4882a593Smuzhiyun }
747*4882a593Smuzhiyun 
kvm_apic_update_ppr(struct kvm_vcpu * vcpu)748*4882a593Smuzhiyun void kvm_apic_update_ppr(struct kvm_vcpu *vcpu)
749*4882a593Smuzhiyun {
750*4882a593Smuzhiyun 	apic_update_ppr(vcpu->arch.apic);
751*4882a593Smuzhiyun }
752*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_apic_update_ppr);
753*4882a593Smuzhiyun 
apic_set_tpr(struct kvm_lapic * apic,u32 tpr)754*4882a593Smuzhiyun static void apic_set_tpr(struct kvm_lapic *apic, u32 tpr)
755*4882a593Smuzhiyun {
756*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_TASKPRI, tpr);
757*4882a593Smuzhiyun 	apic_update_ppr(apic);
758*4882a593Smuzhiyun }
759*4882a593Smuzhiyun 
kvm_apic_broadcast(struct kvm_lapic * apic,u32 mda)760*4882a593Smuzhiyun static bool kvm_apic_broadcast(struct kvm_lapic *apic, u32 mda)
761*4882a593Smuzhiyun {
762*4882a593Smuzhiyun 	return mda == (apic_x2apic_mode(apic) ?
763*4882a593Smuzhiyun 			X2APIC_BROADCAST : APIC_BROADCAST);
764*4882a593Smuzhiyun }
765*4882a593Smuzhiyun 
kvm_apic_match_physical_addr(struct kvm_lapic * apic,u32 mda)766*4882a593Smuzhiyun static bool kvm_apic_match_physical_addr(struct kvm_lapic *apic, u32 mda)
767*4882a593Smuzhiyun {
768*4882a593Smuzhiyun 	if (kvm_apic_broadcast(apic, mda))
769*4882a593Smuzhiyun 		return true;
770*4882a593Smuzhiyun 
771*4882a593Smuzhiyun 	if (apic_x2apic_mode(apic))
772*4882a593Smuzhiyun 		return mda == kvm_x2apic_id(apic);
773*4882a593Smuzhiyun 
774*4882a593Smuzhiyun 	/*
775*4882a593Smuzhiyun 	 * Hotplug hack: Make LAPIC in xAPIC mode also accept interrupts as if
776*4882a593Smuzhiyun 	 * it were in x2APIC mode.  Hotplugged VCPUs start in xAPIC mode and
777*4882a593Smuzhiyun 	 * this allows unique addressing of VCPUs with APIC ID over 0xff.
778*4882a593Smuzhiyun 	 * The 0xff condition is needed because writeable xAPIC ID.
779*4882a593Smuzhiyun 	 */
780*4882a593Smuzhiyun 	if (kvm_x2apic_id(apic) > 0xff && mda == kvm_x2apic_id(apic))
781*4882a593Smuzhiyun 		return true;
782*4882a593Smuzhiyun 
783*4882a593Smuzhiyun 	return mda == kvm_xapic_id(apic);
784*4882a593Smuzhiyun }
785*4882a593Smuzhiyun 
kvm_apic_match_logical_addr(struct kvm_lapic * apic,u32 mda)786*4882a593Smuzhiyun static bool kvm_apic_match_logical_addr(struct kvm_lapic *apic, u32 mda)
787*4882a593Smuzhiyun {
788*4882a593Smuzhiyun 	u32 logical_id;
789*4882a593Smuzhiyun 
790*4882a593Smuzhiyun 	if (kvm_apic_broadcast(apic, mda))
791*4882a593Smuzhiyun 		return true;
792*4882a593Smuzhiyun 
793*4882a593Smuzhiyun 	logical_id = kvm_lapic_get_reg(apic, APIC_LDR);
794*4882a593Smuzhiyun 
795*4882a593Smuzhiyun 	if (apic_x2apic_mode(apic))
796*4882a593Smuzhiyun 		return ((logical_id >> 16) == (mda >> 16))
797*4882a593Smuzhiyun 		       && (logical_id & mda & 0xffff) != 0;
798*4882a593Smuzhiyun 
799*4882a593Smuzhiyun 	logical_id = GET_APIC_LOGICAL_ID(logical_id);
800*4882a593Smuzhiyun 
801*4882a593Smuzhiyun 	switch (kvm_lapic_get_reg(apic, APIC_DFR)) {
802*4882a593Smuzhiyun 	case APIC_DFR_FLAT:
803*4882a593Smuzhiyun 		return (logical_id & mda) != 0;
804*4882a593Smuzhiyun 	case APIC_DFR_CLUSTER:
805*4882a593Smuzhiyun 		return ((logical_id >> 4) == (mda >> 4))
806*4882a593Smuzhiyun 		       && (logical_id & mda & 0xf) != 0;
807*4882a593Smuzhiyun 	default:
808*4882a593Smuzhiyun 		return false;
809*4882a593Smuzhiyun 	}
810*4882a593Smuzhiyun }
811*4882a593Smuzhiyun 
812*4882a593Smuzhiyun /* The KVM local APIC implementation has two quirks:
813*4882a593Smuzhiyun  *
814*4882a593Smuzhiyun  *  - Real hardware delivers interrupts destined to x2APIC ID > 0xff to LAPICs
815*4882a593Smuzhiyun  *    in xAPIC mode if the "destination & 0xff" matches its xAPIC ID.
816*4882a593Smuzhiyun  *    KVM doesn't do that aliasing.
817*4882a593Smuzhiyun  *
818*4882a593Smuzhiyun  *  - in-kernel IOAPIC messages have to be delivered directly to
819*4882a593Smuzhiyun  *    x2APIC, because the kernel does not support interrupt remapping.
820*4882a593Smuzhiyun  *    In order to support broadcast without interrupt remapping, x2APIC
821*4882a593Smuzhiyun  *    rewrites the destination of non-IPI messages from APIC_BROADCAST
822*4882a593Smuzhiyun  *    to X2APIC_BROADCAST.
823*4882a593Smuzhiyun  *
824*4882a593Smuzhiyun  * The broadcast quirk can be disabled with KVM_CAP_X2APIC_API.  This is
825*4882a593Smuzhiyun  * important when userspace wants to use x2APIC-format MSIs, because
826*4882a593Smuzhiyun  * APIC_BROADCAST (0xff) is a legal route for "cluster 0, CPUs 0-7".
827*4882a593Smuzhiyun  */
kvm_apic_mda(struct kvm_vcpu * vcpu,unsigned int dest_id,struct kvm_lapic * source,struct kvm_lapic * target)828*4882a593Smuzhiyun static u32 kvm_apic_mda(struct kvm_vcpu *vcpu, unsigned int dest_id,
829*4882a593Smuzhiyun 		struct kvm_lapic *source, struct kvm_lapic *target)
830*4882a593Smuzhiyun {
831*4882a593Smuzhiyun 	bool ipi = source != NULL;
832*4882a593Smuzhiyun 
833*4882a593Smuzhiyun 	if (!vcpu->kvm->arch.x2apic_broadcast_quirk_disabled &&
834*4882a593Smuzhiyun 	    !ipi && dest_id == APIC_BROADCAST && apic_x2apic_mode(target))
835*4882a593Smuzhiyun 		return X2APIC_BROADCAST;
836*4882a593Smuzhiyun 
837*4882a593Smuzhiyun 	return dest_id;
838*4882a593Smuzhiyun }
839*4882a593Smuzhiyun 
kvm_apic_match_dest(struct kvm_vcpu * vcpu,struct kvm_lapic * source,int shorthand,unsigned int dest,int dest_mode)840*4882a593Smuzhiyun bool kvm_apic_match_dest(struct kvm_vcpu *vcpu, struct kvm_lapic *source,
841*4882a593Smuzhiyun 			   int shorthand, unsigned int dest, int dest_mode)
842*4882a593Smuzhiyun {
843*4882a593Smuzhiyun 	struct kvm_lapic *target = vcpu->arch.apic;
844*4882a593Smuzhiyun 	u32 mda = kvm_apic_mda(vcpu, dest, source, target);
845*4882a593Smuzhiyun 
846*4882a593Smuzhiyun 	ASSERT(target);
847*4882a593Smuzhiyun 	switch (shorthand) {
848*4882a593Smuzhiyun 	case APIC_DEST_NOSHORT:
849*4882a593Smuzhiyun 		if (dest_mode == APIC_DEST_PHYSICAL)
850*4882a593Smuzhiyun 			return kvm_apic_match_physical_addr(target, mda);
851*4882a593Smuzhiyun 		else
852*4882a593Smuzhiyun 			return kvm_apic_match_logical_addr(target, mda);
853*4882a593Smuzhiyun 	case APIC_DEST_SELF:
854*4882a593Smuzhiyun 		return target == source;
855*4882a593Smuzhiyun 	case APIC_DEST_ALLINC:
856*4882a593Smuzhiyun 		return true;
857*4882a593Smuzhiyun 	case APIC_DEST_ALLBUT:
858*4882a593Smuzhiyun 		return target != source;
859*4882a593Smuzhiyun 	default:
860*4882a593Smuzhiyun 		return false;
861*4882a593Smuzhiyun 	}
862*4882a593Smuzhiyun }
863*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_apic_match_dest);
864*4882a593Smuzhiyun 
kvm_vector_to_index(u32 vector,u32 dest_vcpus,const unsigned long * bitmap,u32 bitmap_size)865*4882a593Smuzhiyun int kvm_vector_to_index(u32 vector, u32 dest_vcpus,
866*4882a593Smuzhiyun 		       const unsigned long *bitmap, u32 bitmap_size)
867*4882a593Smuzhiyun {
868*4882a593Smuzhiyun 	u32 mod;
869*4882a593Smuzhiyun 	int i, idx = -1;
870*4882a593Smuzhiyun 
871*4882a593Smuzhiyun 	mod = vector % dest_vcpus;
872*4882a593Smuzhiyun 
873*4882a593Smuzhiyun 	for (i = 0; i <= mod; i++) {
874*4882a593Smuzhiyun 		idx = find_next_bit(bitmap, bitmap_size, idx + 1);
875*4882a593Smuzhiyun 		BUG_ON(idx == bitmap_size);
876*4882a593Smuzhiyun 	}
877*4882a593Smuzhiyun 
878*4882a593Smuzhiyun 	return idx;
879*4882a593Smuzhiyun }
880*4882a593Smuzhiyun 
kvm_apic_disabled_lapic_found(struct kvm * kvm)881*4882a593Smuzhiyun static void kvm_apic_disabled_lapic_found(struct kvm *kvm)
882*4882a593Smuzhiyun {
883*4882a593Smuzhiyun 	if (!kvm->arch.disabled_lapic_found) {
884*4882a593Smuzhiyun 		kvm->arch.disabled_lapic_found = true;
885*4882a593Smuzhiyun 		printk(KERN_INFO
886*4882a593Smuzhiyun 		       "Disabled LAPIC found during irq injection\n");
887*4882a593Smuzhiyun 	}
888*4882a593Smuzhiyun }
889*4882a593Smuzhiyun 
kvm_apic_is_broadcast_dest(struct kvm * kvm,struct kvm_lapic ** src,struct kvm_lapic_irq * irq,struct kvm_apic_map * map)890*4882a593Smuzhiyun static bool kvm_apic_is_broadcast_dest(struct kvm *kvm, struct kvm_lapic **src,
891*4882a593Smuzhiyun 		struct kvm_lapic_irq *irq, struct kvm_apic_map *map)
892*4882a593Smuzhiyun {
893*4882a593Smuzhiyun 	if (kvm->arch.x2apic_broadcast_quirk_disabled) {
894*4882a593Smuzhiyun 		if ((irq->dest_id == APIC_BROADCAST &&
895*4882a593Smuzhiyun 				map->mode != KVM_APIC_MODE_X2APIC))
896*4882a593Smuzhiyun 			return true;
897*4882a593Smuzhiyun 		if (irq->dest_id == X2APIC_BROADCAST)
898*4882a593Smuzhiyun 			return true;
899*4882a593Smuzhiyun 	} else {
900*4882a593Smuzhiyun 		bool x2apic_ipi = src && *src && apic_x2apic_mode(*src);
901*4882a593Smuzhiyun 		if (irq->dest_id == (x2apic_ipi ?
902*4882a593Smuzhiyun 		                     X2APIC_BROADCAST : APIC_BROADCAST))
903*4882a593Smuzhiyun 			return true;
904*4882a593Smuzhiyun 	}
905*4882a593Smuzhiyun 
906*4882a593Smuzhiyun 	return false;
907*4882a593Smuzhiyun }
908*4882a593Smuzhiyun 
909*4882a593Smuzhiyun /* Return true if the interrupt can be handled by using *bitmap as index mask
910*4882a593Smuzhiyun  * for valid destinations in *dst array.
911*4882a593Smuzhiyun  * Return false if kvm_apic_map_get_dest_lapic did nothing useful.
912*4882a593Smuzhiyun  * Note: we may have zero kvm_lapic destinations when we return true, which
913*4882a593Smuzhiyun  * means that the interrupt should be dropped.  In this case, *bitmap would be
914*4882a593Smuzhiyun  * zero and *dst undefined.
915*4882a593Smuzhiyun  */
kvm_apic_map_get_dest_lapic(struct kvm * kvm,struct kvm_lapic ** src,struct kvm_lapic_irq * irq,struct kvm_apic_map * map,struct kvm_lapic *** dst,unsigned long * bitmap)916*4882a593Smuzhiyun static inline bool kvm_apic_map_get_dest_lapic(struct kvm *kvm,
917*4882a593Smuzhiyun 		struct kvm_lapic **src, struct kvm_lapic_irq *irq,
918*4882a593Smuzhiyun 		struct kvm_apic_map *map, struct kvm_lapic ***dst,
919*4882a593Smuzhiyun 		unsigned long *bitmap)
920*4882a593Smuzhiyun {
921*4882a593Smuzhiyun 	int i, lowest;
922*4882a593Smuzhiyun 
923*4882a593Smuzhiyun 	if (irq->shorthand == APIC_DEST_SELF && src) {
924*4882a593Smuzhiyun 		*dst = src;
925*4882a593Smuzhiyun 		*bitmap = 1;
926*4882a593Smuzhiyun 		return true;
927*4882a593Smuzhiyun 	} else if (irq->shorthand)
928*4882a593Smuzhiyun 		return false;
929*4882a593Smuzhiyun 
930*4882a593Smuzhiyun 	if (!map || kvm_apic_is_broadcast_dest(kvm, src, irq, map))
931*4882a593Smuzhiyun 		return false;
932*4882a593Smuzhiyun 
933*4882a593Smuzhiyun 	if (irq->dest_mode == APIC_DEST_PHYSICAL) {
934*4882a593Smuzhiyun 		if (irq->dest_id > map->max_apic_id) {
935*4882a593Smuzhiyun 			*bitmap = 0;
936*4882a593Smuzhiyun 		} else {
937*4882a593Smuzhiyun 			u32 dest_id = array_index_nospec(irq->dest_id, map->max_apic_id + 1);
938*4882a593Smuzhiyun 			*dst = &map->phys_map[dest_id];
939*4882a593Smuzhiyun 			*bitmap = 1;
940*4882a593Smuzhiyun 		}
941*4882a593Smuzhiyun 		return true;
942*4882a593Smuzhiyun 	}
943*4882a593Smuzhiyun 
944*4882a593Smuzhiyun 	*bitmap = 0;
945*4882a593Smuzhiyun 	if (!kvm_apic_map_get_logical_dest(map, irq->dest_id, dst,
946*4882a593Smuzhiyun 				(u16 *)bitmap))
947*4882a593Smuzhiyun 		return false;
948*4882a593Smuzhiyun 
949*4882a593Smuzhiyun 	if (!kvm_lowest_prio_delivery(irq))
950*4882a593Smuzhiyun 		return true;
951*4882a593Smuzhiyun 
952*4882a593Smuzhiyun 	if (!kvm_vector_hashing_enabled()) {
953*4882a593Smuzhiyun 		lowest = -1;
954*4882a593Smuzhiyun 		for_each_set_bit(i, bitmap, 16) {
955*4882a593Smuzhiyun 			if (!(*dst)[i])
956*4882a593Smuzhiyun 				continue;
957*4882a593Smuzhiyun 			if (lowest < 0)
958*4882a593Smuzhiyun 				lowest = i;
959*4882a593Smuzhiyun 			else if (kvm_apic_compare_prio((*dst)[i]->vcpu,
960*4882a593Smuzhiyun 						(*dst)[lowest]->vcpu) < 0)
961*4882a593Smuzhiyun 				lowest = i;
962*4882a593Smuzhiyun 		}
963*4882a593Smuzhiyun 	} else {
964*4882a593Smuzhiyun 		if (!*bitmap)
965*4882a593Smuzhiyun 			return true;
966*4882a593Smuzhiyun 
967*4882a593Smuzhiyun 		lowest = kvm_vector_to_index(irq->vector, hweight16(*bitmap),
968*4882a593Smuzhiyun 				bitmap, 16);
969*4882a593Smuzhiyun 
970*4882a593Smuzhiyun 		if (!(*dst)[lowest]) {
971*4882a593Smuzhiyun 			kvm_apic_disabled_lapic_found(kvm);
972*4882a593Smuzhiyun 			*bitmap = 0;
973*4882a593Smuzhiyun 			return true;
974*4882a593Smuzhiyun 		}
975*4882a593Smuzhiyun 	}
976*4882a593Smuzhiyun 
977*4882a593Smuzhiyun 	*bitmap = (lowest >= 0) ? 1 << lowest : 0;
978*4882a593Smuzhiyun 
979*4882a593Smuzhiyun 	return true;
980*4882a593Smuzhiyun }
981*4882a593Smuzhiyun 
kvm_irq_delivery_to_apic_fast(struct kvm * kvm,struct kvm_lapic * src,struct kvm_lapic_irq * irq,int * r,struct dest_map * dest_map)982*4882a593Smuzhiyun bool kvm_irq_delivery_to_apic_fast(struct kvm *kvm, struct kvm_lapic *src,
983*4882a593Smuzhiyun 		struct kvm_lapic_irq *irq, int *r, struct dest_map *dest_map)
984*4882a593Smuzhiyun {
985*4882a593Smuzhiyun 	struct kvm_apic_map *map;
986*4882a593Smuzhiyun 	unsigned long bitmap;
987*4882a593Smuzhiyun 	struct kvm_lapic **dst = NULL;
988*4882a593Smuzhiyun 	int i;
989*4882a593Smuzhiyun 	bool ret;
990*4882a593Smuzhiyun 
991*4882a593Smuzhiyun 	*r = -1;
992*4882a593Smuzhiyun 
993*4882a593Smuzhiyun 	if (irq->shorthand == APIC_DEST_SELF) {
994*4882a593Smuzhiyun 		if (KVM_BUG_ON(!src, kvm)) {
995*4882a593Smuzhiyun 			*r = 0;
996*4882a593Smuzhiyun 			return true;
997*4882a593Smuzhiyun 		}
998*4882a593Smuzhiyun 		*r = kvm_apic_set_irq(src->vcpu, irq, dest_map);
999*4882a593Smuzhiyun 		return true;
1000*4882a593Smuzhiyun 	}
1001*4882a593Smuzhiyun 
1002*4882a593Smuzhiyun 	rcu_read_lock();
1003*4882a593Smuzhiyun 	map = rcu_dereference(kvm->arch.apic_map);
1004*4882a593Smuzhiyun 
1005*4882a593Smuzhiyun 	ret = kvm_apic_map_get_dest_lapic(kvm, &src, irq, map, &dst, &bitmap);
1006*4882a593Smuzhiyun 	if (ret) {
1007*4882a593Smuzhiyun 		*r = 0;
1008*4882a593Smuzhiyun 		for_each_set_bit(i, &bitmap, 16) {
1009*4882a593Smuzhiyun 			if (!dst[i])
1010*4882a593Smuzhiyun 				continue;
1011*4882a593Smuzhiyun 			*r += kvm_apic_set_irq(dst[i]->vcpu, irq, dest_map);
1012*4882a593Smuzhiyun 		}
1013*4882a593Smuzhiyun 	}
1014*4882a593Smuzhiyun 
1015*4882a593Smuzhiyun 	rcu_read_unlock();
1016*4882a593Smuzhiyun 	return ret;
1017*4882a593Smuzhiyun }
1018*4882a593Smuzhiyun 
1019*4882a593Smuzhiyun /*
1020*4882a593Smuzhiyun  * This routine tries to handle interrupts in posted mode, here is how
1021*4882a593Smuzhiyun  * it deals with different cases:
1022*4882a593Smuzhiyun  * - For single-destination interrupts, handle it in posted mode
1023*4882a593Smuzhiyun  * - Else if vector hashing is enabled and it is a lowest-priority
1024*4882a593Smuzhiyun  *   interrupt, handle it in posted mode and use the following mechanism
1025*4882a593Smuzhiyun  *   to find the destination vCPU.
1026*4882a593Smuzhiyun  *	1. For lowest-priority interrupts, store all the possible
1027*4882a593Smuzhiyun  *	   destination vCPUs in an array.
1028*4882a593Smuzhiyun  *	2. Use "guest vector % max number of destination vCPUs" to find
1029*4882a593Smuzhiyun  *	   the right destination vCPU in the array for the lowest-priority
1030*4882a593Smuzhiyun  *	   interrupt.
1031*4882a593Smuzhiyun  * - Otherwise, use remapped mode to inject the interrupt.
1032*4882a593Smuzhiyun  */
kvm_intr_is_single_vcpu_fast(struct kvm * kvm,struct kvm_lapic_irq * irq,struct kvm_vcpu ** dest_vcpu)1033*4882a593Smuzhiyun bool kvm_intr_is_single_vcpu_fast(struct kvm *kvm, struct kvm_lapic_irq *irq,
1034*4882a593Smuzhiyun 			struct kvm_vcpu **dest_vcpu)
1035*4882a593Smuzhiyun {
1036*4882a593Smuzhiyun 	struct kvm_apic_map *map;
1037*4882a593Smuzhiyun 	unsigned long bitmap;
1038*4882a593Smuzhiyun 	struct kvm_lapic **dst = NULL;
1039*4882a593Smuzhiyun 	bool ret = false;
1040*4882a593Smuzhiyun 
1041*4882a593Smuzhiyun 	if (irq->shorthand)
1042*4882a593Smuzhiyun 		return false;
1043*4882a593Smuzhiyun 
1044*4882a593Smuzhiyun 	rcu_read_lock();
1045*4882a593Smuzhiyun 	map = rcu_dereference(kvm->arch.apic_map);
1046*4882a593Smuzhiyun 
1047*4882a593Smuzhiyun 	if (kvm_apic_map_get_dest_lapic(kvm, NULL, irq, map, &dst, &bitmap) &&
1048*4882a593Smuzhiyun 			hweight16(bitmap) == 1) {
1049*4882a593Smuzhiyun 		unsigned long i = find_first_bit(&bitmap, 16);
1050*4882a593Smuzhiyun 
1051*4882a593Smuzhiyun 		if (dst[i]) {
1052*4882a593Smuzhiyun 			*dest_vcpu = dst[i]->vcpu;
1053*4882a593Smuzhiyun 			ret = true;
1054*4882a593Smuzhiyun 		}
1055*4882a593Smuzhiyun 	}
1056*4882a593Smuzhiyun 
1057*4882a593Smuzhiyun 	rcu_read_unlock();
1058*4882a593Smuzhiyun 	return ret;
1059*4882a593Smuzhiyun }
1060*4882a593Smuzhiyun 
1061*4882a593Smuzhiyun /*
1062*4882a593Smuzhiyun  * Add a pending IRQ into lapic.
1063*4882a593Smuzhiyun  * Return 1 if successfully added and 0 if discarded.
1064*4882a593Smuzhiyun  */
__apic_accept_irq(struct kvm_lapic * apic,int delivery_mode,int vector,int level,int trig_mode,struct dest_map * dest_map)1065*4882a593Smuzhiyun static int __apic_accept_irq(struct kvm_lapic *apic, int delivery_mode,
1066*4882a593Smuzhiyun 			     int vector, int level, int trig_mode,
1067*4882a593Smuzhiyun 			     struct dest_map *dest_map)
1068*4882a593Smuzhiyun {
1069*4882a593Smuzhiyun 	int result = 0;
1070*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu = apic->vcpu;
1071*4882a593Smuzhiyun 
1072*4882a593Smuzhiyun 	trace_kvm_apic_accept_irq(vcpu->vcpu_id, delivery_mode,
1073*4882a593Smuzhiyun 				  trig_mode, vector);
1074*4882a593Smuzhiyun 	switch (delivery_mode) {
1075*4882a593Smuzhiyun 	case APIC_DM_LOWEST:
1076*4882a593Smuzhiyun 		vcpu->arch.apic_arb_prio++;
1077*4882a593Smuzhiyun 		fallthrough;
1078*4882a593Smuzhiyun 	case APIC_DM_FIXED:
1079*4882a593Smuzhiyun 		if (unlikely(trig_mode && !level))
1080*4882a593Smuzhiyun 			break;
1081*4882a593Smuzhiyun 
1082*4882a593Smuzhiyun 		/* FIXME add logic for vcpu on reset */
1083*4882a593Smuzhiyun 		if (unlikely(!apic_enabled(apic)))
1084*4882a593Smuzhiyun 			break;
1085*4882a593Smuzhiyun 
1086*4882a593Smuzhiyun 		result = 1;
1087*4882a593Smuzhiyun 
1088*4882a593Smuzhiyun 		if (dest_map) {
1089*4882a593Smuzhiyun 			__set_bit(vcpu->vcpu_id, dest_map->map);
1090*4882a593Smuzhiyun 			dest_map->vectors[vcpu->vcpu_id] = vector;
1091*4882a593Smuzhiyun 		}
1092*4882a593Smuzhiyun 
1093*4882a593Smuzhiyun 		if (apic_test_vector(vector, apic->regs + APIC_TMR) != !!trig_mode) {
1094*4882a593Smuzhiyun 			if (trig_mode)
1095*4882a593Smuzhiyun 				kvm_lapic_set_vector(vector,
1096*4882a593Smuzhiyun 						     apic->regs + APIC_TMR);
1097*4882a593Smuzhiyun 			else
1098*4882a593Smuzhiyun 				kvm_lapic_clear_vector(vector,
1099*4882a593Smuzhiyun 						       apic->regs + APIC_TMR);
1100*4882a593Smuzhiyun 		}
1101*4882a593Smuzhiyun 
1102*4882a593Smuzhiyun 		if (kvm_x86_ops.deliver_posted_interrupt(vcpu, vector)) {
1103*4882a593Smuzhiyun 			kvm_lapic_set_irr(vector, apic);
1104*4882a593Smuzhiyun 			kvm_make_request(KVM_REQ_EVENT, vcpu);
1105*4882a593Smuzhiyun 			kvm_vcpu_kick(vcpu);
1106*4882a593Smuzhiyun 		}
1107*4882a593Smuzhiyun 		break;
1108*4882a593Smuzhiyun 
1109*4882a593Smuzhiyun 	case APIC_DM_REMRD:
1110*4882a593Smuzhiyun 		result = 1;
1111*4882a593Smuzhiyun 		vcpu->arch.pv.pv_unhalted = 1;
1112*4882a593Smuzhiyun 		kvm_make_request(KVM_REQ_EVENT, vcpu);
1113*4882a593Smuzhiyun 		kvm_vcpu_kick(vcpu);
1114*4882a593Smuzhiyun 		break;
1115*4882a593Smuzhiyun 
1116*4882a593Smuzhiyun 	case APIC_DM_SMI:
1117*4882a593Smuzhiyun 		result = 1;
1118*4882a593Smuzhiyun 		kvm_make_request(KVM_REQ_SMI, vcpu);
1119*4882a593Smuzhiyun 		kvm_vcpu_kick(vcpu);
1120*4882a593Smuzhiyun 		break;
1121*4882a593Smuzhiyun 
1122*4882a593Smuzhiyun 	case APIC_DM_NMI:
1123*4882a593Smuzhiyun 		result = 1;
1124*4882a593Smuzhiyun 		kvm_inject_nmi(vcpu);
1125*4882a593Smuzhiyun 		kvm_vcpu_kick(vcpu);
1126*4882a593Smuzhiyun 		break;
1127*4882a593Smuzhiyun 
1128*4882a593Smuzhiyun 	case APIC_DM_INIT:
1129*4882a593Smuzhiyun 		if (!trig_mode || level) {
1130*4882a593Smuzhiyun 			result = 1;
1131*4882a593Smuzhiyun 			/* assumes that there are only KVM_APIC_INIT/SIPI */
1132*4882a593Smuzhiyun 			apic->pending_events = (1UL << KVM_APIC_INIT);
1133*4882a593Smuzhiyun 			kvm_make_request(KVM_REQ_EVENT, vcpu);
1134*4882a593Smuzhiyun 			kvm_vcpu_kick(vcpu);
1135*4882a593Smuzhiyun 		}
1136*4882a593Smuzhiyun 		break;
1137*4882a593Smuzhiyun 
1138*4882a593Smuzhiyun 	case APIC_DM_STARTUP:
1139*4882a593Smuzhiyun 		result = 1;
1140*4882a593Smuzhiyun 		apic->sipi_vector = vector;
1141*4882a593Smuzhiyun 		/* make sure sipi_vector is visible for the receiver */
1142*4882a593Smuzhiyun 		smp_wmb();
1143*4882a593Smuzhiyun 		set_bit(KVM_APIC_SIPI, &apic->pending_events);
1144*4882a593Smuzhiyun 		kvm_make_request(KVM_REQ_EVENT, vcpu);
1145*4882a593Smuzhiyun 		kvm_vcpu_kick(vcpu);
1146*4882a593Smuzhiyun 		break;
1147*4882a593Smuzhiyun 
1148*4882a593Smuzhiyun 	case APIC_DM_EXTINT:
1149*4882a593Smuzhiyun 		/*
1150*4882a593Smuzhiyun 		 * Should only be called by kvm_apic_local_deliver() with LVT0,
1151*4882a593Smuzhiyun 		 * before NMI watchdog was enabled. Already handled by
1152*4882a593Smuzhiyun 		 * kvm_apic_accept_pic_intr().
1153*4882a593Smuzhiyun 		 */
1154*4882a593Smuzhiyun 		break;
1155*4882a593Smuzhiyun 
1156*4882a593Smuzhiyun 	default:
1157*4882a593Smuzhiyun 		printk(KERN_ERR "TODO: unsupported delivery mode %x\n",
1158*4882a593Smuzhiyun 		       delivery_mode);
1159*4882a593Smuzhiyun 		break;
1160*4882a593Smuzhiyun 	}
1161*4882a593Smuzhiyun 	return result;
1162*4882a593Smuzhiyun }
1163*4882a593Smuzhiyun 
1164*4882a593Smuzhiyun /*
1165*4882a593Smuzhiyun  * This routine identifies the destination vcpus mask meant to receive the
1166*4882a593Smuzhiyun  * IOAPIC interrupts. It either uses kvm_apic_map_get_dest_lapic() to find
1167*4882a593Smuzhiyun  * out the destination vcpus array and set the bitmap or it traverses to
1168*4882a593Smuzhiyun  * each available vcpu to identify the same.
1169*4882a593Smuzhiyun  */
kvm_bitmap_or_dest_vcpus(struct kvm * kvm,struct kvm_lapic_irq * irq,unsigned long * vcpu_bitmap)1170*4882a593Smuzhiyun void kvm_bitmap_or_dest_vcpus(struct kvm *kvm, struct kvm_lapic_irq *irq,
1171*4882a593Smuzhiyun 			      unsigned long *vcpu_bitmap)
1172*4882a593Smuzhiyun {
1173*4882a593Smuzhiyun 	struct kvm_lapic **dest_vcpu = NULL;
1174*4882a593Smuzhiyun 	struct kvm_lapic *src = NULL;
1175*4882a593Smuzhiyun 	struct kvm_apic_map *map;
1176*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu;
1177*4882a593Smuzhiyun 	unsigned long bitmap;
1178*4882a593Smuzhiyun 	int i, vcpu_idx;
1179*4882a593Smuzhiyun 	bool ret;
1180*4882a593Smuzhiyun 
1181*4882a593Smuzhiyun 	rcu_read_lock();
1182*4882a593Smuzhiyun 	map = rcu_dereference(kvm->arch.apic_map);
1183*4882a593Smuzhiyun 
1184*4882a593Smuzhiyun 	ret = kvm_apic_map_get_dest_lapic(kvm, &src, irq, map, &dest_vcpu,
1185*4882a593Smuzhiyun 					  &bitmap);
1186*4882a593Smuzhiyun 	if (ret) {
1187*4882a593Smuzhiyun 		for_each_set_bit(i, &bitmap, 16) {
1188*4882a593Smuzhiyun 			if (!dest_vcpu[i])
1189*4882a593Smuzhiyun 				continue;
1190*4882a593Smuzhiyun 			vcpu_idx = dest_vcpu[i]->vcpu->vcpu_idx;
1191*4882a593Smuzhiyun 			__set_bit(vcpu_idx, vcpu_bitmap);
1192*4882a593Smuzhiyun 		}
1193*4882a593Smuzhiyun 	} else {
1194*4882a593Smuzhiyun 		kvm_for_each_vcpu(i, vcpu, kvm) {
1195*4882a593Smuzhiyun 			if (!kvm_apic_present(vcpu))
1196*4882a593Smuzhiyun 				continue;
1197*4882a593Smuzhiyun 			if (!kvm_apic_match_dest(vcpu, NULL,
1198*4882a593Smuzhiyun 						 irq->shorthand,
1199*4882a593Smuzhiyun 						 irq->dest_id,
1200*4882a593Smuzhiyun 						 irq->dest_mode))
1201*4882a593Smuzhiyun 				continue;
1202*4882a593Smuzhiyun 			__set_bit(i, vcpu_bitmap);
1203*4882a593Smuzhiyun 		}
1204*4882a593Smuzhiyun 	}
1205*4882a593Smuzhiyun 	rcu_read_unlock();
1206*4882a593Smuzhiyun }
1207*4882a593Smuzhiyun 
kvm_apic_compare_prio(struct kvm_vcpu * vcpu1,struct kvm_vcpu * vcpu2)1208*4882a593Smuzhiyun int kvm_apic_compare_prio(struct kvm_vcpu *vcpu1, struct kvm_vcpu *vcpu2)
1209*4882a593Smuzhiyun {
1210*4882a593Smuzhiyun 	return vcpu1->arch.apic_arb_prio - vcpu2->arch.apic_arb_prio;
1211*4882a593Smuzhiyun }
1212*4882a593Smuzhiyun 
kvm_ioapic_handles_vector(struct kvm_lapic * apic,int vector)1213*4882a593Smuzhiyun static bool kvm_ioapic_handles_vector(struct kvm_lapic *apic, int vector)
1214*4882a593Smuzhiyun {
1215*4882a593Smuzhiyun 	return test_bit(vector, apic->vcpu->arch.ioapic_handled_vectors);
1216*4882a593Smuzhiyun }
1217*4882a593Smuzhiyun 
kvm_ioapic_send_eoi(struct kvm_lapic * apic,int vector)1218*4882a593Smuzhiyun static void kvm_ioapic_send_eoi(struct kvm_lapic *apic, int vector)
1219*4882a593Smuzhiyun {
1220*4882a593Smuzhiyun 	int trigger_mode;
1221*4882a593Smuzhiyun 
1222*4882a593Smuzhiyun 	/* Eoi the ioapic only if the ioapic doesn't own the vector. */
1223*4882a593Smuzhiyun 	if (!kvm_ioapic_handles_vector(apic, vector))
1224*4882a593Smuzhiyun 		return;
1225*4882a593Smuzhiyun 
1226*4882a593Smuzhiyun 	/* Request a KVM exit to inform the userspace IOAPIC. */
1227*4882a593Smuzhiyun 	if (irqchip_split(apic->vcpu->kvm)) {
1228*4882a593Smuzhiyun 		apic->vcpu->arch.pending_ioapic_eoi = vector;
1229*4882a593Smuzhiyun 		kvm_make_request(KVM_REQ_IOAPIC_EOI_EXIT, apic->vcpu);
1230*4882a593Smuzhiyun 		return;
1231*4882a593Smuzhiyun 	}
1232*4882a593Smuzhiyun 
1233*4882a593Smuzhiyun 	if (apic_test_vector(vector, apic->regs + APIC_TMR))
1234*4882a593Smuzhiyun 		trigger_mode = IOAPIC_LEVEL_TRIG;
1235*4882a593Smuzhiyun 	else
1236*4882a593Smuzhiyun 		trigger_mode = IOAPIC_EDGE_TRIG;
1237*4882a593Smuzhiyun 
1238*4882a593Smuzhiyun 	kvm_ioapic_update_eoi(apic->vcpu, vector, trigger_mode);
1239*4882a593Smuzhiyun }
1240*4882a593Smuzhiyun 
apic_set_eoi(struct kvm_lapic * apic)1241*4882a593Smuzhiyun static int apic_set_eoi(struct kvm_lapic *apic)
1242*4882a593Smuzhiyun {
1243*4882a593Smuzhiyun 	int vector = apic_find_highest_isr(apic);
1244*4882a593Smuzhiyun 
1245*4882a593Smuzhiyun 	trace_kvm_eoi(apic, vector);
1246*4882a593Smuzhiyun 
1247*4882a593Smuzhiyun 	/*
1248*4882a593Smuzhiyun 	 * Not every write EOI will has corresponding ISR,
1249*4882a593Smuzhiyun 	 * one example is when Kernel check timer on setup_IO_APIC
1250*4882a593Smuzhiyun 	 */
1251*4882a593Smuzhiyun 	if (vector == -1)
1252*4882a593Smuzhiyun 		return vector;
1253*4882a593Smuzhiyun 
1254*4882a593Smuzhiyun 	apic_clear_isr(vector, apic);
1255*4882a593Smuzhiyun 	apic_update_ppr(apic);
1256*4882a593Smuzhiyun 
1257*4882a593Smuzhiyun 	if (test_bit(vector, vcpu_to_synic(apic->vcpu)->vec_bitmap))
1258*4882a593Smuzhiyun 		kvm_hv_synic_send_eoi(apic->vcpu, vector);
1259*4882a593Smuzhiyun 
1260*4882a593Smuzhiyun 	kvm_ioapic_send_eoi(apic, vector);
1261*4882a593Smuzhiyun 	kvm_make_request(KVM_REQ_EVENT, apic->vcpu);
1262*4882a593Smuzhiyun 	return vector;
1263*4882a593Smuzhiyun }
1264*4882a593Smuzhiyun 
1265*4882a593Smuzhiyun /*
1266*4882a593Smuzhiyun  * this interface assumes a trap-like exit, which has already finished
1267*4882a593Smuzhiyun  * desired side effect including vISR and vPPR update.
1268*4882a593Smuzhiyun  */
kvm_apic_set_eoi_accelerated(struct kvm_vcpu * vcpu,int vector)1269*4882a593Smuzhiyun void kvm_apic_set_eoi_accelerated(struct kvm_vcpu *vcpu, int vector)
1270*4882a593Smuzhiyun {
1271*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
1272*4882a593Smuzhiyun 
1273*4882a593Smuzhiyun 	trace_kvm_eoi(apic, vector);
1274*4882a593Smuzhiyun 
1275*4882a593Smuzhiyun 	kvm_ioapic_send_eoi(apic, vector);
1276*4882a593Smuzhiyun 	kvm_make_request(KVM_REQ_EVENT, apic->vcpu);
1277*4882a593Smuzhiyun }
1278*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_apic_set_eoi_accelerated);
1279*4882a593Smuzhiyun 
kvm_apic_send_ipi(struct kvm_lapic * apic,u32 icr_low,u32 icr_high)1280*4882a593Smuzhiyun void kvm_apic_send_ipi(struct kvm_lapic *apic, u32 icr_low, u32 icr_high)
1281*4882a593Smuzhiyun {
1282*4882a593Smuzhiyun 	struct kvm_lapic_irq irq;
1283*4882a593Smuzhiyun 
1284*4882a593Smuzhiyun 	irq.vector = icr_low & APIC_VECTOR_MASK;
1285*4882a593Smuzhiyun 	irq.delivery_mode = icr_low & APIC_MODE_MASK;
1286*4882a593Smuzhiyun 	irq.dest_mode = icr_low & APIC_DEST_MASK;
1287*4882a593Smuzhiyun 	irq.level = (icr_low & APIC_INT_ASSERT) != 0;
1288*4882a593Smuzhiyun 	irq.trig_mode = icr_low & APIC_INT_LEVELTRIG;
1289*4882a593Smuzhiyun 	irq.shorthand = icr_low & APIC_SHORT_MASK;
1290*4882a593Smuzhiyun 	irq.msi_redir_hint = false;
1291*4882a593Smuzhiyun 	if (apic_x2apic_mode(apic))
1292*4882a593Smuzhiyun 		irq.dest_id = icr_high;
1293*4882a593Smuzhiyun 	else
1294*4882a593Smuzhiyun 		irq.dest_id = GET_APIC_DEST_FIELD(icr_high);
1295*4882a593Smuzhiyun 
1296*4882a593Smuzhiyun 	trace_kvm_apic_ipi(icr_low, irq.dest_id);
1297*4882a593Smuzhiyun 
1298*4882a593Smuzhiyun 	kvm_irq_delivery_to_apic(apic->vcpu->kvm, apic, &irq, NULL);
1299*4882a593Smuzhiyun }
1300*4882a593Smuzhiyun 
apic_get_tmcct(struct kvm_lapic * apic)1301*4882a593Smuzhiyun static u32 apic_get_tmcct(struct kvm_lapic *apic)
1302*4882a593Smuzhiyun {
1303*4882a593Smuzhiyun 	ktime_t remaining, now;
1304*4882a593Smuzhiyun 	s64 ns;
1305*4882a593Smuzhiyun 	u32 tmcct;
1306*4882a593Smuzhiyun 
1307*4882a593Smuzhiyun 	ASSERT(apic != NULL);
1308*4882a593Smuzhiyun 
1309*4882a593Smuzhiyun 	/* if initial count is 0, current count should also be 0 */
1310*4882a593Smuzhiyun 	if (kvm_lapic_get_reg(apic, APIC_TMICT) == 0 ||
1311*4882a593Smuzhiyun 		apic->lapic_timer.period == 0)
1312*4882a593Smuzhiyun 		return 0;
1313*4882a593Smuzhiyun 
1314*4882a593Smuzhiyun 	now = ktime_get();
1315*4882a593Smuzhiyun 	remaining = ktime_sub(apic->lapic_timer.target_expiration, now);
1316*4882a593Smuzhiyun 	if (ktime_to_ns(remaining) < 0)
1317*4882a593Smuzhiyun 		remaining = 0;
1318*4882a593Smuzhiyun 
1319*4882a593Smuzhiyun 	ns = mod_64(ktime_to_ns(remaining), apic->lapic_timer.period);
1320*4882a593Smuzhiyun 	tmcct = div64_u64(ns,
1321*4882a593Smuzhiyun 			 (APIC_BUS_CYCLE_NS * apic->divide_count));
1322*4882a593Smuzhiyun 
1323*4882a593Smuzhiyun 	return tmcct;
1324*4882a593Smuzhiyun }
1325*4882a593Smuzhiyun 
__report_tpr_access(struct kvm_lapic * apic,bool write)1326*4882a593Smuzhiyun static void __report_tpr_access(struct kvm_lapic *apic, bool write)
1327*4882a593Smuzhiyun {
1328*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu = apic->vcpu;
1329*4882a593Smuzhiyun 	struct kvm_run *run = vcpu->run;
1330*4882a593Smuzhiyun 
1331*4882a593Smuzhiyun 	kvm_make_request(KVM_REQ_REPORT_TPR_ACCESS, vcpu);
1332*4882a593Smuzhiyun 	run->tpr_access.rip = kvm_rip_read(vcpu);
1333*4882a593Smuzhiyun 	run->tpr_access.is_write = write;
1334*4882a593Smuzhiyun }
1335*4882a593Smuzhiyun 
report_tpr_access(struct kvm_lapic * apic,bool write)1336*4882a593Smuzhiyun static inline void report_tpr_access(struct kvm_lapic *apic, bool write)
1337*4882a593Smuzhiyun {
1338*4882a593Smuzhiyun 	if (apic->vcpu->arch.tpr_access_reporting)
1339*4882a593Smuzhiyun 		__report_tpr_access(apic, write);
1340*4882a593Smuzhiyun }
1341*4882a593Smuzhiyun 
__apic_read(struct kvm_lapic * apic,unsigned int offset)1342*4882a593Smuzhiyun static u32 __apic_read(struct kvm_lapic *apic, unsigned int offset)
1343*4882a593Smuzhiyun {
1344*4882a593Smuzhiyun 	u32 val = 0;
1345*4882a593Smuzhiyun 
1346*4882a593Smuzhiyun 	if (offset >= LAPIC_MMIO_LENGTH)
1347*4882a593Smuzhiyun 		return 0;
1348*4882a593Smuzhiyun 
1349*4882a593Smuzhiyun 	switch (offset) {
1350*4882a593Smuzhiyun 	case APIC_ARBPRI:
1351*4882a593Smuzhiyun 		break;
1352*4882a593Smuzhiyun 
1353*4882a593Smuzhiyun 	case APIC_TMCCT:	/* Timer CCR */
1354*4882a593Smuzhiyun 		if (apic_lvtt_tscdeadline(apic))
1355*4882a593Smuzhiyun 			return 0;
1356*4882a593Smuzhiyun 
1357*4882a593Smuzhiyun 		val = apic_get_tmcct(apic);
1358*4882a593Smuzhiyun 		break;
1359*4882a593Smuzhiyun 	case APIC_PROCPRI:
1360*4882a593Smuzhiyun 		apic_update_ppr(apic);
1361*4882a593Smuzhiyun 		val = kvm_lapic_get_reg(apic, offset);
1362*4882a593Smuzhiyun 		break;
1363*4882a593Smuzhiyun 	case APIC_TASKPRI:
1364*4882a593Smuzhiyun 		report_tpr_access(apic, false);
1365*4882a593Smuzhiyun 		fallthrough;
1366*4882a593Smuzhiyun 	default:
1367*4882a593Smuzhiyun 		val = kvm_lapic_get_reg(apic, offset);
1368*4882a593Smuzhiyun 		break;
1369*4882a593Smuzhiyun 	}
1370*4882a593Smuzhiyun 
1371*4882a593Smuzhiyun 	return val;
1372*4882a593Smuzhiyun }
1373*4882a593Smuzhiyun 
to_lapic(struct kvm_io_device * dev)1374*4882a593Smuzhiyun static inline struct kvm_lapic *to_lapic(struct kvm_io_device *dev)
1375*4882a593Smuzhiyun {
1376*4882a593Smuzhiyun 	return container_of(dev, struct kvm_lapic, dev);
1377*4882a593Smuzhiyun }
1378*4882a593Smuzhiyun 
1379*4882a593Smuzhiyun #define APIC_REG_MASK(reg)	(1ull << ((reg) >> 4))
1380*4882a593Smuzhiyun #define APIC_REGS_MASK(first, count) \
1381*4882a593Smuzhiyun 	(APIC_REG_MASK(first) * ((1ull << (count)) - 1))
1382*4882a593Smuzhiyun 
kvm_lapic_reg_read(struct kvm_lapic * apic,u32 offset,int len,void * data)1383*4882a593Smuzhiyun int kvm_lapic_reg_read(struct kvm_lapic *apic, u32 offset, int len,
1384*4882a593Smuzhiyun 		void *data)
1385*4882a593Smuzhiyun {
1386*4882a593Smuzhiyun 	unsigned char alignment = offset & 0xf;
1387*4882a593Smuzhiyun 	u32 result;
1388*4882a593Smuzhiyun 	/* this bitmask has a bit cleared for each reserved register */
1389*4882a593Smuzhiyun 	u64 valid_reg_mask =
1390*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_ID) |
1391*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_LVR) |
1392*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_TASKPRI) |
1393*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_PROCPRI) |
1394*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_LDR) |
1395*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_DFR) |
1396*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_SPIV) |
1397*4882a593Smuzhiyun 		APIC_REGS_MASK(APIC_ISR, APIC_ISR_NR) |
1398*4882a593Smuzhiyun 		APIC_REGS_MASK(APIC_TMR, APIC_ISR_NR) |
1399*4882a593Smuzhiyun 		APIC_REGS_MASK(APIC_IRR, APIC_ISR_NR) |
1400*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_ESR) |
1401*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_ICR) |
1402*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_ICR2) |
1403*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_LVTT) |
1404*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_LVTTHMR) |
1405*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_LVTPC) |
1406*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_LVT0) |
1407*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_LVT1) |
1408*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_LVTERR) |
1409*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_TMICT) |
1410*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_TMCCT) |
1411*4882a593Smuzhiyun 		APIC_REG_MASK(APIC_TDCR);
1412*4882a593Smuzhiyun 
1413*4882a593Smuzhiyun 	/* ARBPRI is not valid on x2APIC */
1414*4882a593Smuzhiyun 	if (!apic_x2apic_mode(apic))
1415*4882a593Smuzhiyun 		valid_reg_mask |= APIC_REG_MASK(APIC_ARBPRI);
1416*4882a593Smuzhiyun 
1417*4882a593Smuzhiyun 	if (alignment + len > 4)
1418*4882a593Smuzhiyun 		return 1;
1419*4882a593Smuzhiyun 
1420*4882a593Smuzhiyun 	if (offset > 0x3f0 || !(valid_reg_mask & APIC_REG_MASK(offset)))
1421*4882a593Smuzhiyun 		return 1;
1422*4882a593Smuzhiyun 
1423*4882a593Smuzhiyun 	result = __apic_read(apic, offset & ~0xf);
1424*4882a593Smuzhiyun 
1425*4882a593Smuzhiyun 	trace_kvm_apic_read(offset, result);
1426*4882a593Smuzhiyun 
1427*4882a593Smuzhiyun 	switch (len) {
1428*4882a593Smuzhiyun 	case 1:
1429*4882a593Smuzhiyun 	case 2:
1430*4882a593Smuzhiyun 	case 4:
1431*4882a593Smuzhiyun 		memcpy(data, (char *)&result + alignment, len);
1432*4882a593Smuzhiyun 		break;
1433*4882a593Smuzhiyun 	default:
1434*4882a593Smuzhiyun 		printk(KERN_ERR "Local APIC read with len = %x, "
1435*4882a593Smuzhiyun 		       "should be 1,2, or 4 instead\n", len);
1436*4882a593Smuzhiyun 		break;
1437*4882a593Smuzhiyun 	}
1438*4882a593Smuzhiyun 	return 0;
1439*4882a593Smuzhiyun }
1440*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_lapic_reg_read);
1441*4882a593Smuzhiyun 
apic_mmio_in_range(struct kvm_lapic * apic,gpa_t addr)1442*4882a593Smuzhiyun static int apic_mmio_in_range(struct kvm_lapic *apic, gpa_t addr)
1443*4882a593Smuzhiyun {
1444*4882a593Smuzhiyun 	return addr >= apic->base_address &&
1445*4882a593Smuzhiyun 		addr < apic->base_address + LAPIC_MMIO_LENGTH;
1446*4882a593Smuzhiyun }
1447*4882a593Smuzhiyun 
apic_mmio_read(struct kvm_vcpu * vcpu,struct kvm_io_device * this,gpa_t address,int len,void * data)1448*4882a593Smuzhiyun static int apic_mmio_read(struct kvm_vcpu *vcpu, struct kvm_io_device *this,
1449*4882a593Smuzhiyun 			   gpa_t address, int len, void *data)
1450*4882a593Smuzhiyun {
1451*4882a593Smuzhiyun 	struct kvm_lapic *apic = to_lapic(this);
1452*4882a593Smuzhiyun 	u32 offset = address - apic->base_address;
1453*4882a593Smuzhiyun 
1454*4882a593Smuzhiyun 	if (!apic_mmio_in_range(apic, address))
1455*4882a593Smuzhiyun 		return -EOPNOTSUPP;
1456*4882a593Smuzhiyun 
1457*4882a593Smuzhiyun 	if (!kvm_apic_hw_enabled(apic) || apic_x2apic_mode(apic)) {
1458*4882a593Smuzhiyun 		if (!kvm_check_has_quirk(vcpu->kvm,
1459*4882a593Smuzhiyun 					 KVM_X86_QUIRK_LAPIC_MMIO_HOLE))
1460*4882a593Smuzhiyun 			return -EOPNOTSUPP;
1461*4882a593Smuzhiyun 
1462*4882a593Smuzhiyun 		memset(data, 0xff, len);
1463*4882a593Smuzhiyun 		return 0;
1464*4882a593Smuzhiyun 	}
1465*4882a593Smuzhiyun 
1466*4882a593Smuzhiyun 	kvm_lapic_reg_read(apic, offset, len, data);
1467*4882a593Smuzhiyun 
1468*4882a593Smuzhiyun 	return 0;
1469*4882a593Smuzhiyun }
1470*4882a593Smuzhiyun 
update_divide_count(struct kvm_lapic * apic)1471*4882a593Smuzhiyun static void update_divide_count(struct kvm_lapic *apic)
1472*4882a593Smuzhiyun {
1473*4882a593Smuzhiyun 	u32 tmp1, tmp2, tdcr;
1474*4882a593Smuzhiyun 
1475*4882a593Smuzhiyun 	tdcr = kvm_lapic_get_reg(apic, APIC_TDCR);
1476*4882a593Smuzhiyun 	tmp1 = tdcr & 0xf;
1477*4882a593Smuzhiyun 	tmp2 = ((tmp1 & 0x3) | ((tmp1 & 0x8) >> 1)) + 1;
1478*4882a593Smuzhiyun 	apic->divide_count = 0x1 << (tmp2 & 0x7);
1479*4882a593Smuzhiyun }
1480*4882a593Smuzhiyun 
limit_periodic_timer_frequency(struct kvm_lapic * apic)1481*4882a593Smuzhiyun static void limit_periodic_timer_frequency(struct kvm_lapic *apic)
1482*4882a593Smuzhiyun {
1483*4882a593Smuzhiyun 	/*
1484*4882a593Smuzhiyun 	 * Do not allow the guest to program periodic timers with small
1485*4882a593Smuzhiyun 	 * interval, since the hrtimers are not throttled by the host
1486*4882a593Smuzhiyun 	 * scheduler.
1487*4882a593Smuzhiyun 	 */
1488*4882a593Smuzhiyun 	if (apic_lvtt_period(apic) && apic->lapic_timer.period) {
1489*4882a593Smuzhiyun 		s64 min_period = min_timer_period_us * 1000LL;
1490*4882a593Smuzhiyun 
1491*4882a593Smuzhiyun 		if (apic->lapic_timer.period < min_period) {
1492*4882a593Smuzhiyun 			pr_info_ratelimited(
1493*4882a593Smuzhiyun 			    "kvm: vcpu %i: requested %lld ns "
1494*4882a593Smuzhiyun 			    "lapic timer period limited to %lld ns\n",
1495*4882a593Smuzhiyun 			    apic->vcpu->vcpu_id,
1496*4882a593Smuzhiyun 			    apic->lapic_timer.period, min_period);
1497*4882a593Smuzhiyun 			apic->lapic_timer.period = min_period;
1498*4882a593Smuzhiyun 		}
1499*4882a593Smuzhiyun 	}
1500*4882a593Smuzhiyun }
1501*4882a593Smuzhiyun 
1502*4882a593Smuzhiyun static void cancel_hv_timer(struct kvm_lapic *apic);
1503*4882a593Smuzhiyun 
apic_update_lvtt(struct kvm_lapic * apic)1504*4882a593Smuzhiyun static void apic_update_lvtt(struct kvm_lapic *apic)
1505*4882a593Smuzhiyun {
1506*4882a593Smuzhiyun 	u32 timer_mode = kvm_lapic_get_reg(apic, APIC_LVTT) &
1507*4882a593Smuzhiyun 			apic->lapic_timer.timer_mode_mask;
1508*4882a593Smuzhiyun 
1509*4882a593Smuzhiyun 	if (apic->lapic_timer.timer_mode != timer_mode) {
1510*4882a593Smuzhiyun 		if (apic_lvtt_tscdeadline(apic) != (timer_mode ==
1511*4882a593Smuzhiyun 				APIC_LVT_TIMER_TSCDEADLINE)) {
1512*4882a593Smuzhiyun 			hrtimer_cancel(&apic->lapic_timer.timer);
1513*4882a593Smuzhiyun 			preempt_disable();
1514*4882a593Smuzhiyun 			if (apic->lapic_timer.hv_timer_in_use)
1515*4882a593Smuzhiyun 				cancel_hv_timer(apic);
1516*4882a593Smuzhiyun 			preempt_enable();
1517*4882a593Smuzhiyun 			kvm_lapic_set_reg(apic, APIC_TMICT, 0);
1518*4882a593Smuzhiyun 			apic->lapic_timer.period = 0;
1519*4882a593Smuzhiyun 			apic->lapic_timer.tscdeadline = 0;
1520*4882a593Smuzhiyun 		}
1521*4882a593Smuzhiyun 		apic->lapic_timer.timer_mode = timer_mode;
1522*4882a593Smuzhiyun 		limit_periodic_timer_frequency(apic);
1523*4882a593Smuzhiyun 	}
1524*4882a593Smuzhiyun }
1525*4882a593Smuzhiyun 
1526*4882a593Smuzhiyun /*
1527*4882a593Smuzhiyun  * On APICv, this test will cause a busy wait
1528*4882a593Smuzhiyun  * during a higher-priority task.
1529*4882a593Smuzhiyun  */
1530*4882a593Smuzhiyun 
lapic_timer_int_injected(struct kvm_vcpu * vcpu)1531*4882a593Smuzhiyun static bool lapic_timer_int_injected(struct kvm_vcpu *vcpu)
1532*4882a593Smuzhiyun {
1533*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
1534*4882a593Smuzhiyun 	u32 reg = kvm_lapic_get_reg(apic, APIC_LVTT);
1535*4882a593Smuzhiyun 
1536*4882a593Smuzhiyun 	if (kvm_apic_hw_enabled(apic)) {
1537*4882a593Smuzhiyun 		int vec = reg & APIC_VECTOR_MASK;
1538*4882a593Smuzhiyun 		void *bitmap = apic->regs + APIC_ISR;
1539*4882a593Smuzhiyun 
1540*4882a593Smuzhiyun 		if (vcpu->arch.apicv_active)
1541*4882a593Smuzhiyun 			bitmap = apic->regs + APIC_IRR;
1542*4882a593Smuzhiyun 
1543*4882a593Smuzhiyun 		if (apic_test_vector(vec, bitmap))
1544*4882a593Smuzhiyun 			return true;
1545*4882a593Smuzhiyun 	}
1546*4882a593Smuzhiyun 	return false;
1547*4882a593Smuzhiyun }
1548*4882a593Smuzhiyun 
__wait_lapic_expire(struct kvm_vcpu * vcpu,u64 guest_cycles)1549*4882a593Smuzhiyun static inline void __wait_lapic_expire(struct kvm_vcpu *vcpu, u64 guest_cycles)
1550*4882a593Smuzhiyun {
1551*4882a593Smuzhiyun 	u64 timer_advance_ns = vcpu->arch.apic->lapic_timer.timer_advance_ns;
1552*4882a593Smuzhiyun 
1553*4882a593Smuzhiyun 	/*
1554*4882a593Smuzhiyun 	 * If the guest TSC is running at a different ratio than the host, then
1555*4882a593Smuzhiyun 	 * convert the delay to nanoseconds to achieve an accurate delay.  Note
1556*4882a593Smuzhiyun 	 * that __delay() uses delay_tsc whenever the hardware has TSC, thus
1557*4882a593Smuzhiyun 	 * always for VMX enabled hardware.
1558*4882a593Smuzhiyun 	 */
1559*4882a593Smuzhiyun 	if (vcpu->arch.tsc_scaling_ratio == kvm_default_tsc_scaling_ratio) {
1560*4882a593Smuzhiyun 		__delay(min(guest_cycles,
1561*4882a593Smuzhiyun 			nsec_to_cycles(vcpu, timer_advance_ns)));
1562*4882a593Smuzhiyun 	} else {
1563*4882a593Smuzhiyun 		u64 delay_ns = guest_cycles * 1000000ULL;
1564*4882a593Smuzhiyun 		do_div(delay_ns, vcpu->arch.virtual_tsc_khz);
1565*4882a593Smuzhiyun 		ndelay(min_t(u32, delay_ns, timer_advance_ns));
1566*4882a593Smuzhiyun 	}
1567*4882a593Smuzhiyun }
1568*4882a593Smuzhiyun 
adjust_lapic_timer_advance(struct kvm_vcpu * vcpu,s64 advance_expire_delta)1569*4882a593Smuzhiyun static inline void adjust_lapic_timer_advance(struct kvm_vcpu *vcpu,
1570*4882a593Smuzhiyun 					      s64 advance_expire_delta)
1571*4882a593Smuzhiyun {
1572*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
1573*4882a593Smuzhiyun 	u32 timer_advance_ns = apic->lapic_timer.timer_advance_ns;
1574*4882a593Smuzhiyun 	u64 ns;
1575*4882a593Smuzhiyun 
1576*4882a593Smuzhiyun 	/* Do not adjust for tiny fluctuations or large random spikes. */
1577*4882a593Smuzhiyun 	if (abs(advance_expire_delta) > LAPIC_TIMER_ADVANCE_ADJUST_MAX ||
1578*4882a593Smuzhiyun 	    abs(advance_expire_delta) < LAPIC_TIMER_ADVANCE_ADJUST_MIN)
1579*4882a593Smuzhiyun 		return;
1580*4882a593Smuzhiyun 
1581*4882a593Smuzhiyun 	/* too early */
1582*4882a593Smuzhiyun 	if (advance_expire_delta < 0) {
1583*4882a593Smuzhiyun 		ns = -advance_expire_delta * 1000000ULL;
1584*4882a593Smuzhiyun 		do_div(ns, vcpu->arch.virtual_tsc_khz);
1585*4882a593Smuzhiyun 		timer_advance_ns -= ns/LAPIC_TIMER_ADVANCE_ADJUST_STEP;
1586*4882a593Smuzhiyun 	} else {
1587*4882a593Smuzhiyun 	/* too late */
1588*4882a593Smuzhiyun 		ns = advance_expire_delta * 1000000ULL;
1589*4882a593Smuzhiyun 		do_div(ns, vcpu->arch.virtual_tsc_khz);
1590*4882a593Smuzhiyun 		timer_advance_ns += ns/LAPIC_TIMER_ADVANCE_ADJUST_STEP;
1591*4882a593Smuzhiyun 	}
1592*4882a593Smuzhiyun 
1593*4882a593Smuzhiyun 	if (unlikely(timer_advance_ns > LAPIC_TIMER_ADVANCE_NS_MAX))
1594*4882a593Smuzhiyun 		timer_advance_ns = LAPIC_TIMER_ADVANCE_NS_INIT;
1595*4882a593Smuzhiyun 	apic->lapic_timer.timer_advance_ns = timer_advance_ns;
1596*4882a593Smuzhiyun }
1597*4882a593Smuzhiyun 
__kvm_wait_lapic_expire(struct kvm_vcpu * vcpu)1598*4882a593Smuzhiyun static void __kvm_wait_lapic_expire(struct kvm_vcpu *vcpu)
1599*4882a593Smuzhiyun {
1600*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
1601*4882a593Smuzhiyun 	u64 guest_tsc, tsc_deadline;
1602*4882a593Smuzhiyun 
1603*4882a593Smuzhiyun 	tsc_deadline = apic->lapic_timer.expired_tscdeadline;
1604*4882a593Smuzhiyun 	apic->lapic_timer.expired_tscdeadline = 0;
1605*4882a593Smuzhiyun 	guest_tsc = kvm_read_l1_tsc(vcpu, rdtsc());
1606*4882a593Smuzhiyun 	apic->lapic_timer.advance_expire_delta = guest_tsc - tsc_deadline;
1607*4882a593Smuzhiyun 
1608*4882a593Smuzhiyun 	if (guest_tsc < tsc_deadline)
1609*4882a593Smuzhiyun 		__wait_lapic_expire(vcpu, tsc_deadline - guest_tsc);
1610*4882a593Smuzhiyun 
1611*4882a593Smuzhiyun 	if (lapic_timer_advance_dynamic)
1612*4882a593Smuzhiyun 		adjust_lapic_timer_advance(vcpu, apic->lapic_timer.advance_expire_delta);
1613*4882a593Smuzhiyun }
1614*4882a593Smuzhiyun 
kvm_wait_lapic_expire(struct kvm_vcpu * vcpu)1615*4882a593Smuzhiyun void kvm_wait_lapic_expire(struct kvm_vcpu *vcpu)
1616*4882a593Smuzhiyun {
1617*4882a593Smuzhiyun 	if (lapic_in_kernel(vcpu) &&
1618*4882a593Smuzhiyun 	    vcpu->arch.apic->lapic_timer.expired_tscdeadline &&
1619*4882a593Smuzhiyun 	    vcpu->arch.apic->lapic_timer.timer_advance_ns &&
1620*4882a593Smuzhiyun 	    lapic_timer_int_injected(vcpu))
1621*4882a593Smuzhiyun 		__kvm_wait_lapic_expire(vcpu);
1622*4882a593Smuzhiyun }
1623*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_wait_lapic_expire);
1624*4882a593Smuzhiyun 
kvm_apic_inject_pending_timer_irqs(struct kvm_lapic * apic)1625*4882a593Smuzhiyun static void kvm_apic_inject_pending_timer_irqs(struct kvm_lapic *apic)
1626*4882a593Smuzhiyun {
1627*4882a593Smuzhiyun 	struct kvm_timer *ktimer = &apic->lapic_timer;
1628*4882a593Smuzhiyun 
1629*4882a593Smuzhiyun 	kvm_apic_local_deliver(apic, APIC_LVTT);
1630*4882a593Smuzhiyun 	if (apic_lvtt_tscdeadline(apic)) {
1631*4882a593Smuzhiyun 		ktimer->tscdeadline = 0;
1632*4882a593Smuzhiyun 	} else if (apic_lvtt_oneshot(apic)) {
1633*4882a593Smuzhiyun 		ktimer->tscdeadline = 0;
1634*4882a593Smuzhiyun 		ktimer->target_expiration = 0;
1635*4882a593Smuzhiyun 	}
1636*4882a593Smuzhiyun }
1637*4882a593Smuzhiyun 
apic_timer_expired(struct kvm_lapic * apic,bool from_timer_fn)1638*4882a593Smuzhiyun static void apic_timer_expired(struct kvm_lapic *apic, bool from_timer_fn)
1639*4882a593Smuzhiyun {
1640*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu = apic->vcpu;
1641*4882a593Smuzhiyun 	struct kvm_timer *ktimer = &apic->lapic_timer;
1642*4882a593Smuzhiyun 
1643*4882a593Smuzhiyun 	if (atomic_read(&apic->lapic_timer.pending))
1644*4882a593Smuzhiyun 		return;
1645*4882a593Smuzhiyun 
1646*4882a593Smuzhiyun 	if (apic_lvtt_tscdeadline(apic) || ktimer->hv_timer_in_use)
1647*4882a593Smuzhiyun 		ktimer->expired_tscdeadline = ktimer->tscdeadline;
1648*4882a593Smuzhiyun 
1649*4882a593Smuzhiyun 	if (!from_timer_fn && vcpu->arch.apicv_active) {
1650*4882a593Smuzhiyun 		WARN_ON(kvm_get_running_vcpu() != vcpu);
1651*4882a593Smuzhiyun 		kvm_apic_inject_pending_timer_irqs(apic);
1652*4882a593Smuzhiyun 		return;
1653*4882a593Smuzhiyun 	}
1654*4882a593Smuzhiyun 
1655*4882a593Smuzhiyun 	if (kvm_use_posted_timer_interrupt(apic->vcpu)) {
1656*4882a593Smuzhiyun 		/*
1657*4882a593Smuzhiyun 		 * Ensure the guest's timer has truly expired before posting an
1658*4882a593Smuzhiyun 		 * interrupt.  Open code the relevant checks to avoid querying
1659*4882a593Smuzhiyun 		 * lapic_timer_int_injected(), which will be false since the
1660*4882a593Smuzhiyun 		 * interrupt isn't yet injected.  Waiting until after injecting
1661*4882a593Smuzhiyun 		 * is not an option since that won't help a posted interrupt.
1662*4882a593Smuzhiyun 		 */
1663*4882a593Smuzhiyun 		if (vcpu->arch.apic->lapic_timer.expired_tscdeadline &&
1664*4882a593Smuzhiyun 		    vcpu->arch.apic->lapic_timer.timer_advance_ns)
1665*4882a593Smuzhiyun 			__kvm_wait_lapic_expire(vcpu);
1666*4882a593Smuzhiyun 		kvm_apic_inject_pending_timer_irqs(apic);
1667*4882a593Smuzhiyun 		return;
1668*4882a593Smuzhiyun 	}
1669*4882a593Smuzhiyun 
1670*4882a593Smuzhiyun 	atomic_inc(&apic->lapic_timer.pending);
1671*4882a593Smuzhiyun 	kvm_make_request(KVM_REQ_PENDING_TIMER, vcpu);
1672*4882a593Smuzhiyun 	if (from_timer_fn)
1673*4882a593Smuzhiyun 		kvm_vcpu_kick(vcpu);
1674*4882a593Smuzhiyun }
1675*4882a593Smuzhiyun 
start_sw_tscdeadline(struct kvm_lapic * apic)1676*4882a593Smuzhiyun static void start_sw_tscdeadline(struct kvm_lapic *apic)
1677*4882a593Smuzhiyun {
1678*4882a593Smuzhiyun 	struct kvm_timer *ktimer = &apic->lapic_timer;
1679*4882a593Smuzhiyun 	u64 guest_tsc, tscdeadline = ktimer->tscdeadline;
1680*4882a593Smuzhiyun 	u64 ns = 0;
1681*4882a593Smuzhiyun 	ktime_t expire;
1682*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu = apic->vcpu;
1683*4882a593Smuzhiyun 	unsigned long this_tsc_khz = vcpu->arch.virtual_tsc_khz;
1684*4882a593Smuzhiyun 	unsigned long flags;
1685*4882a593Smuzhiyun 	ktime_t now;
1686*4882a593Smuzhiyun 
1687*4882a593Smuzhiyun 	if (unlikely(!tscdeadline || !this_tsc_khz))
1688*4882a593Smuzhiyun 		return;
1689*4882a593Smuzhiyun 
1690*4882a593Smuzhiyun 	local_irq_save(flags);
1691*4882a593Smuzhiyun 
1692*4882a593Smuzhiyun 	now = ktime_get();
1693*4882a593Smuzhiyun 	guest_tsc = kvm_read_l1_tsc(vcpu, rdtsc());
1694*4882a593Smuzhiyun 
1695*4882a593Smuzhiyun 	ns = (tscdeadline - guest_tsc) * 1000000ULL;
1696*4882a593Smuzhiyun 	do_div(ns, this_tsc_khz);
1697*4882a593Smuzhiyun 
1698*4882a593Smuzhiyun 	if (likely(tscdeadline > guest_tsc) &&
1699*4882a593Smuzhiyun 	    likely(ns > apic->lapic_timer.timer_advance_ns)) {
1700*4882a593Smuzhiyun 		expire = ktime_add_ns(now, ns);
1701*4882a593Smuzhiyun 		expire = ktime_sub_ns(expire, ktimer->timer_advance_ns);
1702*4882a593Smuzhiyun 		hrtimer_start(&ktimer->timer, expire, HRTIMER_MODE_ABS_HARD);
1703*4882a593Smuzhiyun 	} else
1704*4882a593Smuzhiyun 		apic_timer_expired(apic, false);
1705*4882a593Smuzhiyun 
1706*4882a593Smuzhiyun 	local_irq_restore(flags);
1707*4882a593Smuzhiyun }
1708*4882a593Smuzhiyun 
tmict_to_ns(struct kvm_lapic * apic,u32 tmict)1709*4882a593Smuzhiyun static inline u64 tmict_to_ns(struct kvm_lapic *apic, u32 tmict)
1710*4882a593Smuzhiyun {
1711*4882a593Smuzhiyun 	return (u64)tmict * APIC_BUS_CYCLE_NS * (u64)apic->divide_count;
1712*4882a593Smuzhiyun }
1713*4882a593Smuzhiyun 
update_target_expiration(struct kvm_lapic * apic,uint32_t old_divisor)1714*4882a593Smuzhiyun static void update_target_expiration(struct kvm_lapic *apic, uint32_t old_divisor)
1715*4882a593Smuzhiyun {
1716*4882a593Smuzhiyun 	ktime_t now, remaining;
1717*4882a593Smuzhiyun 	u64 ns_remaining_old, ns_remaining_new;
1718*4882a593Smuzhiyun 
1719*4882a593Smuzhiyun 	apic->lapic_timer.period =
1720*4882a593Smuzhiyun 			tmict_to_ns(apic, kvm_lapic_get_reg(apic, APIC_TMICT));
1721*4882a593Smuzhiyun 	limit_periodic_timer_frequency(apic);
1722*4882a593Smuzhiyun 
1723*4882a593Smuzhiyun 	now = ktime_get();
1724*4882a593Smuzhiyun 	remaining = ktime_sub(apic->lapic_timer.target_expiration, now);
1725*4882a593Smuzhiyun 	if (ktime_to_ns(remaining) < 0)
1726*4882a593Smuzhiyun 		remaining = 0;
1727*4882a593Smuzhiyun 
1728*4882a593Smuzhiyun 	ns_remaining_old = ktime_to_ns(remaining);
1729*4882a593Smuzhiyun 	ns_remaining_new = mul_u64_u32_div(ns_remaining_old,
1730*4882a593Smuzhiyun 	                                   apic->divide_count, old_divisor);
1731*4882a593Smuzhiyun 
1732*4882a593Smuzhiyun 	apic->lapic_timer.tscdeadline +=
1733*4882a593Smuzhiyun 		nsec_to_cycles(apic->vcpu, ns_remaining_new) -
1734*4882a593Smuzhiyun 		nsec_to_cycles(apic->vcpu, ns_remaining_old);
1735*4882a593Smuzhiyun 	apic->lapic_timer.target_expiration = ktime_add_ns(now, ns_remaining_new);
1736*4882a593Smuzhiyun }
1737*4882a593Smuzhiyun 
set_target_expiration(struct kvm_lapic * apic,u32 count_reg)1738*4882a593Smuzhiyun static bool set_target_expiration(struct kvm_lapic *apic, u32 count_reg)
1739*4882a593Smuzhiyun {
1740*4882a593Smuzhiyun 	ktime_t now;
1741*4882a593Smuzhiyun 	u64 tscl = rdtsc();
1742*4882a593Smuzhiyun 	s64 deadline;
1743*4882a593Smuzhiyun 
1744*4882a593Smuzhiyun 	now = ktime_get();
1745*4882a593Smuzhiyun 	apic->lapic_timer.period =
1746*4882a593Smuzhiyun 			tmict_to_ns(apic, kvm_lapic_get_reg(apic, APIC_TMICT));
1747*4882a593Smuzhiyun 
1748*4882a593Smuzhiyun 	if (!apic->lapic_timer.period) {
1749*4882a593Smuzhiyun 		apic->lapic_timer.tscdeadline = 0;
1750*4882a593Smuzhiyun 		return false;
1751*4882a593Smuzhiyun 	}
1752*4882a593Smuzhiyun 
1753*4882a593Smuzhiyun 	limit_periodic_timer_frequency(apic);
1754*4882a593Smuzhiyun 	deadline = apic->lapic_timer.period;
1755*4882a593Smuzhiyun 
1756*4882a593Smuzhiyun 	if (apic_lvtt_period(apic) || apic_lvtt_oneshot(apic)) {
1757*4882a593Smuzhiyun 		if (unlikely(count_reg != APIC_TMICT)) {
1758*4882a593Smuzhiyun 			deadline = tmict_to_ns(apic,
1759*4882a593Smuzhiyun 				     kvm_lapic_get_reg(apic, count_reg));
1760*4882a593Smuzhiyun 			if (unlikely(deadline <= 0))
1761*4882a593Smuzhiyun 				deadline = apic->lapic_timer.period;
1762*4882a593Smuzhiyun 			else if (unlikely(deadline > apic->lapic_timer.period)) {
1763*4882a593Smuzhiyun 				pr_info_ratelimited(
1764*4882a593Smuzhiyun 				    "kvm: vcpu %i: requested lapic timer restore with "
1765*4882a593Smuzhiyun 				    "starting count register %#x=%u (%lld ns) > initial count (%lld ns). "
1766*4882a593Smuzhiyun 				    "Using initial count to start timer.\n",
1767*4882a593Smuzhiyun 				    apic->vcpu->vcpu_id,
1768*4882a593Smuzhiyun 				    count_reg,
1769*4882a593Smuzhiyun 				    kvm_lapic_get_reg(apic, count_reg),
1770*4882a593Smuzhiyun 				    deadline, apic->lapic_timer.period);
1771*4882a593Smuzhiyun 				kvm_lapic_set_reg(apic, count_reg, 0);
1772*4882a593Smuzhiyun 				deadline = apic->lapic_timer.period;
1773*4882a593Smuzhiyun 			}
1774*4882a593Smuzhiyun 		}
1775*4882a593Smuzhiyun 	}
1776*4882a593Smuzhiyun 
1777*4882a593Smuzhiyun 	apic->lapic_timer.tscdeadline = kvm_read_l1_tsc(apic->vcpu, tscl) +
1778*4882a593Smuzhiyun 		nsec_to_cycles(apic->vcpu, deadline);
1779*4882a593Smuzhiyun 	apic->lapic_timer.target_expiration = ktime_add_ns(now, deadline);
1780*4882a593Smuzhiyun 
1781*4882a593Smuzhiyun 	return true;
1782*4882a593Smuzhiyun }
1783*4882a593Smuzhiyun 
advance_periodic_target_expiration(struct kvm_lapic * apic)1784*4882a593Smuzhiyun static void advance_periodic_target_expiration(struct kvm_lapic *apic)
1785*4882a593Smuzhiyun {
1786*4882a593Smuzhiyun 	ktime_t now = ktime_get();
1787*4882a593Smuzhiyun 	u64 tscl = rdtsc();
1788*4882a593Smuzhiyun 	ktime_t delta;
1789*4882a593Smuzhiyun 
1790*4882a593Smuzhiyun 	/*
1791*4882a593Smuzhiyun 	 * Synchronize both deadlines to the same time source or
1792*4882a593Smuzhiyun 	 * differences in the periods (caused by differences in the
1793*4882a593Smuzhiyun 	 * underlying clocks or numerical approximation errors) will
1794*4882a593Smuzhiyun 	 * cause the two to drift apart over time as the errors
1795*4882a593Smuzhiyun 	 * accumulate.
1796*4882a593Smuzhiyun 	 */
1797*4882a593Smuzhiyun 	apic->lapic_timer.target_expiration =
1798*4882a593Smuzhiyun 		ktime_add_ns(apic->lapic_timer.target_expiration,
1799*4882a593Smuzhiyun 				apic->lapic_timer.period);
1800*4882a593Smuzhiyun 	delta = ktime_sub(apic->lapic_timer.target_expiration, now);
1801*4882a593Smuzhiyun 	apic->lapic_timer.tscdeadline = kvm_read_l1_tsc(apic->vcpu, tscl) +
1802*4882a593Smuzhiyun 		nsec_to_cycles(apic->vcpu, delta);
1803*4882a593Smuzhiyun }
1804*4882a593Smuzhiyun 
start_sw_period(struct kvm_lapic * apic)1805*4882a593Smuzhiyun static void start_sw_period(struct kvm_lapic *apic)
1806*4882a593Smuzhiyun {
1807*4882a593Smuzhiyun 	if (!apic->lapic_timer.period)
1808*4882a593Smuzhiyun 		return;
1809*4882a593Smuzhiyun 
1810*4882a593Smuzhiyun 	if (ktime_after(ktime_get(),
1811*4882a593Smuzhiyun 			apic->lapic_timer.target_expiration)) {
1812*4882a593Smuzhiyun 		apic_timer_expired(apic, false);
1813*4882a593Smuzhiyun 
1814*4882a593Smuzhiyun 		if (apic_lvtt_oneshot(apic))
1815*4882a593Smuzhiyun 			return;
1816*4882a593Smuzhiyun 
1817*4882a593Smuzhiyun 		advance_periodic_target_expiration(apic);
1818*4882a593Smuzhiyun 	}
1819*4882a593Smuzhiyun 
1820*4882a593Smuzhiyun 	hrtimer_start(&apic->lapic_timer.timer,
1821*4882a593Smuzhiyun 		apic->lapic_timer.target_expiration,
1822*4882a593Smuzhiyun 		HRTIMER_MODE_ABS_HARD);
1823*4882a593Smuzhiyun }
1824*4882a593Smuzhiyun 
kvm_lapic_hv_timer_in_use(struct kvm_vcpu * vcpu)1825*4882a593Smuzhiyun bool kvm_lapic_hv_timer_in_use(struct kvm_vcpu *vcpu)
1826*4882a593Smuzhiyun {
1827*4882a593Smuzhiyun 	if (!lapic_in_kernel(vcpu))
1828*4882a593Smuzhiyun 		return false;
1829*4882a593Smuzhiyun 
1830*4882a593Smuzhiyun 	return vcpu->arch.apic->lapic_timer.hv_timer_in_use;
1831*4882a593Smuzhiyun }
1832*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_lapic_hv_timer_in_use);
1833*4882a593Smuzhiyun 
cancel_hv_timer(struct kvm_lapic * apic)1834*4882a593Smuzhiyun static void cancel_hv_timer(struct kvm_lapic *apic)
1835*4882a593Smuzhiyun {
1836*4882a593Smuzhiyun 	WARN_ON(preemptible());
1837*4882a593Smuzhiyun 	WARN_ON(!apic->lapic_timer.hv_timer_in_use);
1838*4882a593Smuzhiyun 	kvm_x86_ops.cancel_hv_timer(apic->vcpu);
1839*4882a593Smuzhiyun 	apic->lapic_timer.hv_timer_in_use = false;
1840*4882a593Smuzhiyun }
1841*4882a593Smuzhiyun 
start_hv_timer(struct kvm_lapic * apic)1842*4882a593Smuzhiyun static bool start_hv_timer(struct kvm_lapic *apic)
1843*4882a593Smuzhiyun {
1844*4882a593Smuzhiyun 	struct kvm_timer *ktimer = &apic->lapic_timer;
1845*4882a593Smuzhiyun 	struct kvm_vcpu *vcpu = apic->vcpu;
1846*4882a593Smuzhiyun 	bool expired;
1847*4882a593Smuzhiyun 
1848*4882a593Smuzhiyun 	WARN_ON(preemptible());
1849*4882a593Smuzhiyun 	if (!kvm_can_use_hv_timer(vcpu))
1850*4882a593Smuzhiyun 		return false;
1851*4882a593Smuzhiyun 
1852*4882a593Smuzhiyun 	if (!ktimer->tscdeadline)
1853*4882a593Smuzhiyun 		return false;
1854*4882a593Smuzhiyun 
1855*4882a593Smuzhiyun 	if (kvm_x86_ops.set_hv_timer(vcpu, ktimer->tscdeadline, &expired))
1856*4882a593Smuzhiyun 		return false;
1857*4882a593Smuzhiyun 
1858*4882a593Smuzhiyun 	ktimer->hv_timer_in_use = true;
1859*4882a593Smuzhiyun 	hrtimer_cancel(&ktimer->timer);
1860*4882a593Smuzhiyun 
1861*4882a593Smuzhiyun 	/*
1862*4882a593Smuzhiyun 	 * To simplify handling the periodic timer, leave the hv timer running
1863*4882a593Smuzhiyun 	 * even if the deadline timer has expired, i.e. rely on the resulting
1864*4882a593Smuzhiyun 	 * VM-Exit to recompute the periodic timer's target expiration.
1865*4882a593Smuzhiyun 	 */
1866*4882a593Smuzhiyun 	if (!apic_lvtt_period(apic)) {
1867*4882a593Smuzhiyun 		/*
1868*4882a593Smuzhiyun 		 * Cancel the hv timer if the sw timer fired while the hv timer
1869*4882a593Smuzhiyun 		 * was being programmed, or if the hv timer itself expired.
1870*4882a593Smuzhiyun 		 */
1871*4882a593Smuzhiyun 		if (atomic_read(&ktimer->pending)) {
1872*4882a593Smuzhiyun 			cancel_hv_timer(apic);
1873*4882a593Smuzhiyun 		} else if (expired) {
1874*4882a593Smuzhiyun 			apic_timer_expired(apic, false);
1875*4882a593Smuzhiyun 			cancel_hv_timer(apic);
1876*4882a593Smuzhiyun 		}
1877*4882a593Smuzhiyun 	}
1878*4882a593Smuzhiyun 
1879*4882a593Smuzhiyun 	trace_kvm_hv_timer_state(vcpu->vcpu_id, ktimer->hv_timer_in_use);
1880*4882a593Smuzhiyun 
1881*4882a593Smuzhiyun 	return true;
1882*4882a593Smuzhiyun }
1883*4882a593Smuzhiyun 
start_sw_timer(struct kvm_lapic * apic)1884*4882a593Smuzhiyun static void start_sw_timer(struct kvm_lapic *apic)
1885*4882a593Smuzhiyun {
1886*4882a593Smuzhiyun 	struct kvm_timer *ktimer = &apic->lapic_timer;
1887*4882a593Smuzhiyun 
1888*4882a593Smuzhiyun 	WARN_ON(preemptible());
1889*4882a593Smuzhiyun 	if (apic->lapic_timer.hv_timer_in_use)
1890*4882a593Smuzhiyun 		cancel_hv_timer(apic);
1891*4882a593Smuzhiyun 	if (!apic_lvtt_period(apic) && atomic_read(&ktimer->pending))
1892*4882a593Smuzhiyun 		return;
1893*4882a593Smuzhiyun 
1894*4882a593Smuzhiyun 	if (apic_lvtt_period(apic) || apic_lvtt_oneshot(apic))
1895*4882a593Smuzhiyun 		start_sw_period(apic);
1896*4882a593Smuzhiyun 	else if (apic_lvtt_tscdeadline(apic))
1897*4882a593Smuzhiyun 		start_sw_tscdeadline(apic);
1898*4882a593Smuzhiyun 	trace_kvm_hv_timer_state(apic->vcpu->vcpu_id, false);
1899*4882a593Smuzhiyun }
1900*4882a593Smuzhiyun 
restart_apic_timer(struct kvm_lapic * apic)1901*4882a593Smuzhiyun static void restart_apic_timer(struct kvm_lapic *apic)
1902*4882a593Smuzhiyun {
1903*4882a593Smuzhiyun 	preempt_disable();
1904*4882a593Smuzhiyun 
1905*4882a593Smuzhiyun 	if (!apic_lvtt_period(apic) && atomic_read(&apic->lapic_timer.pending))
1906*4882a593Smuzhiyun 		goto out;
1907*4882a593Smuzhiyun 
1908*4882a593Smuzhiyun 	if (!start_hv_timer(apic))
1909*4882a593Smuzhiyun 		start_sw_timer(apic);
1910*4882a593Smuzhiyun out:
1911*4882a593Smuzhiyun 	preempt_enable();
1912*4882a593Smuzhiyun }
1913*4882a593Smuzhiyun 
kvm_lapic_expired_hv_timer(struct kvm_vcpu * vcpu)1914*4882a593Smuzhiyun void kvm_lapic_expired_hv_timer(struct kvm_vcpu *vcpu)
1915*4882a593Smuzhiyun {
1916*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
1917*4882a593Smuzhiyun 
1918*4882a593Smuzhiyun 	preempt_disable();
1919*4882a593Smuzhiyun 	/* If the preempt notifier has already run, it also called apic_timer_expired */
1920*4882a593Smuzhiyun 	if (!apic->lapic_timer.hv_timer_in_use)
1921*4882a593Smuzhiyun 		goto out;
1922*4882a593Smuzhiyun 	WARN_ON(rcuwait_active(&vcpu->wait));
1923*4882a593Smuzhiyun 	apic_timer_expired(apic, false);
1924*4882a593Smuzhiyun 	cancel_hv_timer(apic);
1925*4882a593Smuzhiyun 
1926*4882a593Smuzhiyun 	if (apic_lvtt_period(apic) && apic->lapic_timer.period) {
1927*4882a593Smuzhiyun 		advance_periodic_target_expiration(apic);
1928*4882a593Smuzhiyun 		restart_apic_timer(apic);
1929*4882a593Smuzhiyun 	}
1930*4882a593Smuzhiyun out:
1931*4882a593Smuzhiyun 	preempt_enable();
1932*4882a593Smuzhiyun }
1933*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_lapic_expired_hv_timer);
1934*4882a593Smuzhiyun 
kvm_lapic_switch_to_hv_timer(struct kvm_vcpu * vcpu)1935*4882a593Smuzhiyun void kvm_lapic_switch_to_hv_timer(struct kvm_vcpu *vcpu)
1936*4882a593Smuzhiyun {
1937*4882a593Smuzhiyun 	restart_apic_timer(vcpu->arch.apic);
1938*4882a593Smuzhiyun }
1939*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_lapic_switch_to_hv_timer);
1940*4882a593Smuzhiyun 
kvm_lapic_switch_to_sw_timer(struct kvm_vcpu * vcpu)1941*4882a593Smuzhiyun void kvm_lapic_switch_to_sw_timer(struct kvm_vcpu *vcpu)
1942*4882a593Smuzhiyun {
1943*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
1944*4882a593Smuzhiyun 
1945*4882a593Smuzhiyun 	preempt_disable();
1946*4882a593Smuzhiyun 	/* Possibly the TSC deadline timer is not enabled yet */
1947*4882a593Smuzhiyun 	if (apic->lapic_timer.hv_timer_in_use)
1948*4882a593Smuzhiyun 		start_sw_timer(apic);
1949*4882a593Smuzhiyun 	preempt_enable();
1950*4882a593Smuzhiyun }
1951*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_lapic_switch_to_sw_timer);
1952*4882a593Smuzhiyun 
kvm_lapic_restart_hv_timer(struct kvm_vcpu * vcpu)1953*4882a593Smuzhiyun void kvm_lapic_restart_hv_timer(struct kvm_vcpu *vcpu)
1954*4882a593Smuzhiyun {
1955*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
1956*4882a593Smuzhiyun 
1957*4882a593Smuzhiyun 	WARN_ON(!apic->lapic_timer.hv_timer_in_use);
1958*4882a593Smuzhiyun 	restart_apic_timer(apic);
1959*4882a593Smuzhiyun }
1960*4882a593Smuzhiyun 
__start_apic_timer(struct kvm_lapic * apic,u32 count_reg)1961*4882a593Smuzhiyun static void __start_apic_timer(struct kvm_lapic *apic, u32 count_reg)
1962*4882a593Smuzhiyun {
1963*4882a593Smuzhiyun 	atomic_set(&apic->lapic_timer.pending, 0);
1964*4882a593Smuzhiyun 
1965*4882a593Smuzhiyun 	if ((apic_lvtt_period(apic) || apic_lvtt_oneshot(apic))
1966*4882a593Smuzhiyun 	    && !set_target_expiration(apic, count_reg))
1967*4882a593Smuzhiyun 		return;
1968*4882a593Smuzhiyun 
1969*4882a593Smuzhiyun 	restart_apic_timer(apic);
1970*4882a593Smuzhiyun }
1971*4882a593Smuzhiyun 
start_apic_timer(struct kvm_lapic * apic)1972*4882a593Smuzhiyun static void start_apic_timer(struct kvm_lapic *apic)
1973*4882a593Smuzhiyun {
1974*4882a593Smuzhiyun 	__start_apic_timer(apic, APIC_TMICT);
1975*4882a593Smuzhiyun }
1976*4882a593Smuzhiyun 
apic_manage_nmi_watchdog(struct kvm_lapic * apic,u32 lvt0_val)1977*4882a593Smuzhiyun static void apic_manage_nmi_watchdog(struct kvm_lapic *apic, u32 lvt0_val)
1978*4882a593Smuzhiyun {
1979*4882a593Smuzhiyun 	bool lvt0_in_nmi_mode = apic_lvt_nmi_mode(lvt0_val);
1980*4882a593Smuzhiyun 
1981*4882a593Smuzhiyun 	if (apic->lvt0_in_nmi_mode != lvt0_in_nmi_mode) {
1982*4882a593Smuzhiyun 		apic->lvt0_in_nmi_mode = lvt0_in_nmi_mode;
1983*4882a593Smuzhiyun 		if (lvt0_in_nmi_mode) {
1984*4882a593Smuzhiyun 			atomic_inc(&apic->vcpu->kvm->arch.vapics_in_nmi_mode);
1985*4882a593Smuzhiyun 		} else
1986*4882a593Smuzhiyun 			atomic_dec(&apic->vcpu->kvm->arch.vapics_in_nmi_mode);
1987*4882a593Smuzhiyun 	}
1988*4882a593Smuzhiyun }
1989*4882a593Smuzhiyun 
kvm_lapic_reg_write(struct kvm_lapic * apic,u32 reg,u32 val)1990*4882a593Smuzhiyun int kvm_lapic_reg_write(struct kvm_lapic *apic, u32 reg, u32 val)
1991*4882a593Smuzhiyun {
1992*4882a593Smuzhiyun 	int ret = 0;
1993*4882a593Smuzhiyun 
1994*4882a593Smuzhiyun 	trace_kvm_apic_write(reg, val);
1995*4882a593Smuzhiyun 
1996*4882a593Smuzhiyun 	switch (reg) {
1997*4882a593Smuzhiyun 	case APIC_ID:		/* Local APIC ID */
1998*4882a593Smuzhiyun 		if (!apic_x2apic_mode(apic))
1999*4882a593Smuzhiyun 			kvm_apic_set_xapic_id(apic, val >> 24);
2000*4882a593Smuzhiyun 		else
2001*4882a593Smuzhiyun 			ret = 1;
2002*4882a593Smuzhiyun 		break;
2003*4882a593Smuzhiyun 
2004*4882a593Smuzhiyun 	case APIC_TASKPRI:
2005*4882a593Smuzhiyun 		report_tpr_access(apic, true);
2006*4882a593Smuzhiyun 		apic_set_tpr(apic, val & 0xff);
2007*4882a593Smuzhiyun 		break;
2008*4882a593Smuzhiyun 
2009*4882a593Smuzhiyun 	case APIC_EOI:
2010*4882a593Smuzhiyun 		apic_set_eoi(apic);
2011*4882a593Smuzhiyun 		break;
2012*4882a593Smuzhiyun 
2013*4882a593Smuzhiyun 	case APIC_LDR:
2014*4882a593Smuzhiyun 		if (!apic_x2apic_mode(apic))
2015*4882a593Smuzhiyun 			kvm_apic_set_ldr(apic, val & APIC_LDR_MASK);
2016*4882a593Smuzhiyun 		else
2017*4882a593Smuzhiyun 			ret = 1;
2018*4882a593Smuzhiyun 		break;
2019*4882a593Smuzhiyun 
2020*4882a593Smuzhiyun 	case APIC_DFR:
2021*4882a593Smuzhiyun 		if (!apic_x2apic_mode(apic))
2022*4882a593Smuzhiyun 			kvm_apic_set_dfr(apic, val | 0x0FFFFFFF);
2023*4882a593Smuzhiyun 		else
2024*4882a593Smuzhiyun 			ret = 1;
2025*4882a593Smuzhiyun 		break;
2026*4882a593Smuzhiyun 
2027*4882a593Smuzhiyun 	case APIC_SPIV: {
2028*4882a593Smuzhiyun 		u32 mask = 0x3ff;
2029*4882a593Smuzhiyun 		if (kvm_lapic_get_reg(apic, APIC_LVR) & APIC_LVR_DIRECTED_EOI)
2030*4882a593Smuzhiyun 			mask |= APIC_SPIV_DIRECTED_EOI;
2031*4882a593Smuzhiyun 		apic_set_spiv(apic, val & mask);
2032*4882a593Smuzhiyun 		if (!(val & APIC_SPIV_APIC_ENABLED)) {
2033*4882a593Smuzhiyun 			int i;
2034*4882a593Smuzhiyun 			u32 lvt_val;
2035*4882a593Smuzhiyun 
2036*4882a593Smuzhiyun 			for (i = 0; i < KVM_APIC_LVT_NUM; i++) {
2037*4882a593Smuzhiyun 				lvt_val = kvm_lapic_get_reg(apic,
2038*4882a593Smuzhiyun 						       APIC_LVTT + 0x10 * i);
2039*4882a593Smuzhiyun 				kvm_lapic_set_reg(apic, APIC_LVTT + 0x10 * i,
2040*4882a593Smuzhiyun 					     lvt_val | APIC_LVT_MASKED);
2041*4882a593Smuzhiyun 			}
2042*4882a593Smuzhiyun 			apic_update_lvtt(apic);
2043*4882a593Smuzhiyun 			atomic_set(&apic->lapic_timer.pending, 0);
2044*4882a593Smuzhiyun 
2045*4882a593Smuzhiyun 		}
2046*4882a593Smuzhiyun 		break;
2047*4882a593Smuzhiyun 	}
2048*4882a593Smuzhiyun 	case APIC_ICR:
2049*4882a593Smuzhiyun 		/* No delay here, so we always clear the pending bit */
2050*4882a593Smuzhiyun 		val &= ~(1 << 12);
2051*4882a593Smuzhiyun 		kvm_apic_send_ipi(apic, val, kvm_lapic_get_reg(apic, APIC_ICR2));
2052*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_ICR, val);
2053*4882a593Smuzhiyun 		break;
2054*4882a593Smuzhiyun 
2055*4882a593Smuzhiyun 	case APIC_ICR2:
2056*4882a593Smuzhiyun 		if (!apic_x2apic_mode(apic))
2057*4882a593Smuzhiyun 			val &= 0xff000000;
2058*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_ICR2, val);
2059*4882a593Smuzhiyun 		break;
2060*4882a593Smuzhiyun 
2061*4882a593Smuzhiyun 	case APIC_LVT0:
2062*4882a593Smuzhiyun 		apic_manage_nmi_watchdog(apic, val);
2063*4882a593Smuzhiyun 		fallthrough;
2064*4882a593Smuzhiyun 	case APIC_LVTTHMR:
2065*4882a593Smuzhiyun 	case APIC_LVTPC:
2066*4882a593Smuzhiyun 	case APIC_LVT1:
2067*4882a593Smuzhiyun 	case APIC_LVTERR: {
2068*4882a593Smuzhiyun 		/* TODO: Check vector */
2069*4882a593Smuzhiyun 		size_t size;
2070*4882a593Smuzhiyun 		u32 index;
2071*4882a593Smuzhiyun 
2072*4882a593Smuzhiyun 		if (!kvm_apic_sw_enabled(apic))
2073*4882a593Smuzhiyun 			val |= APIC_LVT_MASKED;
2074*4882a593Smuzhiyun 		size = ARRAY_SIZE(apic_lvt_mask);
2075*4882a593Smuzhiyun 		index = array_index_nospec(
2076*4882a593Smuzhiyun 				(reg - APIC_LVTT) >> 4, size);
2077*4882a593Smuzhiyun 		val &= apic_lvt_mask[index];
2078*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, reg, val);
2079*4882a593Smuzhiyun 		break;
2080*4882a593Smuzhiyun 	}
2081*4882a593Smuzhiyun 
2082*4882a593Smuzhiyun 	case APIC_LVTT:
2083*4882a593Smuzhiyun 		if (!kvm_apic_sw_enabled(apic))
2084*4882a593Smuzhiyun 			val |= APIC_LVT_MASKED;
2085*4882a593Smuzhiyun 		val &= (apic_lvt_mask[0] | apic->lapic_timer.timer_mode_mask);
2086*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_LVTT, val);
2087*4882a593Smuzhiyun 		apic_update_lvtt(apic);
2088*4882a593Smuzhiyun 		break;
2089*4882a593Smuzhiyun 
2090*4882a593Smuzhiyun 	case APIC_TMICT:
2091*4882a593Smuzhiyun 		if (apic_lvtt_tscdeadline(apic))
2092*4882a593Smuzhiyun 			break;
2093*4882a593Smuzhiyun 
2094*4882a593Smuzhiyun 		hrtimer_cancel(&apic->lapic_timer.timer);
2095*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_TMICT, val);
2096*4882a593Smuzhiyun 		start_apic_timer(apic);
2097*4882a593Smuzhiyun 		break;
2098*4882a593Smuzhiyun 
2099*4882a593Smuzhiyun 	case APIC_TDCR: {
2100*4882a593Smuzhiyun 		uint32_t old_divisor = apic->divide_count;
2101*4882a593Smuzhiyun 
2102*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_TDCR, val & 0xb);
2103*4882a593Smuzhiyun 		update_divide_count(apic);
2104*4882a593Smuzhiyun 		if (apic->divide_count != old_divisor &&
2105*4882a593Smuzhiyun 				apic->lapic_timer.period) {
2106*4882a593Smuzhiyun 			hrtimer_cancel(&apic->lapic_timer.timer);
2107*4882a593Smuzhiyun 			update_target_expiration(apic, old_divisor);
2108*4882a593Smuzhiyun 			restart_apic_timer(apic);
2109*4882a593Smuzhiyun 		}
2110*4882a593Smuzhiyun 		break;
2111*4882a593Smuzhiyun 	}
2112*4882a593Smuzhiyun 	case APIC_ESR:
2113*4882a593Smuzhiyun 		if (apic_x2apic_mode(apic) && val != 0)
2114*4882a593Smuzhiyun 			ret = 1;
2115*4882a593Smuzhiyun 		break;
2116*4882a593Smuzhiyun 
2117*4882a593Smuzhiyun 	case APIC_SELF_IPI:
2118*4882a593Smuzhiyun 		if (apic_x2apic_mode(apic))
2119*4882a593Smuzhiyun 			kvm_apic_send_ipi(apic, APIC_DEST_SELF | (val & APIC_VECTOR_MASK), 0);
2120*4882a593Smuzhiyun 		else
2121*4882a593Smuzhiyun 			ret = 1;
2122*4882a593Smuzhiyun 		break;
2123*4882a593Smuzhiyun 	default:
2124*4882a593Smuzhiyun 		ret = 1;
2125*4882a593Smuzhiyun 		break;
2126*4882a593Smuzhiyun 	}
2127*4882a593Smuzhiyun 
2128*4882a593Smuzhiyun 	kvm_recalculate_apic_map(apic->vcpu->kvm);
2129*4882a593Smuzhiyun 
2130*4882a593Smuzhiyun 	return ret;
2131*4882a593Smuzhiyun }
2132*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_lapic_reg_write);
2133*4882a593Smuzhiyun 
apic_mmio_write(struct kvm_vcpu * vcpu,struct kvm_io_device * this,gpa_t address,int len,const void * data)2134*4882a593Smuzhiyun static int apic_mmio_write(struct kvm_vcpu *vcpu, struct kvm_io_device *this,
2135*4882a593Smuzhiyun 			    gpa_t address, int len, const void *data)
2136*4882a593Smuzhiyun {
2137*4882a593Smuzhiyun 	struct kvm_lapic *apic = to_lapic(this);
2138*4882a593Smuzhiyun 	unsigned int offset = address - apic->base_address;
2139*4882a593Smuzhiyun 	u32 val;
2140*4882a593Smuzhiyun 
2141*4882a593Smuzhiyun 	if (!apic_mmio_in_range(apic, address))
2142*4882a593Smuzhiyun 		return -EOPNOTSUPP;
2143*4882a593Smuzhiyun 
2144*4882a593Smuzhiyun 	if (!kvm_apic_hw_enabled(apic) || apic_x2apic_mode(apic)) {
2145*4882a593Smuzhiyun 		if (!kvm_check_has_quirk(vcpu->kvm,
2146*4882a593Smuzhiyun 					 KVM_X86_QUIRK_LAPIC_MMIO_HOLE))
2147*4882a593Smuzhiyun 			return -EOPNOTSUPP;
2148*4882a593Smuzhiyun 
2149*4882a593Smuzhiyun 		return 0;
2150*4882a593Smuzhiyun 	}
2151*4882a593Smuzhiyun 
2152*4882a593Smuzhiyun 	/*
2153*4882a593Smuzhiyun 	 * APIC register must be aligned on 128-bits boundary.
2154*4882a593Smuzhiyun 	 * 32/64/128 bits registers must be accessed thru 32 bits.
2155*4882a593Smuzhiyun 	 * Refer SDM 8.4.1
2156*4882a593Smuzhiyun 	 */
2157*4882a593Smuzhiyun 	if (len != 4 || (offset & 0xf))
2158*4882a593Smuzhiyun 		return 0;
2159*4882a593Smuzhiyun 
2160*4882a593Smuzhiyun 	val = *(u32*)data;
2161*4882a593Smuzhiyun 
2162*4882a593Smuzhiyun 	kvm_lapic_reg_write(apic, offset & 0xff0, val);
2163*4882a593Smuzhiyun 
2164*4882a593Smuzhiyun 	return 0;
2165*4882a593Smuzhiyun }
2166*4882a593Smuzhiyun 
kvm_lapic_set_eoi(struct kvm_vcpu * vcpu)2167*4882a593Smuzhiyun void kvm_lapic_set_eoi(struct kvm_vcpu *vcpu)
2168*4882a593Smuzhiyun {
2169*4882a593Smuzhiyun 	kvm_lapic_reg_write(vcpu->arch.apic, APIC_EOI, 0);
2170*4882a593Smuzhiyun }
2171*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_lapic_set_eoi);
2172*4882a593Smuzhiyun 
2173*4882a593Smuzhiyun /* emulate APIC access in a trap manner */
kvm_apic_write_nodecode(struct kvm_vcpu * vcpu,u32 offset)2174*4882a593Smuzhiyun void kvm_apic_write_nodecode(struct kvm_vcpu *vcpu, u32 offset)
2175*4882a593Smuzhiyun {
2176*4882a593Smuzhiyun 	u32 val = 0;
2177*4882a593Smuzhiyun 
2178*4882a593Smuzhiyun 	/* hw has done the conditional check and inst decode */
2179*4882a593Smuzhiyun 	offset &= 0xff0;
2180*4882a593Smuzhiyun 
2181*4882a593Smuzhiyun 	kvm_lapic_reg_read(vcpu->arch.apic, offset, 4, &val);
2182*4882a593Smuzhiyun 
2183*4882a593Smuzhiyun 	/* TODO: optimize to just emulate side effect w/o one more write */
2184*4882a593Smuzhiyun 	kvm_lapic_reg_write(vcpu->arch.apic, offset, val);
2185*4882a593Smuzhiyun }
2186*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_apic_write_nodecode);
2187*4882a593Smuzhiyun 
kvm_free_lapic(struct kvm_vcpu * vcpu)2188*4882a593Smuzhiyun void kvm_free_lapic(struct kvm_vcpu *vcpu)
2189*4882a593Smuzhiyun {
2190*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2191*4882a593Smuzhiyun 
2192*4882a593Smuzhiyun 	if (!vcpu->arch.apic)
2193*4882a593Smuzhiyun 		return;
2194*4882a593Smuzhiyun 
2195*4882a593Smuzhiyun 	hrtimer_cancel(&apic->lapic_timer.timer);
2196*4882a593Smuzhiyun 
2197*4882a593Smuzhiyun 	if (!(vcpu->arch.apic_base & MSR_IA32_APICBASE_ENABLE))
2198*4882a593Smuzhiyun 		static_key_slow_dec_deferred(&apic_hw_disabled);
2199*4882a593Smuzhiyun 
2200*4882a593Smuzhiyun 	if (!apic->sw_enabled)
2201*4882a593Smuzhiyun 		static_key_slow_dec_deferred(&apic_sw_disabled);
2202*4882a593Smuzhiyun 
2203*4882a593Smuzhiyun 	if (apic->regs)
2204*4882a593Smuzhiyun 		free_page((unsigned long)apic->regs);
2205*4882a593Smuzhiyun 
2206*4882a593Smuzhiyun 	kfree(apic);
2207*4882a593Smuzhiyun }
2208*4882a593Smuzhiyun 
2209*4882a593Smuzhiyun /*
2210*4882a593Smuzhiyun  *----------------------------------------------------------------------
2211*4882a593Smuzhiyun  * LAPIC interface
2212*4882a593Smuzhiyun  *----------------------------------------------------------------------
2213*4882a593Smuzhiyun  */
kvm_get_lapic_tscdeadline_msr(struct kvm_vcpu * vcpu)2214*4882a593Smuzhiyun u64 kvm_get_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu)
2215*4882a593Smuzhiyun {
2216*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2217*4882a593Smuzhiyun 
2218*4882a593Smuzhiyun 	if (!kvm_apic_present(vcpu) || !apic_lvtt_tscdeadline(apic))
2219*4882a593Smuzhiyun 		return 0;
2220*4882a593Smuzhiyun 
2221*4882a593Smuzhiyun 	return apic->lapic_timer.tscdeadline;
2222*4882a593Smuzhiyun }
2223*4882a593Smuzhiyun 
kvm_set_lapic_tscdeadline_msr(struct kvm_vcpu * vcpu,u64 data)2224*4882a593Smuzhiyun void kvm_set_lapic_tscdeadline_msr(struct kvm_vcpu *vcpu, u64 data)
2225*4882a593Smuzhiyun {
2226*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2227*4882a593Smuzhiyun 
2228*4882a593Smuzhiyun 	if (!kvm_apic_present(vcpu) || !apic_lvtt_tscdeadline(apic))
2229*4882a593Smuzhiyun 		return;
2230*4882a593Smuzhiyun 
2231*4882a593Smuzhiyun 	hrtimer_cancel(&apic->lapic_timer.timer);
2232*4882a593Smuzhiyun 	apic->lapic_timer.tscdeadline = data;
2233*4882a593Smuzhiyun 	start_apic_timer(apic);
2234*4882a593Smuzhiyun }
2235*4882a593Smuzhiyun 
kvm_lapic_set_tpr(struct kvm_vcpu * vcpu,unsigned long cr8)2236*4882a593Smuzhiyun void kvm_lapic_set_tpr(struct kvm_vcpu *vcpu, unsigned long cr8)
2237*4882a593Smuzhiyun {
2238*4882a593Smuzhiyun 	apic_set_tpr(vcpu->arch.apic, (cr8 & 0x0f) << 4);
2239*4882a593Smuzhiyun }
2240*4882a593Smuzhiyun 
kvm_lapic_get_cr8(struct kvm_vcpu * vcpu)2241*4882a593Smuzhiyun u64 kvm_lapic_get_cr8(struct kvm_vcpu *vcpu)
2242*4882a593Smuzhiyun {
2243*4882a593Smuzhiyun 	u64 tpr;
2244*4882a593Smuzhiyun 
2245*4882a593Smuzhiyun 	tpr = (u64) kvm_lapic_get_reg(vcpu->arch.apic, APIC_TASKPRI);
2246*4882a593Smuzhiyun 
2247*4882a593Smuzhiyun 	return (tpr & 0xf0) >> 4;
2248*4882a593Smuzhiyun }
2249*4882a593Smuzhiyun 
kvm_lapic_set_base(struct kvm_vcpu * vcpu,u64 value)2250*4882a593Smuzhiyun void kvm_lapic_set_base(struct kvm_vcpu *vcpu, u64 value)
2251*4882a593Smuzhiyun {
2252*4882a593Smuzhiyun 	u64 old_value = vcpu->arch.apic_base;
2253*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2254*4882a593Smuzhiyun 
2255*4882a593Smuzhiyun 	if (!apic)
2256*4882a593Smuzhiyun 		value |= MSR_IA32_APICBASE_BSP;
2257*4882a593Smuzhiyun 
2258*4882a593Smuzhiyun 	vcpu->arch.apic_base = value;
2259*4882a593Smuzhiyun 
2260*4882a593Smuzhiyun 	if ((old_value ^ value) & MSR_IA32_APICBASE_ENABLE)
2261*4882a593Smuzhiyun 		kvm_update_cpuid_runtime(vcpu);
2262*4882a593Smuzhiyun 
2263*4882a593Smuzhiyun 	if (!apic)
2264*4882a593Smuzhiyun 		return;
2265*4882a593Smuzhiyun 
2266*4882a593Smuzhiyun 	/* update jump label if enable bit changes */
2267*4882a593Smuzhiyun 	if ((old_value ^ value) & MSR_IA32_APICBASE_ENABLE) {
2268*4882a593Smuzhiyun 		if (value & MSR_IA32_APICBASE_ENABLE) {
2269*4882a593Smuzhiyun 			kvm_apic_set_xapic_id(apic, vcpu->vcpu_id);
2270*4882a593Smuzhiyun 			static_key_slow_dec_deferred(&apic_hw_disabled);
2271*4882a593Smuzhiyun 			/* Check if there are APF page ready requests pending */
2272*4882a593Smuzhiyun 			kvm_make_request(KVM_REQ_APF_READY, vcpu);
2273*4882a593Smuzhiyun 		} else {
2274*4882a593Smuzhiyun 			static_key_slow_inc(&apic_hw_disabled.key);
2275*4882a593Smuzhiyun 			atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
2276*4882a593Smuzhiyun 		}
2277*4882a593Smuzhiyun 	}
2278*4882a593Smuzhiyun 
2279*4882a593Smuzhiyun 	if (((old_value ^ value) & X2APIC_ENABLE) && (value & X2APIC_ENABLE))
2280*4882a593Smuzhiyun 		kvm_apic_set_x2apic_id(apic, vcpu->vcpu_id);
2281*4882a593Smuzhiyun 
2282*4882a593Smuzhiyun 	if ((old_value ^ value) & (MSR_IA32_APICBASE_ENABLE | X2APIC_ENABLE))
2283*4882a593Smuzhiyun 		kvm_x86_ops.set_virtual_apic_mode(vcpu);
2284*4882a593Smuzhiyun 
2285*4882a593Smuzhiyun 	apic->base_address = apic->vcpu->arch.apic_base &
2286*4882a593Smuzhiyun 			     MSR_IA32_APICBASE_BASE;
2287*4882a593Smuzhiyun 
2288*4882a593Smuzhiyun 	if ((value & MSR_IA32_APICBASE_ENABLE) &&
2289*4882a593Smuzhiyun 	     apic->base_address != APIC_DEFAULT_PHYS_BASE)
2290*4882a593Smuzhiyun 		pr_warn_once("APIC base relocation is unsupported by KVM");
2291*4882a593Smuzhiyun }
2292*4882a593Smuzhiyun 
kvm_apic_update_apicv(struct kvm_vcpu * vcpu)2293*4882a593Smuzhiyun void kvm_apic_update_apicv(struct kvm_vcpu *vcpu)
2294*4882a593Smuzhiyun {
2295*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2296*4882a593Smuzhiyun 
2297*4882a593Smuzhiyun 	if (vcpu->arch.apicv_active) {
2298*4882a593Smuzhiyun 		/* irr_pending is always true when apicv is activated. */
2299*4882a593Smuzhiyun 		apic->irr_pending = true;
2300*4882a593Smuzhiyun 		apic->isr_count = 1;
2301*4882a593Smuzhiyun 	} else {
2302*4882a593Smuzhiyun 		apic->irr_pending = (apic_search_irr(apic) != -1);
2303*4882a593Smuzhiyun 		apic->isr_count = count_vectors(apic->regs + APIC_ISR);
2304*4882a593Smuzhiyun 	}
2305*4882a593Smuzhiyun }
2306*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_apic_update_apicv);
2307*4882a593Smuzhiyun 
kvm_lapic_reset(struct kvm_vcpu * vcpu,bool init_event)2308*4882a593Smuzhiyun void kvm_lapic_reset(struct kvm_vcpu *vcpu, bool init_event)
2309*4882a593Smuzhiyun {
2310*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2311*4882a593Smuzhiyun 	int i;
2312*4882a593Smuzhiyun 
2313*4882a593Smuzhiyun 	if (!apic)
2314*4882a593Smuzhiyun 		return;
2315*4882a593Smuzhiyun 
2316*4882a593Smuzhiyun 	/* Stop the timer in case it's a reset to an active apic */
2317*4882a593Smuzhiyun 	hrtimer_cancel(&apic->lapic_timer.timer);
2318*4882a593Smuzhiyun 
2319*4882a593Smuzhiyun 	if (!init_event) {
2320*4882a593Smuzhiyun 		kvm_lapic_set_base(vcpu, APIC_DEFAULT_PHYS_BASE |
2321*4882a593Smuzhiyun 		                         MSR_IA32_APICBASE_ENABLE);
2322*4882a593Smuzhiyun 		kvm_apic_set_xapic_id(apic, vcpu->vcpu_id);
2323*4882a593Smuzhiyun 	}
2324*4882a593Smuzhiyun 	kvm_apic_set_version(apic->vcpu);
2325*4882a593Smuzhiyun 
2326*4882a593Smuzhiyun 	for (i = 0; i < KVM_APIC_LVT_NUM; i++)
2327*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_LVTT + 0x10 * i, APIC_LVT_MASKED);
2328*4882a593Smuzhiyun 	apic_update_lvtt(apic);
2329*4882a593Smuzhiyun 	if (kvm_vcpu_is_reset_bsp(vcpu) &&
2330*4882a593Smuzhiyun 	    kvm_check_has_quirk(vcpu->kvm, KVM_X86_QUIRK_LINT0_REENABLED))
2331*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_LVT0,
2332*4882a593Smuzhiyun 			     SET_APIC_DELIVERY_MODE(0, APIC_MODE_EXTINT));
2333*4882a593Smuzhiyun 	apic_manage_nmi_watchdog(apic, kvm_lapic_get_reg(apic, APIC_LVT0));
2334*4882a593Smuzhiyun 
2335*4882a593Smuzhiyun 	kvm_apic_set_dfr(apic, 0xffffffffU);
2336*4882a593Smuzhiyun 	apic_set_spiv(apic, 0xff);
2337*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_TASKPRI, 0);
2338*4882a593Smuzhiyun 	if (!apic_x2apic_mode(apic))
2339*4882a593Smuzhiyun 		kvm_apic_set_ldr(apic, 0);
2340*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_ESR, 0);
2341*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_ICR, 0);
2342*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_ICR2, 0);
2343*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_TDCR, 0);
2344*4882a593Smuzhiyun 	kvm_lapic_set_reg(apic, APIC_TMICT, 0);
2345*4882a593Smuzhiyun 	for (i = 0; i < 8; i++) {
2346*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_IRR + 0x10 * i, 0);
2347*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_ISR + 0x10 * i, 0);
2348*4882a593Smuzhiyun 		kvm_lapic_set_reg(apic, APIC_TMR + 0x10 * i, 0);
2349*4882a593Smuzhiyun 	}
2350*4882a593Smuzhiyun 	kvm_apic_update_apicv(vcpu);
2351*4882a593Smuzhiyun 	apic->highest_isr_cache = -1;
2352*4882a593Smuzhiyun 	update_divide_count(apic);
2353*4882a593Smuzhiyun 	atomic_set(&apic->lapic_timer.pending, 0);
2354*4882a593Smuzhiyun 	if (kvm_vcpu_is_bsp(vcpu))
2355*4882a593Smuzhiyun 		kvm_lapic_set_base(vcpu,
2356*4882a593Smuzhiyun 				vcpu->arch.apic_base | MSR_IA32_APICBASE_BSP);
2357*4882a593Smuzhiyun 	vcpu->arch.pv_eoi.msr_val = 0;
2358*4882a593Smuzhiyun 	apic_update_ppr(apic);
2359*4882a593Smuzhiyun 	if (vcpu->arch.apicv_active) {
2360*4882a593Smuzhiyun 		kvm_x86_ops.apicv_post_state_restore(vcpu);
2361*4882a593Smuzhiyun 		kvm_x86_ops.hwapic_irr_update(vcpu, -1);
2362*4882a593Smuzhiyun 		kvm_x86_ops.hwapic_isr_update(vcpu, -1);
2363*4882a593Smuzhiyun 	}
2364*4882a593Smuzhiyun 
2365*4882a593Smuzhiyun 	vcpu->arch.apic_arb_prio = 0;
2366*4882a593Smuzhiyun 	vcpu->arch.apic_attention = 0;
2367*4882a593Smuzhiyun 
2368*4882a593Smuzhiyun 	kvm_recalculate_apic_map(vcpu->kvm);
2369*4882a593Smuzhiyun }
2370*4882a593Smuzhiyun 
2371*4882a593Smuzhiyun /*
2372*4882a593Smuzhiyun  *----------------------------------------------------------------------
2373*4882a593Smuzhiyun  * timer interface
2374*4882a593Smuzhiyun  *----------------------------------------------------------------------
2375*4882a593Smuzhiyun  */
2376*4882a593Smuzhiyun 
lapic_is_periodic(struct kvm_lapic * apic)2377*4882a593Smuzhiyun static bool lapic_is_periodic(struct kvm_lapic *apic)
2378*4882a593Smuzhiyun {
2379*4882a593Smuzhiyun 	return apic_lvtt_period(apic);
2380*4882a593Smuzhiyun }
2381*4882a593Smuzhiyun 
apic_has_pending_timer(struct kvm_vcpu * vcpu)2382*4882a593Smuzhiyun int apic_has_pending_timer(struct kvm_vcpu *vcpu)
2383*4882a593Smuzhiyun {
2384*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2385*4882a593Smuzhiyun 
2386*4882a593Smuzhiyun 	if (apic_enabled(apic) && apic_lvt_enabled(apic, APIC_LVTT))
2387*4882a593Smuzhiyun 		return atomic_read(&apic->lapic_timer.pending);
2388*4882a593Smuzhiyun 
2389*4882a593Smuzhiyun 	return 0;
2390*4882a593Smuzhiyun }
2391*4882a593Smuzhiyun 
kvm_apic_local_deliver(struct kvm_lapic * apic,int lvt_type)2392*4882a593Smuzhiyun int kvm_apic_local_deliver(struct kvm_lapic *apic, int lvt_type)
2393*4882a593Smuzhiyun {
2394*4882a593Smuzhiyun 	u32 reg = kvm_lapic_get_reg(apic, lvt_type);
2395*4882a593Smuzhiyun 	int vector, mode, trig_mode;
2396*4882a593Smuzhiyun 
2397*4882a593Smuzhiyun 	if (kvm_apic_hw_enabled(apic) && !(reg & APIC_LVT_MASKED)) {
2398*4882a593Smuzhiyun 		vector = reg & APIC_VECTOR_MASK;
2399*4882a593Smuzhiyun 		mode = reg & APIC_MODE_MASK;
2400*4882a593Smuzhiyun 		trig_mode = reg & APIC_LVT_LEVEL_TRIGGER;
2401*4882a593Smuzhiyun 		return __apic_accept_irq(apic, mode, vector, 1, trig_mode,
2402*4882a593Smuzhiyun 					NULL);
2403*4882a593Smuzhiyun 	}
2404*4882a593Smuzhiyun 	return 0;
2405*4882a593Smuzhiyun }
2406*4882a593Smuzhiyun 
kvm_apic_nmi_wd_deliver(struct kvm_vcpu * vcpu)2407*4882a593Smuzhiyun void kvm_apic_nmi_wd_deliver(struct kvm_vcpu *vcpu)
2408*4882a593Smuzhiyun {
2409*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2410*4882a593Smuzhiyun 
2411*4882a593Smuzhiyun 	if (apic)
2412*4882a593Smuzhiyun 		kvm_apic_local_deliver(apic, APIC_LVT0);
2413*4882a593Smuzhiyun }
2414*4882a593Smuzhiyun 
2415*4882a593Smuzhiyun static const struct kvm_io_device_ops apic_mmio_ops = {
2416*4882a593Smuzhiyun 	.read     = apic_mmio_read,
2417*4882a593Smuzhiyun 	.write    = apic_mmio_write,
2418*4882a593Smuzhiyun };
2419*4882a593Smuzhiyun 
apic_timer_fn(struct hrtimer * data)2420*4882a593Smuzhiyun static enum hrtimer_restart apic_timer_fn(struct hrtimer *data)
2421*4882a593Smuzhiyun {
2422*4882a593Smuzhiyun 	struct kvm_timer *ktimer = container_of(data, struct kvm_timer, timer);
2423*4882a593Smuzhiyun 	struct kvm_lapic *apic = container_of(ktimer, struct kvm_lapic, lapic_timer);
2424*4882a593Smuzhiyun 
2425*4882a593Smuzhiyun 	apic_timer_expired(apic, true);
2426*4882a593Smuzhiyun 
2427*4882a593Smuzhiyun 	if (lapic_is_periodic(apic)) {
2428*4882a593Smuzhiyun 		advance_periodic_target_expiration(apic);
2429*4882a593Smuzhiyun 		hrtimer_add_expires_ns(&ktimer->timer, ktimer->period);
2430*4882a593Smuzhiyun 		return HRTIMER_RESTART;
2431*4882a593Smuzhiyun 	} else
2432*4882a593Smuzhiyun 		return HRTIMER_NORESTART;
2433*4882a593Smuzhiyun }
2434*4882a593Smuzhiyun 
kvm_create_lapic(struct kvm_vcpu * vcpu,int timer_advance_ns)2435*4882a593Smuzhiyun int kvm_create_lapic(struct kvm_vcpu *vcpu, int timer_advance_ns)
2436*4882a593Smuzhiyun {
2437*4882a593Smuzhiyun 	struct kvm_lapic *apic;
2438*4882a593Smuzhiyun 
2439*4882a593Smuzhiyun 	ASSERT(vcpu != NULL);
2440*4882a593Smuzhiyun 
2441*4882a593Smuzhiyun 	apic = kzalloc(sizeof(*apic), GFP_KERNEL_ACCOUNT);
2442*4882a593Smuzhiyun 	if (!apic)
2443*4882a593Smuzhiyun 		goto nomem;
2444*4882a593Smuzhiyun 
2445*4882a593Smuzhiyun 	vcpu->arch.apic = apic;
2446*4882a593Smuzhiyun 
2447*4882a593Smuzhiyun 	apic->regs = (void *)get_zeroed_page(GFP_KERNEL_ACCOUNT);
2448*4882a593Smuzhiyun 	if (!apic->regs) {
2449*4882a593Smuzhiyun 		printk(KERN_ERR "malloc apic regs error for vcpu %x\n",
2450*4882a593Smuzhiyun 		       vcpu->vcpu_id);
2451*4882a593Smuzhiyun 		goto nomem_free_apic;
2452*4882a593Smuzhiyun 	}
2453*4882a593Smuzhiyun 	apic->vcpu = vcpu;
2454*4882a593Smuzhiyun 
2455*4882a593Smuzhiyun 	hrtimer_init(&apic->lapic_timer.timer, CLOCK_MONOTONIC,
2456*4882a593Smuzhiyun 		     HRTIMER_MODE_ABS_HARD);
2457*4882a593Smuzhiyun 	apic->lapic_timer.timer.function = apic_timer_fn;
2458*4882a593Smuzhiyun 	if (timer_advance_ns == -1) {
2459*4882a593Smuzhiyun 		apic->lapic_timer.timer_advance_ns = LAPIC_TIMER_ADVANCE_NS_INIT;
2460*4882a593Smuzhiyun 		lapic_timer_advance_dynamic = true;
2461*4882a593Smuzhiyun 	} else {
2462*4882a593Smuzhiyun 		apic->lapic_timer.timer_advance_ns = timer_advance_ns;
2463*4882a593Smuzhiyun 		lapic_timer_advance_dynamic = false;
2464*4882a593Smuzhiyun 	}
2465*4882a593Smuzhiyun 
2466*4882a593Smuzhiyun 	/*
2467*4882a593Smuzhiyun 	 * APIC is created enabled. This will prevent kvm_lapic_set_base from
2468*4882a593Smuzhiyun 	 * thinking that APIC state has changed.
2469*4882a593Smuzhiyun 	 */
2470*4882a593Smuzhiyun 	vcpu->arch.apic_base = MSR_IA32_APICBASE_ENABLE;
2471*4882a593Smuzhiyun 	static_key_slow_inc(&apic_sw_disabled.key); /* sw disabled at reset */
2472*4882a593Smuzhiyun 	kvm_iodevice_init(&apic->dev, &apic_mmio_ops);
2473*4882a593Smuzhiyun 
2474*4882a593Smuzhiyun 	return 0;
2475*4882a593Smuzhiyun nomem_free_apic:
2476*4882a593Smuzhiyun 	kfree(apic);
2477*4882a593Smuzhiyun 	vcpu->arch.apic = NULL;
2478*4882a593Smuzhiyun nomem:
2479*4882a593Smuzhiyun 	return -ENOMEM;
2480*4882a593Smuzhiyun }
2481*4882a593Smuzhiyun 
kvm_apic_has_interrupt(struct kvm_vcpu * vcpu)2482*4882a593Smuzhiyun int kvm_apic_has_interrupt(struct kvm_vcpu *vcpu)
2483*4882a593Smuzhiyun {
2484*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2485*4882a593Smuzhiyun 	u32 ppr;
2486*4882a593Smuzhiyun 
2487*4882a593Smuzhiyun 	if (!kvm_apic_present(vcpu))
2488*4882a593Smuzhiyun 		return -1;
2489*4882a593Smuzhiyun 
2490*4882a593Smuzhiyun 	__apic_update_ppr(apic, &ppr);
2491*4882a593Smuzhiyun 	return apic_has_interrupt_for_ppr(apic, ppr);
2492*4882a593Smuzhiyun }
2493*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(kvm_apic_has_interrupt);
2494*4882a593Smuzhiyun 
kvm_apic_accept_pic_intr(struct kvm_vcpu * vcpu)2495*4882a593Smuzhiyun int kvm_apic_accept_pic_intr(struct kvm_vcpu *vcpu)
2496*4882a593Smuzhiyun {
2497*4882a593Smuzhiyun 	u32 lvt0 = kvm_lapic_get_reg(vcpu->arch.apic, APIC_LVT0);
2498*4882a593Smuzhiyun 
2499*4882a593Smuzhiyun 	if (!kvm_apic_hw_enabled(vcpu->arch.apic))
2500*4882a593Smuzhiyun 		return 1;
2501*4882a593Smuzhiyun 	if ((lvt0 & APIC_LVT_MASKED) == 0 &&
2502*4882a593Smuzhiyun 	    GET_APIC_DELIVERY_MODE(lvt0) == APIC_MODE_EXTINT)
2503*4882a593Smuzhiyun 		return 1;
2504*4882a593Smuzhiyun 	return 0;
2505*4882a593Smuzhiyun }
2506*4882a593Smuzhiyun 
kvm_inject_apic_timer_irqs(struct kvm_vcpu * vcpu)2507*4882a593Smuzhiyun void kvm_inject_apic_timer_irqs(struct kvm_vcpu *vcpu)
2508*4882a593Smuzhiyun {
2509*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2510*4882a593Smuzhiyun 
2511*4882a593Smuzhiyun 	if (atomic_read(&apic->lapic_timer.pending) > 0) {
2512*4882a593Smuzhiyun 		kvm_apic_inject_pending_timer_irqs(apic);
2513*4882a593Smuzhiyun 		atomic_set(&apic->lapic_timer.pending, 0);
2514*4882a593Smuzhiyun 	}
2515*4882a593Smuzhiyun }
2516*4882a593Smuzhiyun 
kvm_get_apic_interrupt(struct kvm_vcpu * vcpu)2517*4882a593Smuzhiyun int kvm_get_apic_interrupt(struct kvm_vcpu *vcpu)
2518*4882a593Smuzhiyun {
2519*4882a593Smuzhiyun 	int vector = kvm_apic_has_interrupt(vcpu);
2520*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2521*4882a593Smuzhiyun 	u32 ppr;
2522*4882a593Smuzhiyun 
2523*4882a593Smuzhiyun 	if (vector == -1)
2524*4882a593Smuzhiyun 		return -1;
2525*4882a593Smuzhiyun 
2526*4882a593Smuzhiyun 	/*
2527*4882a593Smuzhiyun 	 * We get here even with APIC virtualization enabled, if doing
2528*4882a593Smuzhiyun 	 * nested virtualization and L1 runs with the "acknowledge interrupt
2529*4882a593Smuzhiyun 	 * on exit" mode.  Then we cannot inject the interrupt via RVI,
2530*4882a593Smuzhiyun 	 * because the process would deliver it through the IDT.
2531*4882a593Smuzhiyun 	 */
2532*4882a593Smuzhiyun 
2533*4882a593Smuzhiyun 	apic_clear_irr(vector, apic);
2534*4882a593Smuzhiyun 	if (test_bit(vector, vcpu_to_synic(vcpu)->auto_eoi_bitmap)) {
2535*4882a593Smuzhiyun 		/*
2536*4882a593Smuzhiyun 		 * For auto-EOI interrupts, there might be another pending
2537*4882a593Smuzhiyun 		 * interrupt above PPR, so check whether to raise another
2538*4882a593Smuzhiyun 		 * KVM_REQ_EVENT.
2539*4882a593Smuzhiyun 		 */
2540*4882a593Smuzhiyun 		apic_update_ppr(apic);
2541*4882a593Smuzhiyun 	} else {
2542*4882a593Smuzhiyun 		/*
2543*4882a593Smuzhiyun 		 * For normal interrupts, PPR has been raised and there cannot
2544*4882a593Smuzhiyun 		 * be a higher-priority pending interrupt---except if there was
2545*4882a593Smuzhiyun 		 * a concurrent interrupt injection, but that would have
2546*4882a593Smuzhiyun 		 * triggered KVM_REQ_EVENT already.
2547*4882a593Smuzhiyun 		 */
2548*4882a593Smuzhiyun 		apic_set_isr(vector, apic);
2549*4882a593Smuzhiyun 		__apic_update_ppr(apic, &ppr);
2550*4882a593Smuzhiyun 	}
2551*4882a593Smuzhiyun 
2552*4882a593Smuzhiyun 	return vector;
2553*4882a593Smuzhiyun }
2554*4882a593Smuzhiyun 
kvm_apic_state_fixup(struct kvm_vcpu * vcpu,struct kvm_lapic_state * s,bool set)2555*4882a593Smuzhiyun static int kvm_apic_state_fixup(struct kvm_vcpu *vcpu,
2556*4882a593Smuzhiyun 		struct kvm_lapic_state *s, bool set)
2557*4882a593Smuzhiyun {
2558*4882a593Smuzhiyun 	if (apic_x2apic_mode(vcpu->arch.apic)) {
2559*4882a593Smuzhiyun 		u32 *id = (u32 *)(s->regs + APIC_ID);
2560*4882a593Smuzhiyun 		u32 *ldr = (u32 *)(s->regs + APIC_LDR);
2561*4882a593Smuzhiyun 
2562*4882a593Smuzhiyun 		if (vcpu->kvm->arch.x2apic_format) {
2563*4882a593Smuzhiyun 			if (*id != vcpu->vcpu_id)
2564*4882a593Smuzhiyun 				return -EINVAL;
2565*4882a593Smuzhiyun 		} else {
2566*4882a593Smuzhiyun 			if (set)
2567*4882a593Smuzhiyun 				*id >>= 24;
2568*4882a593Smuzhiyun 			else
2569*4882a593Smuzhiyun 				*id <<= 24;
2570*4882a593Smuzhiyun 		}
2571*4882a593Smuzhiyun 
2572*4882a593Smuzhiyun 		/* In x2APIC mode, the LDR is fixed and based on the id */
2573*4882a593Smuzhiyun 		if (set)
2574*4882a593Smuzhiyun 			*ldr = kvm_apic_calc_x2apic_ldr(*id);
2575*4882a593Smuzhiyun 	}
2576*4882a593Smuzhiyun 
2577*4882a593Smuzhiyun 	return 0;
2578*4882a593Smuzhiyun }
2579*4882a593Smuzhiyun 
kvm_apic_get_state(struct kvm_vcpu * vcpu,struct kvm_lapic_state * s)2580*4882a593Smuzhiyun int kvm_apic_get_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s)
2581*4882a593Smuzhiyun {
2582*4882a593Smuzhiyun 	memcpy(s->regs, vcpu->arch.apic->regs, sizeof(*s));
2583*4882a593Smuzhiyun 
2584*4882a593Smuzhiyun 	/*
2585*4882a593Smuzhiyun 	 * Get calculated timer current count for remaining timer period (if
2586*4882a593Smuzhiyun 	 * any) and store it in the returned register set.
2587*4882a593Smuzhiyun 	 */
2588*4882a593Smuzhiyun 	__kvm_lapic_set_reg(s->regs, APIC_TMCCT,
2589*4882a593Smuzhiyun 			    __apic_read(vcpu->arch.apic, APIC_TMCCT));
2590*4882a593Smuzhiyun 
2591*4882a593Smuzhiyun 	return kvm_apic_state_fixup(vcpu, s, false);
2592*4882a593Smuzhiyun }
2593*4882a593Smuzhiyun 
kvm_apic_set_state(struct kvm_vcpu * vcpu,struct kvm_lapic_state * s)2594*4882a593Smuzhiyun int kvm_apic_set_state(struct kvm_vcpu *vcpu, struct kvm_lapic_state *s)
2595*4882a593Smuzhiyun {
2596*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2597*4882a593Smuzhiyun 	int r;
2598*4882a593Smuzhiyun 
2599*4882a593Smuzhiyun 	kvm_lapic_set_base(vcpu, vcpu->arch.apic_base);
2600*4882a593Smuzhiyun 	/* set SPIV separately to get count of SW disabled APICs right */
2601*4882a593Smuzhiyun 	apic_set_spiv(apic, *((u32 *)(s->regs + APIC_SPIV)));
2602*4882a593Smuzhiyun 
2603*4882a593Smuzhiyun 	r = kvm_apic_state_fixup(vcpu, s, true);
2604*4882a593Smuzhiyun 	if (r) {
2605*4882a593Smuzhiyun 		kvm_recalculate_apic_map(vcpu->kvm);
2606*4882a593Smuzhiyun 		return r;
2607*4882a593Smuzhiyun 	}
2608*4882a593Smuzhiyun 	memcpy(vcpu->arch.apic->regs, s->regs, sizeof(*s));
2609*4882a593Smuzhiyun 
2610*4882a593Smuzhiyun 	atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
2611*4882a593Smuzhiyun 	kvm_recalculate_apic_map(vcpu->kvm);
2612*4882a593Smuzhiyun 	kvm_apic_set_version(vcpu);
2613*4882a593Smuzhiyun 
2614*4882a593Smuzhiyun 	apic_update_ppr(apic);
2615*4882a593Smuzhiyun 	hrtimer_cancel(&apic->lapic_timer.timer);
2616*4882a593Smuzhiyun 	apic_update_lvtt(apic);
2617*4882a593Smuzhiyun 	apic_manage_nmi_watchdog(apic, kvm_lapic_get_reg(apic, APIC_LVT0));
2618*4882a593Smuzhiyun 	update_divide_count(apic);
2619*4882a593Smuzhiyun 	__start_apic_timer(apic, APIC_TMCCT);
2620*4882a593Smuzhiyun 	kvm_apic_update_apicv(vcpu);
2621*4882a593Smuzhiyun 	apic->highest_isr_cache = -1;
2622*4882a593Smuzhiyun 	if (vcpu->arch.apicv_active) {
2623*4882a593Smuzhiyun 		kvm_x86_ops.apicv_post_state_restore(vcpu);
2624*4882a593Smuzhiyun 		kvm_x86_ops.hwapic_irr_update(vcpu,
2625*4882a593Smuzhiyun 				apic_find_highest_irr(apic));
2626*4882a593Smuzhiyun 		kvm_x86_ops.hwapic_isr_update(vcpu,
2627*4882a593Smuzhiyun 				apic_find_highest_isr(apic));
2628*4882a593Smuzhiyun 	}
2629*4882a593Smuzhiyun 	kvm_make_request(KVM_REQ_EVENT, vcpu);
2630*4882a593Smuzhiyun 	if (ioapic_in_kernel(vcpu->kvm))
2631*4882a593Smuzhiyun 		kvm_rtc_eoi_tracking_restore_one(vcpu);
2632*4882a593Smuzhiyun 
2633*4882a593Smuzhiyun 	vcpu->arch.apic_arb_prio = 0;
2634*4882a593Smuzhiyun 
2635*4882a593Smuzhiyun 	return 0;
2636*4882a593Smuzhiyun }
2637*4882a593Smuzhiyun 
__kvm_migrate_apic_timer(struct kvm_vcpu * vcpu)2638*4882a593Smuzhiyun void __kvm_migrate_apic_timer(struct kvm_vcpu *vcpu)
2639*4882a593Smuzhiyun {
2640*4882a593Smuzhiyun 	struct hrtimer *timer;
2641*4882a593Smuzhiyun 
2642*4882a593Smuzhiyun 	if (!lapic_in_kernel(vcpu) ||
2643*4882a593Smuzhiyun 		kvm_can_post_timer_interrupt(vcpu))
2644*4882a593Smuzhiyun 		return;
2645*4882a593Smuzhiyun 
2646*4882a593Smuzhiyun 	timer = &vcpu->arch.apic->lapic_timer.timer;
2647*4882a593Smuzhiyun 	if (hrtimer_cancel(timer))
2648*4882a593Smuzhiyun 		hrtimer_start_expires(timer, HRTIMER_MODE_ABS_HARD);
2649*4882a593Smuzhiyun }
2650*4882a593Smuzhiyun 
2651*4882a593Smuzhiyun /*
2652*4882a593Smuzhiyun  * apic_sync_pv_eoi_from_guest - called on vmexit or cancel interrupt
2653*4882a593Smuzhiyun  *
2654*4882a593Smuzhiyun  * Detect whether guest triggered PV EOI since the
2655*4882a593Smuzhiyun  * last entry. If yes, set EOI on guests's behalf.
2656*4882a593Smuzhiyun  * Clear PV EOI in guest memory in any case.
2657*4882a593Smuzhiyun  */
apic_sync_pv_eoi_from_guest(struct kvm_vcpu * vcpu,struct kvm_lapic * apic)2658*4882a593Smuzhiyun static void apic_sync_pv_eoi_from_guest(struct kvm_vcpu *vcpu,
2659*4882a593Smuzhiyun 					struct kvm_lapic *apic)
2660*4882a593Smuzhiyun {
2661*4882a593Smuzhiyun 	bool pending;
2662*4882a593Smuzhiyun 	int vector;
2663*4882a593Smuzhiyun 	/*
2664*4882a593Smuzhiyun 	 * PV EOI state is derived from KVM_APIC_PV_EOI_PENDING in host
2665*4882a593Smuzhiyun 	 * and KVM_PV_EOI_ENABLED in guest memory as follows:
2666*4882a593Smuzhiyun 	 *
2667*4882a593Smuzhiyun 	 * KVM_APIC_PV_EOI_PENDING is unset:
2668*4882a593Smuzhiyun 	 * 	-> host disabled PV EOI.
2669*4882a593Smuzhiyun 	 * KVM_APIC_PV_EOI_PENDING is set, KVM_PV_EOI_ENABLED is set:
2670*4882a593Smuzhiyun 	 * 	-> host enabled PV EOI, guest did not execute EOI yet.
2671*4882a593Smuzhiyun 	 * KVM_APIC_PV_EOI_PENDING is set, KVM_PV_EOI_ENABLED is unset:
2672*4882a593Smuzhiyun 	 * 	-> host enabled PV EOI, guest executed EOI.
2673*4882a593Smuzhiyun 	 */
2674*4882a593Smuzhiyun 	BUG_ON(!pv_eoi_enabled(vcpu));
2675*4882a593Smuzhiyun 	pending = pv_eoi_get_pending(vcpu);
2676*4882a593Smuzhiyun 	/*
2677*4882a593Smuzhiyun 	 * Clear pending bit in any case: it will be set again on vmentry.
2678*4882a593Smuzhiyun 	 * While this might not be ideal from performance point of view,
2679*4882a593Smuzhiyun 	 * this makes sure pv eoi is only enabled when we know it's safe.
2680*4882a593Smuzhiyun 	 */
2681*4882a593Smuzhiyun 	pv_eoi_clr_pending(vcpu);
2682*4882a593Smuzhiyun 	if (pending)
2683*4882a593Smuzhiyun 		return;
2684*4882a593Smuzhiyun 	vector = apic_set_eoi(apic);
2685*4882a593Smuzhiyun 	trace_kvm_pv_eoi(apic, vector);
2686*4882a593Smuzhiyun }
2687*4882a593Smuzhiyun 
kvm_lapic_sync_from_vapic(struct kvm_vcpu * vcpu)2688*4882a593Smuzhiyun void kvm_lapic_sync_from_vapic(struct kvm_vcpu *vcpu)
2689*4882a593Smuzhiyun {
2690*4882a593Smuzhiyun 	u32 data;
2691*4882a593Smuzhiyun 
2692*4882a593Smuzhiyun 	if (test_bit(KVM_APIC_PV_EOI_PENDING, &vcpu->arch.apic_attention))
2693*4882a593Smuzhiyun 		apic_sync_pv_eoi_from_guest(vcpu, vcpu->arch.apic);
2694*4882a593Smuzhiyun 
2695*4882a593Smuzhiyun 	if (!test_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention))
2696*4882a593Smuzhiyun 		return;
2697*4882a593Smuzhiyun 
2698*4882a593Smuzhiyun 	if (kvm_read_guest_cached(vcpu->kvm, &vcpu->arch.apic->vapic_cache, &data,
2699*4882a593Smuzhiyun 				  sizeof(u32)))
2700*4882a593Smuzhiyun 		return;
2701*4882a593Smuzhiyun 
2702*4882a593Smuzhiyun 	apic_set_tpr(vcpu->arch.apic, data & 0xff);
2703*4882a593Smuzhiyun }
2704*4882a593Smuzhiyun 
2705*4882a593Smuzhiyun /*
2706*4882a593Smuzhiyun  * apic_sync_pv_eoi_to_guest - called before vmentry
2707*4882a593Smuzhiyun  *
2708*4882a593Smuzhiyun  * Detect whether it's safe to enable PV EOI and
2709*4882a593Smuzhiyun  * if yes do so.
2710*4882a593Smuzhiyun  */
apic_sync_pv_eoi_to_guest(struct kvm_vcpu * vcpu,struct kvm_lapic * apic)2711*4882a593Smuzhiyun static void apic_sync_pv_eoi_to_guest(struct kvm_vcpu *vcpu,
2712*4882a593Smuzhiyun 					struct kvm_lapic *apic)
2713*4882a593Smuzhiyun {
2714*4882a593Smuzhiyun 	if (!pv_eoi_enabled(vcpu) ||
2715*4882a593Smuzhiyun 	    /* IRR set or many bits in ISR: could be nested. */
2716*4882a593Smuzhiyun 	    apic->irr_pending ||
2717*4882a593Smuzhiyun 	    /* Cache not set: could be safe but we don't bother. */
2718*4882a593Smuzhiyun 	    apic->highest_isr_cache == -1 ||
2719*4882a593Smuzhiyun 	    /* Need EOI to update ioapic. */
2720*4882a593Smuzhiyun 	    kvm_ioapic_handles_vector(apic, apic->highest_isr_cache)) {
2721*4882a593Smuzhiyun 		/*
2722*4882a593Smuzhiyun 		 * PV EOI was disabled by apic_sync_pv_eoi_from_guest
2723*4882a593Smuzhiyun 		 * so we need not do anything here.
2724*4882a593Smuzhiyun 		 */
2725*4882a593Smuzhiyun 		return;
2726*4882a593Smuzhiyun 	}
2727*4882a593Smuzhiyun 
2728*4882a593Smuzhiyun 	pv_eoi_set_pending(apic->vcpu);
2729*4882a593Smuzhiyun }
2730*4882a593Smuzhiyun 
kvm_lapic_sync_to_vapic(struct kvm_vcpu * vcpu)2731*4882a593Smuzhiyun void kvm_lapic_sync_to_vapic(struct kvm_vcpu *vcpu)
2732*4882a593Smuzhiyun {
2733*4882a593Smuzhiyun 	u32 data, tpr;
2734*4882a593Smuzhiyun 	int max_irr, max_isr;
2735*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2736*4882a593Smuzhiyun 
2737*4882a593Smuzhiyun 	apic_sync_pv_eoi_to_guest(vcpu, apic);
2738*4882a593Smuzhiyun 
2739*4882a593Smuzhiyun 	if (!test_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention))
2740*4882a593Smuzhiyun 		return;
2741*4882a593Smuzhiyun 
2742*4882a593Smuzhiyun 	tpr = kvm_lapic_get_reg(apic, APIC_TASKPRI) & 0xff;
2743*4882a593Smuzhiyun 	max_irr = apic_find_highest_irr(apic);
2744*4882a593Smuzhiyun 	if (max_irr < 0)
2745*4882a593Smuzhiyun 		max_irr = 0;
2746*4882a593Smuzhiyun 	max_isr = apic_find_highest_isr(apic);
2747*4882a593Smuzhiyun 	if (max_isr < 0)
2748*4882a593Smuzhiyun 		max_isr = 0;
2749*4882a593Smuzhiyun 	data = (tpr & 0xff) | ((max_isr & 0xf0) << 8) | (max_irr << 24);
2750*4882a593Smuzhiyun 
2751*4882a593Smuzhiyun 	kvm_write_guest_cached(vcpu->kvm, &vcpu->arch.apic->vapic_cache, &data,
2752*4882a593Smuzhiyun 				sizeof(u32));
2753*4882a593Smuzhiyun }
2754*4882a593Smuzhiyun 
kvm_lapic_set_vapic_addr(struct kvm_vcpu * vcpu,gpa_t vapic_addr)2755*4882a593Smuzhiyun int kvm_lapic_set_vapic_addr(struct kvm_vcpu *vcpu, gpa_t vapic_addr)
2756*4882a593Smuzhiyun {
2757*4882a593Smuzhiyun 	if (vapic_addr) {
2758*4882a593Smuzhiyun 		if (kvm_gfn_to_hva_cache_init(vcpu->kvm,
2759*4882a593Smuzhiyun 					&vcpu->arch.apic->vapic_cache,
2760*4882a593Smuzhiyun 					vapic_addr, sizeof(u32)))
2761*4882a593Smuzhiyun 			return -EINVAL;
2762*4882a593Smuzhiyun 		__set_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention);
2763*4882a593Smuzhiyun 	} else {
2764*4882a593Smuzhiyun 		__clear_bit(KVM_APIC_CHECK_VAPIC, &vcpu->arch.apic_attention);
2765*4882a593Smuzhiyun 	}
2766*4882a593Smuzhiyun 
2767*4882a593Smuzhiyun 	vcpu->arch.apic->vapic_addr = vapic_addr;
2768*4882a593Smuzhiyun 	return 0;
2769*4882a593Smuzhiyun }
2770*4882a593Smuzhiyun 
kvm_x2apic_msr_write(struct kvm_vcpu * vcpu,u32 msr,u64 data)2771*4882a593Smuzhiyun int kvm_x2apic_msr_write(struct kvm_vcpu *vcpu, u32 msr, u64 data)
2772*4882a593Smuzhiyun {
2773*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2774*4882a593Smuzhiyun 	u32 reg = (msr - APIC_BASE_MSR) << 4;
2775*4882a593Smuzhiyun 
2776*4882a593Smuzhiyun 	if (!lapic_in_kernel(vcpu) || !apic_x2apic_mode(apic))
2777*4882a593Smuzhiyun 		return 1;
2778*4882a593Smuzhiyun 
2779*4882a593Smuzhiyun 	if (reg == APIC_ICR2)
2780*4882a593Smuzhiyun 		return 1;
2781*4882a593Smuzhiyun 
2782*4882a593Smuzhiyun 	/* if this is ICR write vector before command */
2783*4882a593Smuzhiyun 	if (reg == APIC_ICR)
2784*4882a593Smuzhiyun 		kvm_lapic_reg_write(apic, APIC_ICR2, (u32)(data >> 32));
2785*4882a593Smuzhiyun 	return kvm_lapic_reg_write(apic, reg, (u32)data);
2786*4882a593Smuzhiyun }
2787*4882a593Smuzhiyun 
kvm_x2apic_msr_read(struct kvm_vcpu * vcpu,u32 msr,u64 * data)2788*4882a593Smuzhiyun int kvm_x2apic_msr_read(struct kvm_vcpu *vcpu, u32 msr, u64 *data)
2789*4882a593Smuzhiyun {
2790*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2791*4882a593Smuzhiyun 	u32 reg = (msr - APIC_BASE_MSR) << 4, low, high = 0;
2792*4882a593Smuzhiyun 
2793*4882a593Smuzhiyun 	if (!lapic_in_kernel(vcpu) || !apic_x2apic_mode(apic))
2794*4882a593Smuzhiyun 		return 1;
2795*4882a593Smuzhiyun 
2796*4882a593Smuzhiyun 	if (reg == APIC_DFR || reg == APIC_ICR2)
2797*4882a593Smuzhiyun 		return 1;
2798*4882a593Smuzhiyun 
2799*4882a593Smuzhiyun 	if (kvm_lapic_reg_read(apic, reg, 4, &low))
2800*4882a593Smuzhiyun 		return 1;
2801*4882a593Smuzhiyun 	if (reg == APIC_ICR)
2802*4882a593Smuzhiyun 		kvm_lapic_reg_read(apic, APIC_ICR2, 4, &high);
2803*4882a593Smuzhiyun 
2804*4882a593Smuzhiyun 	*data = (((u64)high) << 32) | low;
2805*4882a593Smuzhiyun 
2806*4882a593Smuzhiyun 	return 0;
2807*4882a593Smuzhiyun }
2808*4882a593Smuzhiyun 
kvm_hv_vapic_msr_write(struct kvm_vcpu * vcpu,u32 reg,u64 data)2809*4882a593Smuzhiyun int kvm_hv_vapic_msr_write(struct kvm_vcpu *vcpu, u32 reg, u64 data)
2810*4882a593Smuzhiyun {
2811*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2812*4882a593Smuzhiyun 
2813*4882a593Smuzhiyun 	if (!lapic_in_kernel(vcpu))
2814*4882a593Smuzhiyun 		return 1;
2815*4882a593Smuzhiyun 
2816*4882a593Smuzhiyun 	/* if this is ICR write vector before command */
2817*4882a593Smuzhiyun 	if (reg == APIC_ICR)
2818*4882a593Smuzhiyun 		kvm_lapic_reg_write(apic, APIC_ICR2, (u32)(data >> 32));
2819*4882a593Smuzhiyun 	return kvm_lapic_reg_write(apic, reg, (u32)data);
2820*4882a593Smuzhiyun }
2821*4882a593Smuzhiyun 
kvm_hv_vapic_msr_read(struct kvm_vcpu * vcpu,u32 reg,u64 * data)2822*4882a593Smuzhiyun int kvm_hv_vapic_msr_read(struct kvm_vcpu *vcpu, u32 reg, u64 *data)
2823*4882a593Smuzhiyun {
2824*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2825*4882a593Smuzhiyun 	u32 low, high = 0;
2826*4882a593Smuzhiyun 
2827*4882a593Smuzhiyun 	if (!lapic_in_kernel(vcpu))
2828*4882a593Smuzhiyun 		return 1;
2829*4882a593Smuzhiyun 
2830*4882a593Smuzhiyun 	if (kvm_lapic_reg_read(apic, reg, 4, &low))
2831*4882a593Smuzhiyun 		return 1;
2832*4882a593Smuzhiyun 	if (reg == APIC_ICR)
2833*4882a593Smuzhiyun 		kvm_lapic_reg_read(apic, APIC_ICR2, 4, &high);
2834*4882a593Smuzhiyun 
2835*4882a593Smuzhiyun 	*data = (((u64)high) << 32) | low;
2836*4882a593Smuzhiyun 
2837*4882a593Smuzhiyun 	return 0;
2838*4882a593Smuzhiyun }
2839*4882a593Smuzhiyun 
kvm_lapic_enable_pv_eoi(struct kvm_vcpu * vcpu,u64 data,unsigned long len)2840*4882a593Smuzhiyun int kvm_lapic_enable_pv_eoi(struct kvm_vcpu *vcpu, u64 data, unsigned long len)
2841*4882a593Smuzhiyun {
2842*4882a593Smuzhiyun 	u64 addr = data & ~KVM_MSR_ENABLED;
2843*4882a593Smuzhiyun 	struct gfn_to_hva_cache *ghc = &vcpu->arch.pv_eoi.data;
2844*4882a593Smuzhiyun 	unsigned long new_len;
2845*4882a593Smuzhiyun 
2846*4882a593Smuzhiyun 	if (!IS_ALIGNED(addr, 4))
2847*4882a593Smuzhiyun 		return 1;
2848*4882a593Smuzhiyun 
2849*4882a593Smuzhiyun 	vcpu->arch.pv_eoi.msr_val = data;
2850*4882a593Smuzhiyun 	if (!pv_eoi_enabled(vcpu))
2851*4882a593Smuzhiyun 		return 0;
2852*4882a593Smuzhiyun 
2853*4882a593Smuzhiyun 	if (addr == ghc->gpa && len <= ghc->len)
2854*4882a593Smuzhiyun 		new_len = ghc->len;
2855*4882a593Smuzhiyun 	else
2856*4882a593Smuzhiyun 		new_len = len;
2857*4882a593Smuzhiyun 
2858*4882a593Smuzhiyun 	return kvm_gfn_to_hva_cache_init(vcpu->kvm, ghc, addr, new_len);
2859*4882a593Smuzhiyun }
2860*4882a593Smuzhiyun 
kvm_apic_accept_events(struct kvm_vcpu * vcpu)2861*4882a593Smuzhiyun void kvm_apic_accept_events(struct kvm_vcpu *vcpu)
2862*4882a593Smuzhiyun {
2863*4882a593Smuzhiyun 	struct kvm_lapic *apic = vcpu->arch.apic;
2864*4882a593Smuzhiyun 	u8 sipi_vector;
2865*4882a593Smuzhiyun 	unsigned long pe;
2866*4882a593Smuzhiyun 
2867*4882a593Smuzhiyun 	if (!lapic_in_kernel(vcpu) || !apic->pending_events)
2868*4882a593Smuzhiyun 		return;
2869*4882a593Smuzhiyun 
2870*4882a593Smuzhiyun 	/*
2871*4882a593Smuzhiyun 	 * INITs are latched while CPU is in specific states
2872*4882a593Smuzhiyun 	 * (SMM, VMX non-root mode, SVM with GIF=0).
2873*4882a593Smuzhiyun 	 * Because a CPU cannot be in these states immediately
2874*4882a593Smuzhiyun 	 * after it has processed an INIT signal (and thus in
2875*4882a593Smuzhiyun 	 * KVM_MP_STATE_INIT_RECEIVED state), just eat SIPIs
2876*4882a593Smuzhiyun 	 * and leave the INIT pending.
2877*4882a593Smuzhiyun 	 */
2878*4882a593Smuzhiyun 	if (kvm_vcpu_latch_init(vcpu)) {
2879*4882a593Smuzhiyun 		WARN_ON_ONCE(vcpu->arch.mp_state == KVM_MP_STATE_INIT_RECEIVED);
2880*4882a593Smuzhiyun 		if (test_bit(KVM_APIC_SIPI, &apic->pending_events))
2881*4882a593Smuzhiyun 			clear_bit(KVM_APIC_SIPI, &apic->pending_events);
2882*4882a593Smuzhiyun 		return;
2883*4882a593Smuzhiyun 	}
2884*4882a593Smuzhiyun 
2885*4882a593Smuzhiyun 	pe = xchg(&apic->pending_events, 0);
2886*4882a593Smuzhiyun 	if (test_bit(KVM_APIC_INIT, &pe)) {
2887*4882a593Smuzhiyun 		kvm_vcpu_reset(vcpu, true);
2888*4882a593Smuzhiyun 		if (kvm_vcpu_is_bsp(apic->vcpu))
2889*4882a593Smuzhiyun 			vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE;
2890*4882a593Smuzhiyun 		else
2891*4882a593Smuzhiyun 			vcpu->arch.mp_state = KVM_MP_STATE_INIT_RECEIVED;
2892*4882a593Smuzhiyun 	}
2893*4882a593Smuzhiyun 	if (test_bit(KVM_APIC_SIPI, &pe) &&
2894*4882a593Smuzhiyun 	    vcpu->arch.mp_state == KVM_MP_STATE_INIT_RECEIVED) {
2895*4882a593Smuzhiyun 		/* evaluate pending_events before reading the vector */
2896*4882a593Smuzhiyun 		smp_rmb();
2897*4882a593Smuzhiyun 		sipi_vector = apic->sipi_vector;
2898*4882a593Smuzhiyun 		kvm_vcpu_deliver_sipi_vector(vcpu, sipi_vector);
2899*4882a593Smuzhiyun 		vcpu->arch.mp_state = KVM_MP_STATE_RUNNABLE;
2900*4882a593Smuzhiyun 	}
2901*4882a593Smuzhiyun }
2902*4882a593Smuzhiyun 
kvm_lapic_init(void)2903*4882a593Smuzhiyun void kvm_lapic_init(void)
2904*4882a593Smuzhiyun {
2905*4882a593Smuzhiyun 	/* do not patch jump label more than once per second */
2906*4882a593Smuzhiyun 	jump_label_rate_limit(&apic_hw_disabled, HZ);
2907*4882a593Smuzhiyun 	jump_label_rate_limit(&apic_sw_disabled, HZ);
2908*4882a593Smuzhiyun }
2909*4882a593Smuzhiyun 
kvm_lapic_exit(void)2910*4882a593Smuzhiyun void kvm_lapic_exit(void)
2911*4882a593Smuzhiyun {
2912*4882a593Smuzhiyun 	static_key_deferred_flush(&apic_hw_disabled);
2913*4882a593Smuzhiyun 	static_key_deferred_flush(&apic_sw_disabled);
2914*4882a593Smuzhiyun }
2915