xref: /openbmc/linux/arch/powerpc/kvm/e500.h (revision 95e9fd10)
1 /*
2  * Copyright (C) 2008-2011 Freescale Semiconductor, Inc. All rights reserved.
3  *
4  * Author: Yu Liu <yu.liu@freescale.com>
5  *         Scott Wood <scottwood@freescale.com>
6  *         Ashish Kalra <ashish.kalra@freescale.com>
7  *         Varun Sethi <varun.sethi@freescale.com>
8  *
9  * Description:
10  * This file is based on arch/powerpc/kvm/44x_tlb.h and
11  * arch/powerpc/include/asm/kvm_44x.h by Hollis Blanchard <hollisb@us.ibm.com>,
12  * Copyright IBM Corp. 2007-2008
13  *
14  * This program is free software; you can redistribute it and/or modify
15  * it under the terms of the GNU General Public License, version 2, as
16  * published by the Free Software Foundation.
17  */
18 
19 #ifndef KVM_E500_H
20 #define KVM_E500_H
21 
22 #include <linux/kvm_host.h>
23 #include <asm/mmu-book3e.h>
24 #include <asm/tlb.h>
25 
26 #define E500_PID_NUM   3
27 #define E500_TLB_NUM   2
28 
29 #define E500_TLB_VALID 1
30 #define E500_TLB_DIRTY 2
31 #define E500_TLB_BITMAP 4
32 
33 struct tlbe_ref {
34 	pfn_t pfn;
35 	unsigned int flags; /* E500_TLB_* */
36 };
37 
38 struct tlbe_priv {
39 	struct tlbe_ref ref; /* TLB0 only -- TLB1 uses tlb_refs */
40 };
41 
42 #ifdef CONFIG_KVM_E500V2
43 struct vcpu_id_table;
44 #endif
45 
46 struct kvmppc_e500_tlb_params {
47 	int entries, ways, sets;
48 };
49 
50 struct kvmppc_vcpu_e500 {
51 	struct kvm_vcpu vcpu;
52 
53 	/* Unmodified copy of the guest's TLB -- shared with host userspace. */
54 	struct kvm_book3e_206_tlb_entry *gtlb_arch;
55 
56 	/* Starting entry number in gtlb_arch[] */
57 	int gtlb_offset[E500_TLB_NUM];
58 
59 	/* KVM internal information associated with each guest TLB entry */
60 	struct tlbe_priv *gtlb_priv[E500_TLB_NUM];
61 
62 	struct kvmppc_e500_tlb_params gtlb_params[E500_TLB_NUM];
63 
64 	unsigned int gtlb_nv[E500_TLB_NUM];
65 
66 	/*
67 	 * information associated with each host TLB entry --
68 	 * TLB1 only for now.  If/when guest TLB1 entries can be
69 	 * mapped with host TLB0, this will be used for that too.
70 	 *
71 	 * We don't want to use this for guest TLB0 because then we'd
72 	 * have the overhead of doing the translation again even if
73 	 * the entry is still in the guest TLB (e.g. we swapped out
74 	 * and back, and our host TLB entries got evicted).
75 	 */
76 	struct tlbe_ref *tlb_refs[E500_TLB_NUM];
77 	unsigned int host_tlb1_nv;
78 
79 	u32 svr;
80 	u32 l1csr0;
81 	u32 l1csr1;
82 	u32 hid0;
83 	u32 hid1;
84 	u64 mcar;
85 
86 	struct page **shared_tlb_pages;
87 	int num_shared_tlb_pages;
88 
89 	u64 *g2h_tlb1_map;
90 	unsigned int *h2g_tlb1_rmap;
91 
92 	/* Minimum and maximum address mapped my TLB1 */
93 	unsigned long tlb1_min_eaddr;
94 	unsigned long tlb1_max_eaddr;
95 
96 #ifdef CONFIG_KVM_E500V2
97 	u32 pid[E500_PID_NUM];
98 
99 	/* vcpu id table */
100 	struct vcpu_id_table *idt;
101 #endif
102 };
103 
104 static inline struct kvmppc_vcpu_e500 *to_e500(struct kvm_vcpu *vcpu)
105 {
106 	return container_of(vcpu, struct kvmppc_vcpu_e500, vcpu);
107 }
108 
109 
110 /* This geometry is the legacy default -- can be overridden by userspace */
111 #define KVM_E500_TLB0_WAY_SIZE		128
112 #define KVM_E500_TLB0_WAY_NUM		2
113 
114 #define KVM_E500_TLB0_SIZE  (KVM_E500_TLB0_WAY_SIZE * KVM_E500_TLB0_WAY_NUM)
115 #define KVM_E500_TLB1_SIZE  16
116 
117 #define index_of(tlbsel, esel)	(((tlbsel) << 16) | ((esel) & 0xFFFF))
118 #define tlbsel_of(index)	((index) >> 16)
119 #define esel_of(index)		((index) & 0xFFFF)
120 
121 #define E500_TLB_USER_PERM_MASK (MAS3_UX|MAS3_UR|MAS3_UW)
122 #define E500_TLB_SUPER_PERM_MASK (MAS3_SX|MAS3_SR|MAS3_SW)
123 #define MAS2_ATTRIB_MASK \
124 	  (MAS2_X0 | MAS2_X1)
125 #define MAS3_ATTRIB_MASK \
126 	  (MAS3_U0 | MAS3_U1 | MAS3_U2 | MAS3_U3 \
127 	   | E500_TLB_USER_PERM_MASK | E500_TLB_SUPER_PERM_MASK)
128 
129 int kvmppc_e500_emul_mt_mmucsr0(struct kvmppc_vcpu_e500 *vcpu_e500,
130 				ulong value);
131 int kvmppc_e500_emul_tlbwe(struct kvm_vcpu *vcpu);
132 int kvmppc_e500_emul_tlbre(struct kvm_vcpu *vcpu);
133 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, int ra, int rb);
134 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int rt, int ra, int rb);
135 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, int rb);
136 int kvmppc_e500_tlb_init(struct kvmppc_vcpu_e500 *vcpu_e500);
137 void kvmppc_e500_tlb_uninit(struct kvmppc_vcpu_e500 *vcpu_e500);
138 
139 void kvmppc_get_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
140 int kvmppc_set_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs);
141 
142 
143 #ifdef CONFIG_KVM_E500V2
144 unsigned int kvmppc_e500_get_sid(struct kvmppc_vcpu_e500 *vcpu_e500,
145 				 unsigned int as, unsigned int gid,
146 				 unsigned int pr, int avoid_recursion);
147 #endif
148 
149 /* TLB helper functions */
150 static inline unsigned int
151 get_tlb_size(const struct kvm_book3e_206_tlb_entry *tlbe)
152 {
153 	return (tlbe->mas1 >> 7) & 0x1f;
154 }
155 
156 static inline gva_t get_tlb_eaddr(const struct kvm_book3e_206_tlb_entry *tlbe)
157 {
158 	return tlbe->mas2 & 0xfffff000;
159 }
160 
161 static inline u64 get_tlb_bytes(const struct kvm_book3e_206_tlb_entry *tlbe)
162 {
163 	unsigned int pgsize = get_tlb_size(tlbe);
164 	return 1ULL << 10 << pgsize;
165 }
166 
167 static inline gva_t get_tlb_end(const struct kvm_book3e_206_tlb_entry *tlbe)
168 {
169 	u64 bytes = get_tlb_bytes(tlbe);
170 	return get_tlb_eaddr(tlbe) + bytes - 1;
171 }
172 
173 static inline u64 get_tlb_raddr(const struct kvm_book3e_206_tlb_entry *tlbe)
174 {
175 	return tlbe->mas7_3 & ~0xfffULL;
176 }
177 
178 static inline unsigned int
179 get_tlb_tid(const struct kvm_book3e_206_tlb_entry *tlbe)
180 {
181 	return (tlbe->mas1 >> 16) & 0xff;
182 }
183 
184 static inline unsigned int
185 get_tlb_ts(const struct kvm_book3e_206_tlb_entry *tlbe)
186 {
187 	return (tlbe->mas1 >> 12) & 0x1;
188 }
189 
190 static inline unsigned int
191 get_tlb_v(const struct kvm_book3e_206_tlb_entry *tlbe)
192 {
193 	return (tlbe->mas1 >> 31) & 0x1;
194 }
195 
196 static inline unsigned int
197 get_tlb_iprot(const struct kvm_book3e_206_tlb_entry *tlbe)
198 {
199 	return (tlbe->mas1 >> 30) & 0x1;
200 }
201 
202 static inline unsigned int
203 get_tlb_tsize(const struct kvm_book3e_206_tlb_entry *tlbe)
204 {
205 	return (tlbe->mas1 & MAS1_TSIZE_MASK) >> MAS1_TSIZE_SHIFT;
206 }
207 
208 static inline unsigned int get_cur_pid(struct kvm_vcpu *vcpu)
209 {
210 	return vcpu->arch.pid & 0xff;
211 }
212 
213 static inline unsigned int get_cur_as(struct kvm_vcpu *vcpu)
214 {
215 	return !!(vcpu->arch.shared->msr & (MSR_IS | MSR_DS));
216 }
217 
218 static inline unsigned int get_cur_pr(struct kvm_vcpu *vcpu)
219 {
220 	return !!(vcpu->arch.shared->msr & MSR_PR);
221 }
222 
223 static inline unsigned int get_cur_spid(const struct kvm_vcpu *vcpu)
224 {
225 	return (vcpu->arch.shared->mas6 >> 16) & 0xff;
226 }
227 
228 static inline unsigned int get_cur_sas(const struct kvm_vcpu *vcpu)
229 {
230 	return vcpu->arch.shared->mas6 & 0x1;
231 }
232 
233 static inline unsigned int get_tlb_tlbsel(const struct kvm_vcpu *vcpu)
234 {
235 	/*
236 	 * Manual says that tlbsel has 2 bits wide.
237 	 * Since we only have two TLBs, only lower bit is used.
238 	 */
239 	return (vcpu->arch.shared->mas0 >> 28) & 0x1;
240 }
241 
242 static inline unsigned int get_tlb_nv_bit(const struct kvm_vcpu *vcpu)
243 {
244 	return vcpu->arch.shared->mas0 & 0xfff;
245 }
246 
247 static inline unsigned int get_tlb_esel_bit(const struct kvm_vcpu *vcpu)
248 {
249 	return (vcpu->arch.shared->mas0 >> 16) & 0xfff;
250 }
251 
252 static inline int tlbe_is_host_safe(const struct kvm_vcpu *vcpu,
253 			const struct kvm_book3e_206_tlb_entry *tlbe)
254 {
255 	gpa_t gpa;
256 
257 	if (!get_tlb_v(tlbe))
258 		return 0;
259 
260 #ifndef CONFIG_KVM_BOOKE_HV
261 	/* Does it match current guest AS? */
262 	/* XXX what about IS != DS? */
263 	if (get_tlb_ts(tlbe) != !!(vcpu->arch.shared->msr & MSR_IS))
264 		return 0;
265 #endif
266 
267 	gpa = get_tlb_raddr(tlbe);
268 	if (!gfn_to_memslot(vcpu->kvm, gpa >> PAGE_SHIFT))
269 		/* Mapping is not for RAM. */
270 		return 0;
271 
272 	return 1;
273 }
274 
275 static inline struct kvm_book3e_206_tlb_entry *get_entry(
276 	struct kvmppc_vcpu_e500 *vcpu_e500, int tlbsel, int entry)
277 {
278 	int offset = vcpu_e500->gtlb_offset[tlbsel];
279 	return &vcpu_e500->gtlb_arch[offset + entry];
280 }
281 
282 void kvmppc_e500_tlbil_one(struct kvmppc_vcpu_e500 *vcpu_e500,
283 			   struct kvm_book3e_206_tlb_entry *gtlbe);
284 void kvmppc_e500_tlbil_all(struct kvmppc_vcpu_e500 *vcpu_e500);
285 
286 #ifdef CONFIG_KVM_BOOKE_HV
287 #define kvmppc_e500_get_tlb_stid(vcpu, gtlbe)       get_tlb_tid(gtlbe)
288 #define get_tlbmiss_tid(vcpu)           get_cur_pid(vcpu)
289 #define get_tlb_sts(gtlbe)              (gtlbe->mas1 & MAS1_TS)
290 #else
291 unsigned int kvmppc_e500_get_tlb_stid(struct kvm_vcpu *vcpu,
292 				      struct kvm_book3e_206_tlb_entry *gtlbe);
293 
294 static inline unsigned int get_tlbmiss_tid(struct kvm_vcpu *vcpu)
295 {
296 	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
297 	unsigned int tidseld = (vcpu->arch.shared->mas4 >> 16) & 0xf;
298 
299 	return vcpu_e500->pid[tidseld];
300 }
301 
302 /* Force TS=1 for all guest mappings. */
303 #define get_tlb_sts(gtlbe)              (MAS1_TS)
304 #endif /* !BOOKE_HV */
305 
306 #endif /* KVM_E500_H */
307