~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/powerpc/kvm/e500_mmu.c

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 // SPDX-License-Identifier: GPL-2.0-only
  2 /*
  3  * Copyright (C) 2008-2013 Freescale Semiconductor, Inc. All rights reserved.
  4  *
  5  * Author: Yu Liu, yu.liu@freescale.com
  6  *         Scott Wood, scottwood@freescale.com
  7  *         Ashish Kalra, ashish.kalra@freescale.com
  8  *         Varun Sethi, varun.sethi@freescale.com
  9  *         Alexander Graf, agraf@suse.de
 10  *
 11  * Description:
 12  * This file is based on arch/powerpc/kvm/44x_tlb.c,
 13  * by Hollis Blanchard <hollisb@us.ibm.com>.
 14  */
 15 
 16 #include <linux/kernel.h>
 17 #include <linux/types.h>
 18 #include <linux/slab.h>
 19 #include <linux/string.h>
 20 #include <linux/kvm.h>
 21 #include <linux/kvm_host.h>
 22 #include <linux/highmem.h>
 23 #include <linux/log2.h>
 24 #include <linux/uaccess.h>
 25 #include <linux/sched.h>
 26 #include <linux/rwsem.h>
 27 #include <linux/vmalloc.h>
 28 #include <linux/hugetlb.h>
 29 #include <asm/kvm_ppc.h>
 30 
 31 #include "e500.h"
 32 #include "trace_booke.h"
 33 #include "timing.h"
 34 #include "e500_mmu_host.h"
 35 
 36 static inline unsigned int gtlb0_get_next_victim(
 37                 struct kvmppc_vcpu_e500 *vcpu_e500)
 38 {
 39         unsigned int victim;
 40 
 41         victim = vcpu_e500->gtlb_nv[0]++;
 42         if (unlikely(vcpu_e500->gtlb_nv[0] >= vcpu_e500->gtlb_params[0].ways))
 43                 vcpu_e500->gtlb_nv[0] = 0;
 44 
 45         return victim;
 46 }
 47 
 48 static int tlb0_set_base(gva_t addr, int sets, int ways)
 49 {
 50         int set_base;
 51 
 52         set_base = (addr >> PAGE_SHIFT) & (sets - 1);
 53         set_base *= ways;
 54 
 55         return set_base;
 56 }
 57 
 58 static int gtlb0_set_base(struct kvmppc_vcpu_e500 *vcpu_e500, gva_t addr)
 59 {
 60         return tlb0_set_base(addr, vcpu_e500->gtlb_params[0].sets,
 61                              vcpu_e500->gtlb_params[0].ways);
 62 }
 63 
 64 static unsigned int get_tlb_esel(struct kvm_vcpu *vcpu, int tlbsel)
 65 {
 66         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
 67         int esel = get_tlb_esel_bit(vcpu);
 68 
 69         if (tlbsel == 0) {
 70                 esel &= vcpu_e500->gtlb_params[0].ways - 1;
 71                 esel += gtlb0_set_base(vcpu_e500, vcpu->arch.shared->mas2);
 72         } else {
 73                 esel &= vcpu_e500->gtlb_params[tlbsel].entries - 1;
 74         }
 75 
 76         return esel;
 77 }
 78 
 79 /* Search the guest TLB for a matching entry. */
 80 static int kvmppc_e500_tlb_index(struct kvmppc_vcpu_e500 *vcpu_e500,
 81                 gva_t eaddr, int tlbsel, unsigned int pid, int as)
 82 {
 83         int size = vcpu_e500->gtlb_params[tlbsel].entries;
 84         unsigned int set_base, offset;
 85         int i;
 86 
 87         if (tlbsel == 0) {
 88                 set_base = gtlb0_set_base(vcpu_e500, eaddr);
 89                 size = vcpu_e500->gtlb_params[0].ways;
 90         } else {
 91                 if (eaddr < vcpu_e500->tlb1_min_eaddr ||
 92                                 eaddr > vcpu_e500->tlb1_max_eaddr)
 93                         return -1;
 94                 set_base = 0;
 95         }
 96 
 97         offset = vcpu_e500->gtlb_offset[tlbsel];
 98 
 99         for (i = 0; i < size; i++) {
100                 struct kvm_book3e_206_tlb_entry *tlbe =
101                         &vcpu_e500->gtlb_arch[offset + set_base + i];
102                 unsigned int tid;
103 
104                 if (eaddr < get_tlb_eaddr(tlbe))
105                         continue;
106 
107                 if (eaddr > get_tlb_end(tlbe))
108                         continue;
109 
110                 tid = get_tlb_tid(tlbe);
111                 if (tid && (tid != pid))
112                         continue;
113 
114                 if (!get_tlb_v(tlbe))
115                         continue;
116 
117                 if (get_tlb_ts(tlbe) != as && as != -1)
118                         continue;
119 
120                 return set_base + i;
121         }
122 
123         return -1;
124 }
125 
126 static inline void kvmppc_e500_deliver_tlb_miss(struct kvm_vcpu *vcpu,
127                 gva_t eaddr, int as)
128 {
129         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
130         unsigned int victim, tsized;
131         int tlbsel;
132 
133         /* since we only have two TLBs, only lower bit is used. */
134         tlbsel = (vcpu->arch.shared->mas4 >> 28) & 0x1;
135         victim = (tlbsel == 0) ? gtlb0_get_next_victim(vcpu_e500) : 0;
136         tsized = (vcpu->arch.shared->mas4 >> 7) & 0x1f;
137 
138         vcpu->arch.shared->mas0 = MAS0_TLBSEL(tlbsel) | MAS0_ESEL(victim)
139                 | MAS0_NV(vcpu_e500->gtlb_nv[tlbsel]);
140         vcpu->arch.shared->mas1 = MAS1_VALID | (as ? MAS1_TS : 0)
141                 | MAS1_TID(get_tlbmiss_tid(vcpu))
142                 | MAS1_TSIZE(tsized);
143         vcpu->arch.shared->mas2 = (eaddr & MAS2_EPN)
144                 | (vcpu->arch.shared->mas4 & MAS2_ATTRIB_MASK);
145         vcpu->arch.shared->mas7_3 &= MAS3_U0 | MAS3_U1 | MAS3_U2 | MAS3_U3;
146         vcpu->arch.shared->mas6 = (vcpu->arch.shared->mas6 & MAS6_SPID1)
147                 | (get_cur_pid(vcpu) << 16)
148                 | (as ? MAS6_SAS : 0);
149 }
150 
151 static void kvmppc_recalc_tlb1map_range(struct kvmppc_vcpu_e500 *vcpu_e500)
152 {
153         int size = vcpu_e500->gtlb_params[1].entries;
154         unsigned int offset;
155         gva_t eaddr;
156         int i;
157 
158         vcpu_e500->tlb1_min_eaddr = ~0UL;
159         vcpu_e500->tlb1_max_eaddr = 0;
160         offset = vcpu_e500->gtlb_offset[1];
161 
162         for (i = 0; i < size; i++) {
163                 struct kvm_book3e_206_tlb_entry *tlbe =
164                         &vcpu_e500->gtlb_arch[offset + i];
165 
166                 if (!get_tlb_v(tlbe))
167                         continue;
168 
169                 eaddr = get_tlb_eaddr(tlbe);
170                 vcpu_e500->tlb1_min_eaddr =
171                                 min(vcpu_e500->tlb1_min_eaddr, eaddr);
172 
173                 eaddr = get_tlb_end(tlbe);
174                 vcpu_e500->tlb1_max_eaddr =
175                                 max(vcpu_e500->tlb1_max_eaddr, eaddr);
176         }
177 }
178 
179 static int kvmppc_need_recalc_tlb1map_range(struct kvmppc_vcpu_e500 *vcpu_e500,
180                                 struct kvm_book3e_206_tlb_entry *gtlbe)
181 {
182         unsigned long start, end, size;
183 
184         size = get_tlb_bytes(gtlbe);
185         start = get_tlb_eaddr(gtlbe) & ~(size - 1);
186         end = start + size - 1;
187 
188         return vcpu_e500->tlb1_min_eaddr == start ||
189                         vcpu_e500->tlb1_max_eaddr == end;
190 }
191 
192 /* This function is supposed to be called for a adding a new valid tlb entry */
193 static void kvmppc_set_tlb1map_range(struct kvm_vcpu *vcpu,
194                                 struct kvm_book3e_206_tlb_entry *gtlbe)
195 {
196         unsigned long start, end, size;
197         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
198 
199         if (!get_tlb_v(gtlbe))
200                 return;
201 
202         size = get_tlb_bytes(gtlbe);
203         start = get_tlb_eaddr(gtlbe) & ~(size - 1);
204         end = start + size - 1;
205 
206         vcpu_e500->tlb1_min_eaddr = min(vcpu_e500->tlb1_min_eaddr, start);
207         vcpu_e500->tlb1_max_eaddr = max(vcpu_e500->tlb1_max_eaddr, end);
208 }
209 
210 static inline int kvmppc_e500_gtlbe_invalidate(
211                                 struct kvmppc_vcpu_e500 *vcpu_e500,
212                                 int tlbsel, int esel)
213 {
214         struct kvm_book3e_206_tlb_entry *gtlbe =
215                 get_entry(vcpu_e500, tlbsel, esel);
216 
217         if (unlikely(get_tlb_iprot(gtlbe)))
218                 return -1;
219 
220         if (tlbsel == 1 && kvmppc_need_recalc_tlb1map_range(vcpu_e500, gtlbe))
221                 kvmppc_recalc_tlb1map_range(vcpu_e500);
222 
223         gtlbe->mas1 = 0;
224 
225         return 0;
226 }
227 
228 int kvmppc_e500_emul_mt_mmucsr0(struct kvmppc_vcpu_e500 *vcpu_e500, ulong value)
229 {
230         int esel;
231 
232         if (value & MMUCSR0_TLB0FI)
233                 for (esel = 0; esel < vcpu_e500->gtlb_params[0].entries; esel++)
234                         kvmppc_e500_gtlbe_invalidate(vcpu_e500, 0, esel);
235         if (value & MMUCSR0_TLB1FI)
236                 for (esel = 0; esel < vcpu_e500->gtlb_params[1].entries; esel++)
237                         kvmppc_e500_gtlbe_invalidate(vcpu_e500, 1, esel);
238 
239         /* Invalidate all host shadow mappings */
240         kvmppc_core_flush_tlb(&vcpu_e500->vcpu);
241 
242         return EMULATE_DONE;
243 }
244 
245 int kvmppc_e500_emul_tlbivax(struct kvm_vcpu *vcpu, gva_t ea)
246 {
247         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
248         unsigned int ia;
249         int esel, tlbsel;
250 
251         ia = (ea >> 2) & 0x1;
252 
253         /* since we only have two TLBs, only lower bit is used. */
254         tlbsel = (ea >> 3) & 0x1;
255 
256         if (ia) {
257                 /* invalidate all entries */
258                 for (esel = 0; esel < vcpu_e500->gtlb_params[tlbsel].entries;
259                      esel++)
260                         kvmppc_e500_gtlbe_invalidate(vcpu_e500, tlbsel, esel);
261         } else {
262                 ea &= 0xfffff000;
263                 esel = kvmppc_e500_tlb_index(vcpu_e500, ea, tlbsel,
264                                 get_cur_pid(vcpu), -1);
265                 if (esel >= 0)
266                         kvmppc_e500_gtlbe_invalidate(vcpu_e500, tlbsel, esel);
267         }
268 
269         /* Invalidate all host shadow mappings */
270         kvmppc_core_flush_tlb(&vcpu_e500->vcpu);
271 
272         return EMULATE_DONE;
273 }
274 
275 static void tlbilx_all(struct kvmppc_vcpu_e500 *vcpu_e500, int tlbsel,
276                        int pid, int type)
277 {
278         struct kvm_book3e_206_tlb_entry *tlbe;
279         int tid, esel;
280 
281         /* invalidate all entries */
282         for (esel = 0; esel < vcpu_e500->gtlb_params[tlbsel].entries; esel++) {
283                 tlbe = get_entry(vcpu_e500, tlbsel, esel);
284                 tid = get_tlb_tid(tlbe);
285                 if (type == 0 || tid == pid) {
286                         inval_gtlbe_on_host(vcpu_e500, tlbsel, esel);
287                         kvmppc_e500_gtlbe_invalidate(vcpu_e500, tlbsel, esel);
288                 }
289         }
290 }
291 
292 static void tlbilx_one(struct kvmppc_vcpu_e500 *vcpu_e500, int pid,
293                        gva_t ea)
294 {
295         int tlbsel, esel;
296 
297         for (tlbsel = 0; tlbsel < 2; tlbsel++) {
298                 esel = kvmppc_e500_tlb_index(vcpu_e500, ea, tlbsel, pid, -1);
299                 if (esel >= 0) {
300                         inval_gtlbe_on_host(vcpu_e500, tlbsel, esel);
301                         kvmppc_e500_gtlbe_invalidate(vcpu_e500, tlbsel, esel);
302                         break;
303                 }
304         }
305 }
306 
307 int kvmppc_e500_emul_tlbilx(struct kvm_vcpu *vcpu, int type, gva_t ea)
308 {
309         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
310         int pid = get_cur_spid(vcpu);
311 
312         if (type == 0 || type == 1) {
313                 tlbilx_all(vcpu_e500, 0, pid, type);
314                 tlbilx_all(vcpu_e500, 1, pid, type);
315         } else if (type == 3) {
316                 tlbilx_one(vcpu_e500, pid, ea);
317         }
318 
319         return EMULATE_DONE;
320 }
321 
322 int kvmppc_e500_emul_tlbre(struct kvm_vcpu *vcpu)
323 {
324         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
325         int tlbsel, esel;
326         struct kvm_book3e_206_tlb_entry *gtlbe;
327 
328         tlbsel = get_tlb_tlbsel(vcpu);
329         esel = get_tlb_esel(vcpu, tlbsel);
330 
331         gtlbe = get_entry(vcpu_e500, tlbsel, esel);
332         vcpu->arch.shared->mas0 &= ~MAS0_NV(~0);
333         vcpu->arch.shared->mas0 |= MAS0_NV(vcpu_e500->gtlb_nv[tlbsel]);
334         vcpu->arch.shared->mas1 = gtlbe->mas1;
335         vcpu->arch.shared->mas2 = gtlbe->mas2;
336         vcpu->arch.shared->mas7_3 = gtlbe->mas7_3;
337 
338         return EMULATE_DONE;
339 }
340 
341 int kvmppc_e500_emul_tlbsx(struct kvm_vcpu *vcpu, gva_t ea)
342 {
343         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
344         int as = !!get_cur_sas(vcpu);
345         unsigned int pid = get_cur_spid(vcpu);
346         int esel, tlbsel;
347         struct kvm_book3e_206_tlb_entry *gtlbe = NULL;
348 
349         for (tlbsel = 0; tlbsel < 2; tlbsel++) {
350                 esel = kvmppc_e500_tlb_index(vcpu_e500, ea, tlbsel, pid, as);
351                 if (esel >= 0) {
352                         gtlbe = get_entry(vcpu_e500, tlbsel, esel);
353                         break;
354                 }
355         }
356 
357         if (gtlbe) {
358                 esel &= vcpu_e500->gtlb_params[tlbsel].ways - 1;
359 
360                 vcpu->arch.shared->mas0 = MAS0_TLBSEL(tlbsel) | MAS0_ESEL(esel)
361                         | MAS0_NV(vcpu_e500->gtlb_nv[tlbsel]);
362                 vcpu->arch.shared->mas1 = gtlbe->mas1;
363                 vcpu->arch.shared->mas2 = gtlbe->mas2;
364                 vcpu->arch.shared->mas7_3 = gtlbe->mas7_3;
365         } else {
366                 int victim;
367 
368                 /* since we only have two TLBs, only lower bit is used. */
369                 tlbsel = vcpu->arch.shared->mas4 >> 28 & 0x1;
370                 victim = (tlbsel == 0) ? gtlb0_get_next_victim(vcpu_e500) : 0;
371 
372                 vcpu->arch.shared->mas0 = MAS0_TLBSEL(tlbsel)
373                         | MAS0_ESEL(victim)
374                         | MAS0_NV(vcpu_e500->gtlb_nv[tlbsel]);
375                 vcpu->arch.shared->mas1 =
376                           (vcpu->arch.shared->mas6 & MAS6_SPID0)
377                         | ((vcpu->arch.shared->mas6 & MAS6_SAS) ? MAS1_TS : 0)
378                         | (vcpu->arch.shared->mas4 & MAS4_TSIZED(~0));
379                 vcpu->arch.shared->mas2 &= MAS2_EPN;
380                 vcpu->arch.shared->mas2 |= vcpu->arch.shared->mas4 &
381                                            MAS2_ATTRIB_MASK;
382                 vcpu->arch.shared->mas7_3 &= MAS3_U0 | MAS3_U1 |
383                                              MAS3_U2 | MAS3_U3;
384         }
385 
386         kvmppc_set_exit_type(vcpu, EMULATED_TLBSX_EXITS);
387         return EMULATE_DONE;
388 }
389 
390 int kvmppc_e500_emul_tlbwe(struct kvm_vcpu *vcpu)
391 {
392         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
393         struct kvm_book3e_206_tlb_entry *gtlbe;
394         int tlbsel, esel;
395         int recal = 0;
396         int idx;
397 
398         tlbsel = get_tlb_tlbsel(vcpu);
399         esel = get_tlb_esel(vcpu, tlbsel);
400 
401         gtlbe = get_entry(vcpu_e500, tlbsel, esel);
402 
403         if (get_tlb_v(gtlbe)) {
404                 inval_gtlbe_on_host(vcpu_e500, tlbsel, esel);
405                 if ((tlbsel == 1) &&
406                         kvmppc_need_recalc_tlb1map_range(vcpu_e500, gtlbe))
407                         recal = 1;
408         }
409 
410         gtlbe->mas1 = vcpu->arch.shared->mas1;
411         gtlbe->mas2 = vcpu->arch.shared->mas2;
412         if (!(vcpu->arch.shared->msr & MSR_CM))
413                 gtlbe->mas2 &= 0xffffffffUL;
414         gtlbe->mas7_3 = vcpu->arch.shared->mas7_3;
415 
416         trace_kvm_booke206_gtlb_write(vcpu->arch.shared->mas0, gtlbe->mas1,
417                                       gtlbe->mas2, gtlbe->mas7_3);
418 
419         if (tlbsel == 1) {
420                 /*
421                  * If a valid tlb1 entry is overwritten then recalculate the
422                  * min/max TLB1 map address range otherwise no need to look
423                  * in tlb1 array.
424                  */
425                 if (recal)
426                         kvmppc_recalc_tlb1map_range(vcpu_e500);
427                 else
428                         kvmppc_set_tlb1map_range(vcpu, gtlbe);
429         }
430 
431         idx = srcu_read_lock(&vcpu->kvm->srcu);
432 
433         /* Invalidate shadow mappings for the about-to-be-clobbered TLBE. */
434         if (tlbe_is_host_safe(vcpu, gtlbe)) {
435                 u64 eaddr = get_tlb_eaddr(gtlbe);
436                 u64 raddr = get_tlb_raddr(gtlbe);
437 
438                 if (tlbsel == 0) {
439                         gtlbe->mas1 &= ~MAS1_TSIZE(~0);
440                         gtlbe->mas1 |= MAS1_TSIZE(BOOK3E_PAGESZ_4K);
441                 }
442 
443                 /* Premap the faulting page */
444                 kvmppc_mmu_map(vcpu, eaddr, raddr, index_of(tlbsel, esel));
445         }
446 
447         srcu_read_unlock(&vcpu->kvm->srcu, idx);
448 
449         kvmppc_set_exit_type(vcpu, EMULATED_TLBWE_EXITS);
450         return EMULATE_DONE;
451 }
452 
453 static int kvmppc_e500_tlb_search(struct kvm_vcpu *vcpu,
454                                   gva_t eaddr, unsigned int pid, int as)
455 {
456         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
457         int esel, tlbsel;
458 
459         for (tlbsel = 0; tlbsel < 2; tlbsel++) {
460                 esel = kvmppc_e500_tlb_index(vcpu_e500, eaddr, tlbsel, pid, as);
461                 if (esel >= 0)
462                         return index_of(tlbsel, esel);
463         }
464 
465         return -1;
466 }
467 
468 /* 'linear_address' is actually an encoding of AS|PID|EADDR . */
469 int kvmppc_core_vcpu_translate(struct kvm_vcpu *vcpu,
470                                struct kvm_translation *tr)
471 {
472         int index;
473         gva_t eaddr;
474         u8 pid;
475         u8 as;
476 
477         eaddr = tr->linear_address;
478         pid = (tr->linear_address >> 32) & 0xff;
479         as = (tr->linear_address >> 40) & 0x1;
480 
481         index = kvmppc_e500_tlb_search(vcpu, eaddr, pid, as);
482         if (index < 0) {
483                 tr->valid = 0;
484                 return 0;
485         }
486 
487         tr->physical_address = kvmppc_mmu_xlate(vcpu, index, eaddr);
488         /* XXX what does "writeable" and "usermode" even mean? */
489         tr->valid = 1;
490 
491         return 0;
492 }
493 
494 
495 int kvmppc_mmu_itlb_index(struct kvm_vcpu *vcpu, gva_t eaddr)
496 {
497         unsigned int as = !!(vcpu->arch.shared->msr & MSR_IS);
498 
499         return kvmppc_e500_tlb_search(vcpu, eaddr, get_cur_pid(vcpu), as);
500 }
501 
502 int kvmppc_mmu_dtlb_index(struct kvm_vcpu *vcpu, gva_t eaddr)
503 {
504         unsigned int as = !!(vcpu->arch.shared->msr & MSR_DS);
505 
506         return kvmppc_e500_tlb_search(vcpu, eaddr, get_cur_pid(vcpu), as);
507 }
508 
509 void kvmppc_mmu_itlb_miss(struct kvm_vcpu *vcpu)
510 {
511         unsigned int as = !!(vcpu->arch.shared->msr & MSR_IS);
512 
513         kvmppc_e500_deliver_tlb_miss(vcpu, vcpu->arch.regs.nip, as);
514 }
515 
516 void kvmppc_mmu_dtlb_miss(struct kvm_vcpu *vcpu)
517 {
518         unsigned int as = !!(vcpu->arch.shared->msr & MSR_DS);
519 
520         kvmppc_e500_deliver_tlb_miss(vcpu, vcpu->arch.fault_dear, as);
521 }
522 
523 gpa_t kvmppc_mmu_xlate(struct kvm_vcpu *vcpu, unsigned int index,
524                         gva_t eaddr)
525 {
526         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
527         struct kvm_book3e_206_tlb_entry *gtlbe;
528         u64 pgmask;
529 
530         gtlbe = get_entry(vcpu_e500, tlbsel_of(index), esel_of(index));
531         pgmask = get_tlb_bytes(gtlbe) - 1;
532 
533         return get_tlb_raddr(gtlbe) | (eaddr & pgmask);
534 }
535 
536 /*****************************************/
537 
538 static void free_gtlb(struct kvmppc_vcpu_e500 *vcpu_e500)
539 {
540         int i;
541 
542         kvmppc_core_flush_tlb(&vcpu_e500->vcpu);
543         kfree(vcpu_e500->g2h_tlb1_map);
544         kfree(vcpu_e500->gtlb_priv[0]);
545         kfree(vcpu_e500->gtlb_priv[1]);
546 
547         if (vcpu_e500->shared_tlb_pages) {
548                 vfree((void *)(round_down((uintptr_t)vcpu_e500->gtlb_arch,
549                                           PAGE_SIZE)));
550 
551                 for (i = 0; i < vcpu_e500->num_shared_tlb_pages; i++) {
552                         set_page_dirty_lock(vcpu_e500->shared_tlb_pages[i]);
553                         put_page(vcpu_e500->shared_tlb_pages[i]);
554                 }
555 
556                 vcpu_e500->num_shared_tlb_pages = 0;
557 
558                 kfree(vcpu_e500->shared_tlb_pages);
559                 vcpu_e500->shared_tlb_pages = NULL;
560         } else {
561                 kfree(vcpu_e500->gtlb_arch);
562         }
563 
564         vcpu_e500->gtlb_arch = NULL;
565 }
566 
567 void kvmppc_get_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
568 {
569         sregs->u.e.mas0 = vcpu->arch.shared->mas0;
570         sregs->u.e.mas1 = vcpu->arch.shared->mas1;
571         sregs->u.e.mas2 = vcpu->arch.shared->mas2;
572         sregs->u.e.mas7_3 = vcpu->arch.shared->mas7_3;
573         sregs->u.e.mas4 = vcpu->arch.shared->mas4;
574         sregs->u.e.mas6 = vcpu->arch.shared->mas6;
575 
576         sregs->u.e.mmucfg = vcpu->arch.mmucfg;
577         sregs->u.e.tlbcfg[0] = vcpu->arch.tlbcfg[0];
578         sregs->u.e.tlbcfg[1] = vcpu->arch.tlbcfg[1];
579         sregs->u.e.tlbcfg[2] = 0;
580         sregs->u.e.tlbcfg[3] = 0;
581 }
582 
583 int kvmppc_set_sregs_e500_tlb(struct kvm_vcpu *vcpu, struct kvm_sregs *sregs)
584 {
585         if (sregs->u.e.features & KVM_SREGS_E_ARCH206_MMU) {
586                 vcpu->arch.shared->mas0 = sregs->u.e.mas0;
587                 vcpu->arch.shared->mas1 = sregs->u.e.mas1;
588                 vcpu->arch.shared->mas2 = sregs->u.e.mas2;
589                 vcpu->arch.shared->mas7_3 = sregs->u.e.mas7_3;
590                 vcpu->arch.shared->mas4 = sregs->u.e.mas4;
591                 vcpu->arch.shared->mas6 = sregs->u.e.mas6;
592         }
593 
594         return 0;
595 }
596 
597 int kvmppc_get_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id,
598                                 union kvmppc_one_reg *val)
599 {
600         int r = 0;
601         long int i;
602 
603         switch (id) {
604         case KVM_REG_PPC_MAS0:
605                 *val = get_reg_val(id, vcpu->arch.shared->mas0);
606                 break;
607         case KVM_REG_PPC_MAS1:
608                 *val = get_reg_val(id, vcpu->arch.shared->mas1);
609                 break;
610         case KVM_REG_PPC_MAS2:
611                 *val = get_reg_val(id, vcpu->arch.shared->mas2);
612                 break;
613         case KVM_REG_PPC_MAS7_3:
614                 *val = get_reg_val(id, vcpu->arch.shared->mas7_3);
615                 break;
616         case KVM_REG_PPC_MAS4:
617                 *val = get_reg_val(id, vcpu->arch.shared->mas4);
618                 break;
619         case KVM_REG_PPC_MAS6:
620                 *val = get_reg_val(id, vcpu->arch.shared->mas6);
621                 break;
622         case KVM_REG_PPC_MMUCFG:
623                 *val = get_reg_val(id, vcpu->arch.mmucfg);
624                 break;
625         case KVM_REG_PPC_EPTCFG:
626                 *val = get_reg_val(id, vcpu->arch.eptcfg);
627                 break;
628         case KVM_REG_PPC_TLB0CFG:
629         case KVM_REG_PPC_TLB1CFG:
630         case KVM_REG_PPC_TLB2CFG:
631         case KVM_REG_PPC_TLB3CFG:
632                 i = id - KVM_REG_PPC_TLB0CFG;
633                 *val = get_reg_val(id, vcpu->arch.tlbcfg[i]);
634                 break;
635         case KVM_REG_PPC_TLB0PS:
636         case KVM_REG_PPC_TLB1PS:
637         case KVM_REG_PPC_TLB2PS:
638         case KVM_REG_PPC_TLB3PS:
639                 i = id - KVM_REG_PPC_TLB0PS;
640                 *val = get_reg_val(id, vcpu->arch.tlbps[i]);
641                 break;
642         default:
643                 r = -EINVAL;
644                 break;
645         }
646 
647         return r;
648 }
649 
650 int kvmppc_set_one_reg_e500_tlb(struct kvm_vcpu *vcpu, u64 id,
651                                union kvmppc_one_reg *val)
652 {
653         int r = 0;
654         long int i;
655 
656         switch (id) {
657         case KVM_REG_PPC_MAS0:
658                 vcpu->arch.shared->mas0 = set_reg_val(id, *val);
659                 break;
660         case KVM_REG_PPC_MAS1:
661                 vcpu->arch.shared->mas1 = set_reg_val(id, *val);
662                 break;
663         case KVM_REG_PPC_MAS2:
664                 vcpu->arch.shared->mas2 = set_reg_val(id, *val);
665                 break;
666         case KVM_REG_PPC_MAS7_3:
667                 vcpu->arch.shared->mas7_3 = set_reg_val(id, *val);
668                 break;
669         case KVM_REG_PPC_MAS4:
670                 vcpu->arch.shared->mas4 = set_reg_val(id, *val);
671                 break;
672         case KVM_REG_PPC_MAS6:
673                 vcpu->arch.shared->mas6 = set_reg_val(id, *val);
674                 break;
675         /* Only allow MMU registers to be set to the config supported by KVM */
676         case KVM_REG_PPC_MMUCFG: {
677                 u32 reg = set_reg_val(id, *val);
678                 if (reg != vcpu->arch.mmucfg)
679                         r = -EINVAL;
680                 break;
681         }
682         case KVM_REG_PPC_EPTCFG: {
683                 u32 reg = set_reg_val(id, *val);
684                 if (reg != vcpu->arch.eptcfg)
685                         r = -EINVAL;
686                 break;
687         }
688         case KVM_REG_PPC_TLB0CFG:
689         case KVM_REG_PPC_TLB1CFG:
690         case KVM_REG_PPC_TLB2CFG:
691         case KVM_REG_PPC_TLB3CFG: {
692                 /* MMU geometry (N_ENTRY/ASSOC) can be set only using SW_TLB */
693                 u32 reg = set_reg_val(id, *val);
694                 i = id - KVM_REG_PPC_TLB0CFG;
695                 if (reg != vcpu->arch.tlbcfg[i])
696                         r = -EINVAL;
697                 break;
698         }
699         case KVM_REG_PPC_TLB0PS:
700         case KVM_REG_PPC_TLB1PS:
701         case KVM_REG_PPC_TLB2PS:
702         case KVM_REG_PPC_TLB3PS: {
703                 u32 reg = set_reg_val(id, *val);
704                 i = id - KVM_REG_PPC_TLB0PS;
705                 if (reg != vcpu->arch.tlbps[i])
706                         r = -EINVAL;
707                 break;
708         }
709         default:
710                 r = -EINVAL;
711                 break;
712         }
713 
714         return r;
715 }
716 
717 static int vcpu_mmu_geometry_update(struct kvm_vcpu *vcpu,
718                 struct kvm_book3e_206_tlb_params *params)
719 {
720         vcpu->arch.tlbcfg[0] &= ~(TLBnCFG_N_ENTRY | TLBnCFG_ASSOC);
721         if (params->tlb_sizes[0] <= 2048)
722                 vcpu->arch.tlbcfg[0] |= params->tlb_sizes[0];
723         vcpu->arch.tlbcfg[0] |= params->tlb_ways[0] << TLBnCFG_ASSOC_SHIFT;
724 
725         vcpu->arch.tlbcfg[1] &= ~(TLBnCFG_N_ENTRY | TLBnCFG_ASSOC);
726         vcpu->arch.tlbcfg[1] |= params->tlb_sizes[1];
727         vcpu->arch.tlbcfg[1] |= params->tlb_ways[1] << TLBnCFG_ASSOC_SHIFT;
728         return 0;
729 }
730 
731 int kvm_vcpu_ioctl_config_tlb(struct kvm_vcpu *vcpu,
732                               struct kvm_config_tlb *cfg)
733 {
734         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
735         struct kvm_book3e_206_tlb_params params;
736         char *virt;
737         struct page **pages;
738         struct tlbe_priv *privs[2] = {};
739         u64 *g2h_bitmap;
740         size_t array_len;
741         u32 sets;
742         int num_pages, ret, i;
743 
744         if (cfg->mmu_type != KVM_MMU_FSL_BOOKE_NOHV)
745                 return -EINVAL;
746 
747         if (copy_from_user(&params, (void __user *)(uintptr_t)cfg->params,
748                            sizeof(params)))
749                 return -EFAULT;
750 
751         if (params.tlb_sizes[1] > 64)
752                 return -EINVAL;
753         if (params.tlb_ways[1] != params.tlb_sizes[1])
754                 return -EINVAL;
755         if (params.tlb_sizes[2] != 0 || params.tlb_sizes[3] != 0)
756                 return -EINVAL;
757         if (params.tlb_ways[2] != 0 || params.tlb_ways[3] != 0)
758                 return -EINVAL;
759 
760         if (!is_power_of_2(params.tlb_ways[0]))
761                 return -EINVAL;
762 
763         sets = params.tlb_sizes[0] >> ilog2(params.tlb_ways[0]);
764         if (!is_power_of_2(sets))
765                 return -EINVAL;
766 
767         array_len = params.tlb_sizes[0] + params.tlb_sizes[1];
768         array_len *= sizeof(struct kvm_book3e_206_tlb_entry);
769 
770         if (cfg->array_len < array_len)
771                 return -EINVAL;
772 
773         num_pages = DIV_ROUND_UP(cfg->array + array_len - 1, PAGE_SIZE) -
774                     cfg->array / PAGE_SIZE;
775         pages = kmalloc_array(num_pages, sizeof(*pages), GFP_KERNEL);
776         if (!pages)
777                 return -ENOMEM;
778 
779         ret = get_user_pages_fast(cfg->array, num_pages, FOLL_WRITE, pages);
780         if (ret < 0)
781                 goto free_pages;
782 
783         if (ret != num_pages) {
784                 num_pages = ret;
785                 ret = -EFAULT;
786                 goto put_pages;
787         }
788 
789         virt = vmap(pages, num_pages, VM_MAP, PAGE_KERNEL);
790         if (!virt) {
791                 ret = -ENOMEM;
792                 goto put_pages;
793         }
794 
795         privs[0] = kcalloc(params.tlb_sizes[0], sizeof(*privs[0]), GFP_KERNEL);
796         if (!privs[0]) {
797                 ret = -ENOMEM;
798                 goto put_pages;
799         }
800 
801         privs[1] = kcalloc(params.tlb_sizes[1], sizeof(*privs[1]), GFP_KERNEL);
802         if (!privs[1]) {
803                 ret = -ENOMEM;
804                 goto free_privs_first;
805         }
806 
807         g2h_bitmap = kcalloc(params.tlb_sizes[1],
808                              sizeof(*g2h_bitmap),
809                              GFP_KERNEL);
810         if (!g2h_bitmap) {
811                 ret = -ENOMEM;
812                 goto free_privs_second;
813         }
814 
815         free_gtlb(vcpu_e500);
816 
817         vcpu_e500->gtlb_priv[0] = privs[0];
818         vcpu_e500->gtlb_priv[1] = privs[1];
819         vcpu_e500->g2h_tlb1_map = g2h_bitmap;
820 
821         vcpu_e500->gtlb_arch = (struct kvm_book3e_206_tlb_entry *)
822                 (virt + (cfg->array & (PAGE_SIZE - 1)));
823 
824         vcpu_e500->gtlb_params[0].entries = params.tlb_sizes[0];
825         vcpu_e500->gtlb_params[1].entries = params.tlb_sizes[1];
826 
827         vcpu_e500->gtlb_offset[0] = 0;
828         vcpu_e500->gtlb_offset[1] = params.tlb_sizes[0];
829 
830         /* Update vcpu's MMU geometry based on SW_TLB input */
831         vcpu_mmu_geometry_update(vcpu, &params);
832 
833         vcpu_e500->shared_tlb_pages = pages;
834         vcpu_e500->num_shared_tlb_pages = num_pages;
835 
836         vcpu_e500->gtlb_params[0].ways = params.tlb_ways[0];
837         vcpu_e500->gtlb_params[0].sets = sets;
838 
839         vcpu_e500->gtlb_params[1].ways = params.tlb_sizes[1];
840         vcpu_e500->gtlb_params[1].sets = 1;
841 
842         kvmppc_recalc_tlb1map_range(vcpu_e500);
843         return 0;
844  free_privs_second:
845         kfree(privs[1]);
846  free_privs_first:
847         kfree(privs[0]);
848  put_pages:
849         for (i = 0; i < num_pages; i++)
850                 put_page(pages[i]);
851  free_pages:
852         kfree(pages);
853         return ret;
854 }
855 
856 int kvm_vcpu_ioctl_dirty_tlb(struct kvm_vcpu *vcpu,
857                              struct kvm_dirty_tlb *dirty)
858 {
859         struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
860         kvmppc_recalc_tlb1map_range(vcpu_e500);
861         kvmppc_core_flush_tlb(vcpu);
862         return 0;
863 }
864 
865 /* Vcpu's MMU default configuration */
866 static int vcpu_mmu_init(struct kvm_vcpu *vcpu,
867                        struct kvmppc_e500_tlb_params *params)
868 {
869         /* Initialize RASIZE, PIDSIZE, NTLBS and MAVN fields with host values*/
870         vcpu->arch.mmucfg = mfspr(SPRN_MMUCFG) & ~MMUCFG_LPIDSIZE;
871 
872         /* Initialize TLBnCFG fields with host values and SW_TLB geometry*/
873         vcpu->arch.tlbcfg[0] = mfspr(SPRN_TLB0CFG) &
874                              ~(TLBnCFG_N_ENTRY | TLBnCFG_ASSOC);
875         vcpu->arch.tlbcfg[0] |= params[0].entries;
876         vcpu->arch.tlbcfg[0] |= params[0].ways << TLBnCFG_ASSOC_SHIFT;
877 
878         vcpu->arch.tlbcfg[1] = mfspr(SPRN_TLB1CFG) &
879                              ~(TLBnCFG_N_ENTRY | TLBnCFG_ASSOC);
880         vcpu->arch.tlbcfg[1] |= params[1].entries;
881         vcpu->arch.tlbcfg[1] |= params[1].ways << TLBnCFG_ASSOC_SHIFT;
882 
883         if (has_feature(vcpu, VCPU_FTR_MMU_V2)) {
884                 vcpu->arch.tlbps[0] = mfspr(SPRN_TLB0PS);
885                 vcpu->arch.tlbps[1] = mfspr(SPRN_TLB1PS);
886 
887                 vcpu->arch.mmucfg &= ~MMUCFG_LRAT;
888 
889                 /* Guest mmu emulation currently doesn't handle E.PT */
890                 vcpu->arch.eptcfg = 0;
891                 vcpu->arch.tlbcfg[0] &= ~TLBnCFG_PT;
892                 vcpu->arch.tlbcfg[1] &= ~TLBnCFG_IND;
893         }
894 
895         return 0;
896 }
897 
898 int kvmppc_e500_tlb_init(struct kvmppc_vcpu_e500 *vcpu_e500)
899 {
900         struct kvm_vcpu *vcpu = &vcpu_e500->vcpu;
901 
902         if (e500_mmu_host_init(vcpu_e500))
903                 goto free_vcpu;
904 
905         vcpu_e500->gtlb_params[0].entries = KVM_E500_TLB0_SIZE;
906         vcpu_e500->gtlb_params[1].entries = KVM_E500_TLB1_SIZE;
907 
908         vcpu_e500->gtlb_params[0].ways = KVM_E500_TLB0_WAY_NUM;
909         vcpu_e500->gtlb_params[0].sets =
910                 KVM_E500_TLB0_SIZE / KVM_E500_TLB0_WAY_NUM;
911 
912         vcpu_e500->gtlb_params[1].ways = KVM_E500_TLB1_SIZE;
913         vcpu_e500->gtlb_params[1].sets = 1;
914 
915         vcpu_e500->gtlb_arch = kmalloc_array(KVM_E500_TLB0_SIZE +
916                                              KVM_E500_TLB1_SIZE,
917                                              sizeof(*vcpu_e500->gtlb_arch),
918                                              GFP_KERNEL);
919         if (!vcpu_e500->gtlb_arch)
920                 return -ENOMEM;
921 
922         vcpu_e500->gtlb_offset[0] = 0;
923         vcpu_e500->gtlb_offset[1] = KVM_E500_TLB0_SIZE;
924 
925         vcpu_e500->gtlb_priv[0] = kcalloc(vcpu_e500->gtlb_params[0].entries,
926                                           sizeof(struct tlbe_ref),
927                                           GFP_KERNEL);
928         if (!vcpu_e500->gtlb_priv[0])
929                 goto free_vcpu;
930 
931         vcpu_e500->gtlb_priv[1] = kcalloc(vcpu_e500->gtlb_params[1].entries,
932                                           sizeof(struct tlbe_ref),
933                                           GFP_KERNEL);
934         if (!vcpu_e500->gtlb_priv[1])
935                 goto free_vcpu;
936 
937         vcpu_e500->g2h_tlb1_map = kcalloc(vcpu_e500->gtlb_params[1].entries,
938                                           sizeof(*vcpu_e500->g2h_tlb1_map),
939                                           GFP_KERNEL);
940         if (!vcpu_e500->g2h_tlb1_map)
941                 goto free_vcpu;
942 
943         vcpu_mmu_init(vcpu, vcpu_e500->gtlb_params);
944 
945         kvmppc_recalc_tlb1map_range(vcpu_e500);
946         return 0;
947  free_vcpu:
948         free_gtlb(vcpu_e500);
949         return -1;
950 }
951 
952 void kvmppc_e500_tlb_uninit(struct kvmppc_vcpu_e500 *vcpu_e500)
953 {
954         free_gtlb(vcpu_e500);
955         e500_mmu_host_uninit(vcpu_e500);
956 }
957 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php