~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/mips/kvm/msa.S

Version: ~ [ linux-6.12-rc7 ] ~ [ linux-6.11.7 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.60 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.116 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.171 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.229 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.285 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.323 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.12 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

Diff markup

Differences between /arch/mips/kvm/msa.S (Architecture sparc64) and /arch/mips/kvm/msa.S (Architecture mips)


  1 /*                                                  1 /*
  2  * This file is subject to the terms and condi      2  * This file is subject to the terms and conditions of the GNU General Public
  3  * License.  See the file "COPYING" in the mai      3  * License.  See the file "COPYING" in the main directory of this archive
  4  * for more details.                                4  * for more details.
  5  *                                                  5  *
  6  * MIPS SIMD Architecture (MSA) context handli      6  * MIPS SIMD Architecture (MSA) context handling code for KVM.
  7  *                                                  7  *
  8  * Copyright (C) 2015 Imagination Technologies      8  * Copyright (C) 2015 Imagination Technologies Ltd.
  9  */                                                 9  */
 10                                                    10 
 11 #include <asm/asm.h>                               11 #include <asm/asm.h>
 12 #include <asm/asm-offsets.h>                       12 #include <asm/asm-offsets.h>
 13 #include <asm/asmmacro.h>                          13 #include <asm/asmmacro.h>
 14 #include <asm/regdef.h>                            14 #include <asm/regdef.h>
 15                                                    15 
 16         .set    noreorder                          16         .set    noreorder
 17         .set    noat                               17         .set    noat
 18                                                    18 
 19 LEAF(__kvm_save_msa)                               19 LEAF(__kvm_save_msa)
 20         st_d    0,  VCPU_FPR0,  a0                 20         st_d    0,  VCPU_FPR0,  a0
 21         st_d    1,  VCPU_FPR1,  a0                 21         st_d    1,  VCPU_FPR1,  a0
 22         st_d    2,  VCPU_FPR2,  a0                 22         st_d    2,  VCPU_FPR2,  a0
 23         st_d    3,  VCPU_FPR3,  a0                 23         st_d    3,  VCPU_FPR3,  a0
 24         st_d    4,  VCPU_FPR4,  a0                 24         st_d    4,  VCPU_FPR4,  a0
 25         st_d    5,  VCPU_FPR5,  a0                 25         st_d    5,  VCPU_FPR5,  a0
 26         st_d    6,  VCPU_FPR6,  a0                 26         st_d    6,  VCPU_FPR6,  a0
 27         st_d    7,  VCPU_FPR7,  a0                 27         st_d    7,  VCPU_FPR7,  a0
 28         st_d    8,  VCPU_FPR8,  a0                 28         st_d    8,  VCPU_FPR8,  a0
 29         st_d    9,  VCPU_FPR9,  a0                 29         st_d    9,  VCPU_FPR9,  a0
 30         st_d    10, VCPU_FPR10, a0                 30         st_d    10, VCPU_FPR10, a0
 31         st_d    11, VCPU_FPR11, a0                 31         st_d    11, VCPU_FPR11, a0
 32         st_d    12, VCPU_FPR12, a0                 32         st_d    12, VCPU_FPR12, a0
 33         st_d    13, VCPU_FPR13, a0                 33         st_d    13, VCPU_FPR13, a0
 34         st_d    14, VCPU_FPR14, a0                 34         st_d    14, VCPU_FPR14, a0
 35         st_d    15, VCPU_FPR15, a0                 35         st_d    15, VCPU_FPR15, a0
 36         st_d    16, VCPU_FPR16, a0                 36         st_d    16, VCPU_FPR16, a0
 37         st_d    17, VCPU_FPR17, a0                 37         st_d    17, VCPU_FPR17, a0
 38         st_d    18, VCPU_FPR18, a0                 38         st_d    18, VCPU_FPR18, a0
 39         st_d    19, VCPU_FPR19, a0                 39         st_d    19, VCPU_FPR19, a0
 40         st_d    20, VCPU_FPR20, a0                 40         st_d    20, VCPU_FPR20, a0
 41         st_d    21, VCPU_FPR21, a0                 41         st_d    21, VCPU_FPR21, a0
 42         st_d    22, VCPU_FPR22, a0                 42         st_d    22, VCPU_FPR22, a0
 43         st_d    23, VCPU_FPR23, a0                 43         st_d    23, VCPU_FPR23, a0
 44         st_d    24, VCPU_FPR24, a0                 44         st_d    24, VCPU_FPR24, a0
 45         st_d    25, VCPU_FPR25, a0                 45         st_d    25, VCPU_FPR25, a0
 46         st_d    26, VCPU_FPR26, a0                 46         st_d    26, VCPU_FPR26, a0
 47         st_d    27, VCPU_FPR27, a0                 47         st_d    27, VCPU_FPR27, a0
 48         st_d    28, VCPU_FPR28, a0                 48         st_d    28, VCPU_FPR28, a0
 49         st_d    29, VCPU_FPR29, a0                 49         st_d    29, VCPU_FPR29, a0
 50         st_d    30, VCPU_FPR30, a0                 50         st_d    30, VCPU_FPR30, a0
 51         st_d    31, VCPU_FPR31, a0                 51         st_d    31, VCPU_FPR31, a0
 52         jr      ra                                 52         jr      ra
 53          nop                                       53          nop
 54         END(__kvm_save_msa)                        54         END(__kvm_save_msa)
 55                                                    55 
 56 LEAF(__kvm_restore_msa)                            56 LEAF(__kvm_restore_msa)
 57         ld_d    0,  VCPU_FPR0,  a0                 57         ld_d    0,  VCPU_FPR0,  a0
 58         ld_d    1,  VCPU_FPR1,  a0                 58         ld_d    1,  VCPU_FPR1,  a0
 59         ld_d    2,  VCPU_FPR2,  a0                 59         ld_d    2,  VCPU_FPR2,  a0
 60         ld_d    3,  VCPU_FPR3,  a0                 60         ld_d    3,  VCPU_FPR3,  a0
 61         ld_d    4,  VCPU_FPR4,  a0                 61         ld_d    4,  VCPU_FPR4,  a0
 62         ld_d    5,  VCPU_FPR5,  a0                 62         ld_d    5,  VCPU_FPR5,  a0
 63         ld_d    6,  VCPU_FPR6,  a0                 63         ld_d    6,  VCPU_FPR6,  a0
 64         ld_d    7,  VCPU_FPR7,  a0                 64         ld_d    7,  VCPU_FPR7,  a0
 65         ld_d    8,  VCPU_FPR8,  a0                 65         ld_d    8,  VCPU_FPR8,  a0
 66         ld_d    9,  VCPU_FPR9,  a0                 66         ld_d    9,  VCPU_FPR9,  a0
 67         ld_d    10, VCPU_FPR10, a0                 67         ld_d    10, VCPU_FPR10, a0
 68         ld_d    11, VCPU_FPR11, a0                 68         ld_d    11, VCPU_FPR11, a0
 69         ld_d    12, VCPU_FPR12, a0                 69         ld_d    12, VCPU_FPR12, a0
 70         ld_d    13, VCPU_FPR13, a0                 70         ld_d    13, VCPU_FPR13, a0
 71         ld_d    14, VCPU_FPR14, a0                 71         ld_d    14, VCPU_FPR14, a0
 72         ld_d    15, VCPU_FPR15, a0                 72         ld_d    15, VCPU_FPR15, a0
 73         ld_d    16, VCPU_FPR16, a0                 73         ld_d    16, VCPU_FPR16, a0
 74         ld_d    17, VCPU_FPR17, a0                 74         ld_d    17, VCPU_FPR17, a0
 75         ld_d    18, VCPU_FPR18, a0                 75         ld_d    18, VCPU_FPR18, a0
 76         ld_d    19, VCPU_FPR19, a0                 76         ld_d    19, VCPU_FPR19, a0
 77         ld_d    20, VCPU_FPR20, a0                 77         ld_d    20, VCPU_FPR20, a0
 78         ld_d    21, VCPU_FPR21, a0                 78         ld_d    21, VCPU_FPR21, a0
 79         ld_d    22, VCPU_FPR22, a0                 79         ld_d    22, VCPU_FPR22, a0
 80         ld_d    23, VCPU_FPR23, a0                 80         ld_d    23, VCPU_FPR23, a0
 81         ld_d    24, VCPU_FPR24, a0                 81         ld_d    24, VCPU_FPR24, a0
 82         ld_d    25, VCPU_FPR25, a0                 82         ld_d    25, VCPU_FPR25, a0
 83         ld_d    26, VCPU_FPR26, a0                 83         ld_d    26, VCPU_FPR26, a0
 84         ld_d    27, VCPU_FPR27, a0                 84         ld_d    27, VCPU_FPR27, a0
 85         ld_d    28, VCPU_FPR28, a0                 85         ld_d    28, VCPU_FPR28, a0
 86         ld_d    29, VCPU_FPR29, a0                 86         ld_d    29, VCPU_FPR29, a0
 87         ld_d    30, VCPU_FPR30, a0                 87         ld_d    30, VCPU_FPR30, a0
 88         ld_d    31, VCPU_FPR31, a0                 88         ld_d    31, VCPU_FPR31, a0
 89         jr      ra                                 89         jr      ra
 90          nop                                       90          nop
 91         END(__kvm_restore_msa)                     91         END(__kvm_restore_msa)
 92                                                    92 
 93         .macro  kvm_restore_msa_upper   wr, of     93         .macro  kvm_restore_msa_upper   wr, off, base
 94         .set    push                               94         .set    push
 95         .set    noat                               95         .set    noat
 96 #ifdef CONFIG_64BIT                                96 #ifdef CONFIG_64BIT
 97         ld      $1, \off(\base)                    97         ld      $1, \off(\base)
 98         insert_d \wr, 1                            98         insert_d \wr, 1
 99 #elif defined(CONFIG_CPU_LITTLE_ENDIAN)            99 #elif defined(CONFIG_CPU_LITTLE_ENDIAN)
100         lw      $1, \off(\base)                   100         lw      $1, \off(\base)
101         insert_w \wr, 2                           101         insert_w \wr, 2
102         lw      $1, (\off+4)(\base)               102         lw      $1, (\off+4)(\base)
103         insert_w \wr, 3                           103         insert_w \wr, 3
104 #else /* CONFIG_CPU_BIG_ENDIAN */                 104 #else /* CONFIG_CPU_BIG_ENDIAN */
105         lw      $1, (\off+4)(\base)               105         lw      $1, (\off+4)(\base)
106         insert_w \wr, 2                           106         insert_w \wr, 2
107         lw      $1, \off(\base)                   107         lw      $1, \off(\base)
108         insert_w \wr, 3                           108         insert_w \wr, 3
109 #endif                                            109 #endif
110         .set    pop                               110         .set    pop
111         .endm                                     111         .endm
112                                                   112 
113 LEAF(__kvm_restore_msa_upper)                     113 LEAF(__kvm_restore_msa_upper)
114         kvm_restore_msa_upper   0,  VCPU_FPR0     114         kvm_restore_msa_upper   0,  VCPU_FPR0 +8, a0
115         kvm_restore_msa_upper   1,  VCPU_FPR1     115         kvm_restore_msa_upper   1,  VCPU_FPR1 +8, a0
116         kvm_restore_msa_upper   2,  VCPU_FPR2     116         kvm_restore_msa_upper   2,  VCPU_FPR2 +8, a0
117         kvm_restore_msa_upper   3,  VCPU_FPR3     117         kvm_restore_msa_upper   3,  VCPU_FPR3 +8, a0
118         kvm_restore_msa_upper   4,  VCPU_FPR4     118         kvm_restore_msa_upper   4,  VCPU_FPR4 +8, a0
119         kvm_restore_msa_upper   5,  VCPU_FPR5     119         kvm_restore_msa_upper   5,  VCPU_FPR5 +8, a0
120         kvm_restore_msa_upper   6,  VCPU_FPR6     120         kvm_restore_msa_upper   6,  VCPU_FPR6 +8, a0
121         kvm_restore_msa_upper   7,  VCPU_FPR7     121         kvm_restore_msa_upper   7,  VCPU_FPR7 +8, a0
122         kvm_restore_msa_upper   8,  VCPU_FPR8     122         kvm_restore_msa_upper   8,  VCPU_FPR8 +8, a0
123         kvm_restore_msa_upper   9,  VCPU_FPR9     123         kvm_restore_msa_upper   9,  VCPU_FPR9 +8, a0
124         kvm_restore_msa_upper   10, VCPU_FPR10    124         kvm_restore_msa_upper   10, VCPU_FPR10+8, a0
125         kvm_restore_msa_upper   11, VCPU_FPR11    125         kvm_restore_msa_upper   11, VCPU_FPR11+8, a0
126         kvm_restore_msa_upper   12, VCPU_FPR12    126         kvm_restore_msa_upper   12, VCPU_FPR12+8, a0
127         kvm_restore_msa_upper   13, VCPU_FPR13    127         kvm_restore_msa_upper   13, VCPU_FPR13+8, a0
128         kvm_restore_msa_upper   14, VCPU_FPR14    128         kvm_restore_msa_upper   14, VCPU_FPR14+8, a0
129         kvm_restore_msa_upper   15, VCPU_FPR15    129         kvm_restore_msa_upper   15, VCPU_FPR15+8, a0
130         kvm_restore_msa_upper   16, VCPU_FPR16    130         kvm_restore_msa_upper   16, VCPU_FPR16+8, a0
131         kvm_restore_msa_upper   17, VCPU_FPR17    131         kvm_restore_msa_upper   17, VCPU_FPR17+8, a0
132         kvm_restore_msa_upper   18, VCPU_FPR18    132         kvm_restore_msa_upper   18, VCPU_FPR18+8, a0
133         kvm_restore_msa_upper   19, VCPU_FPR19    133         kvm_restore_msa_upper   19, VCPU_FPR19+8, a0
134         kvm_restore_msa_upper   20, VCPU_FPR20    134         kvm_restore_msa_upper   20, VCPU_FPR20+8, a0
135         kvm_restore_msa_upper   21, VCPU_FPR21    135         kvm_restore_msa_upper   21, VCPU_FPR21+8, a0
136         kvm_restore_msa_upper   22, VCPU_FPR22    136         kvm_restore_msa_upper   22, VCPU_FPR22+8, a0
137         kvm_restore_msa_upper   23, VCPU_FPR23    137         kvm_restore_msa_upper   23, VCPU_FPR23+8, a0
138         kvm_restore_msa_upper   24, VCPU_FPR24    138         kvm_restore_msa_upper   24, VCPU_FPR24+8, a0
139         kvm_restore_msa_upper   25, VCPU_FPR25    139         kvm_restore_msa_upper   25, VCPU_FPR25+8, a0
140         kvm_restore_msa_upper   26, VCPU_FPR26    140         kvm_restore_msa_upper   26, VCPU_FPR26+8, a0
141         kvm_restore_msa_upper   27, VCPU_FPR27    141         kvm_restore_msa_upper   27, VCPU_FPR27+8, a0
142         kvm_restore_msa_upper   28, VCPU_FPR28    142         kvm_restore_msa_upper   28, VCPU_FPR28+8, a0
143         kvm_restore_msa_upper   29, VCPU_FPR29    143         kvm_restore_msa_upper   29, VCPU_FPR29+8, a0
144         kvm_restore_msa_upper   30, VCPU_FPR30    144         kvm_restore_msa_upper   30, VCPU_FPR30+8, a0
145         kvm_restore_msa_upper   31, VCPU_FPR31    145         kvm_restore_msa_upper   31, VCPU_FPR31+8, a0
146         jr      ra                                146         jr      ra
147          nop                                      147          nop
148         END(__kvm_restore_msa_upper)              148         END(__kvm_restore_msa_upper)
149                                                   149 
150 LEAF(__kvm_restore_msacsr)                        150 LEAF(__kvm_restore_msacsr)
151         lw      t0, VCPU_MSA_CSR(a0)              151         lw      t0, VCPU_MSA_CSR(a0)
152         /*                                        152         /*
153          * The ctcmsa must stay at this offset    153          * The ctcmsa must stay at this offset in __kvm_restore_msacsr.
154          * See kvm_mips_csr_die_notify() which    154          * See kvm_mips_csr_die_notify() which handles t0 containing a value
155          * which triggers an MSA FP Exception,    155          * which triggers an MSA FP Exception, which must be stepped over and
156          * ignored since the set cause bits mu    156          * ignored since the set cause bits must remain there for the guest.
157          */                                       157          */
158         _ctcmsa MSA_CSR, t0                       158         _ctcmsa MSA_CSR, t0
159         jr      ra                                159         jr      ra
160          nop                                      160          nop
161         END(__kvm_restore_msacsr)                 161         END(__kvm_restore_msacsr)
                                                      

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php