~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/riscv/kernel/entry.S

Version: ~ [ linux-6.12-rc7 ] ~ [ linux-6.11.7 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.60 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.116 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.171 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.229 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.285 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.323 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.12 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

Diff markup

Differences between /arch/riscv/kernel/entry.S (Version linux-6.12-rc7) and /arch/sparc/kernel/entry.S (Version linux-4.4.302)


  1 /* SPDX-License-Identifier: GPL-2.0-only */    !!   1 /* arch/sparc/kernel/entry.S:  Sparc trap low-level entry points.
  2 /*                                             !!   2  *
  3  * Copyright (C) 2012 Regents of the Universit !!   3  * Copyright (C) 1995, 2007 David S. Miller (davem@davemloft.net)
  4  * Copyright (C) 2017 SiFive                   !!   4  * Copyright (C) 1996 Eddie C. Dost   (ecd@skynet.be)
                                                   >>   5  * Copyright (C) 1996 Miguel de Icaza (miguel@nuclecu.unam.mx)
                                                   >>   6  * Copyright (C) 1996-1999 Jakub Jelinek   (jj@sunsite.mff.cuni.cz)
                                                   >>   7  * Copyright (C) 1997 Anton Blanchard (anton@progsoc.uts.edu.au)
  5  */                                                 8  */
  6                                                     9 
  7 #include <linux/init.h>                        << 
  8 #include <linux/linkage.h>                         10 #include <linux/linkage.h>
                                                   >>  11 #include <linux/errno.h>
  9                                                    12 
 10 #include <asm/asm.h>                           !!  13 #include <asm/head.h>
 11 #include <asm/csr.h>                           !!  14 #include <asm/asi.h>
 12 #include <asm/scs.h>                           !!  15 #include <asm/smp.h>
 13 #include <asm/unistd.h>                        !!  16 #include <asm/contregs.h>
                                                   >>  17 #include <asm/ptrace.h>
                                                   >>  18 #include <asm/asm-offsets.h>
                                                   >>  19 #include <asm/psr.h>
                                                   >>  20 #include <asm/vaddrs.h>
 14 #include <asm/page.h>                              21 #include <asm/page.h>
                                                   >>  22 #include <asm/pgtable.h>
                                                   >>  23 #include <asm/winmacro.h>
                                                   >>  24 #include <asm/signal.h>
                                                   >>  25 #include <asm/obio.h>
                                                   >>  26 #include <asm/mxcc.h>
 15 #include <asm/thread_info.h>                       27 #include <asm/thread_info.h>
 16 #include <asm/asm-offsets.h>                   !!  28 #include <asm/param.h>
 17 #include <asm/errata_list.h>                   !!  29 #include <asm/unistd.h>
 18 #include <linux/sizes.h>                       !!  30 
                                                   >>  31 #include <asm/asmmacro.h>
 19                                                    32 
 20         .section .irqentry.text, "ax"          !!  33 #define curptr      g6
 21                                                    34 
 22 .macro new_vmalloc_check                       !!  35 /* These are just handy. */
 23         REG_S   a0, TASK_TI_A0(tp)             !!  36 #define _SV     save    %sp, -STACKFRAME_SZ, %sp
 24         csrr    a0, CSR_CAUSE                  !!  37 #define _RS     restore 
 25         /* Exclude IRQs */                     !!  38 
 26         blt     a0, zero, _new_vmalloc_restore !!  39 #define FLUSH_ALL_KERNEL_WINDOWS \
 27                                                !!  40         _SV; _SV; _SV; _SV; _SV; _SV; _SV; \
 28         REG_S   a1, TASK_TI_A1(tp)             !!  41         _RS; _RS; _RS; _RS; _RS; _RS; _RS;
 29         /* Only check new_vmalloc if we are in !!  42 
 30         li      a1, EXC_LOAD_PAGE_FAULT        !!  43         .text
 31         beq     a0, a1, _new_vmalloc_kernel_ad !!  44 
 32         li      a1, EXC_STORE_PAGE_FAULT       !!  45 #ifdef CONFIG_KGDB
 33         beq     a0, a1, _new_vmalloc_kernel_ad !!  46         .align  4
 34         li      a1, EXC_INST_PAGE_FAULT        !!  47         .globl          arch_kgdb_breakpoint
 35         bne     a0, a1, _new_vmalloc_restore_c !!  48         .type           arch_kgdb_breakpoint,#function
 36                                                !!  49 arch_kgdb_breakpoint:
 37 _new_vmalloc_kernel_address:                   !!  50         ta              0x7d
 38         /* Is it a kernel address? */          !!  51         retl
 39         csrr    a0, CSR_TVAL                   !!  52          nop
 40         bge     a0, zero, _new_vmalloc_restore !!  53         .size           arch_kgdb_breakpoint,.-arch_kgdb_breakpoint
                                                   >>  54 #endif
 41                                                    55 
 42         /* Check if a new vmalloc mapping appe !!  56 #if defined(CONFIG_BLK_DEV_FD) || defined(CONFIG_BLK_DEV_FD_MODULE)
 43         REG_S   a2, TASK_TI_A2(tp)             !!  57         .align  4
                                                   >>  58         .globl  floppy_hardint
                                                   >>  59 floppy_hardint:
 44         /*                                         60         /*
 45          * Computes:                           !!  61          * This code cannot touch registers %l0 %l1 and %l2
 46          * a0 = &new_vmalloc[BIT_WORD(cpu)]    !!  62          * because SAVE_ALL depends on their values. It depends
 47          * a1 = BIT_MASK(cpu)                  !!  63          * on %l3 also, but we regenerate it before a call.
                                                   >>  64          * Other registers are:
                                                   >>  65          * %l3 -- base address of fdc registers
                                                   >>  66          * %l4 -- pdma_vaddr
                                                   >>  67          * %l5 -- scratch for ld/st address
                                                   >>  68          * %l6 -- pdma_size
                                                   >>  69          * %l7 -- scratch [floppy byte, ld/st address, aux. data]
 48          */                                        70          */
 49         REG_L   a2, TASK_TI_CPU(tp)            !!  71 
 50         /*                                     !!  72         /* Do we have work to do? */
 51          * Compute the new_vmalloc element pos !!  73         sethi   %hi(doing_pdma), %l7
 52          * (cpu / 64) * 8 = (cpu >> 6) << 3    !!  74         ld      [%l7 + %lo(doing_pdma)], %l7
                                                   >>  75         cmp     %l7, 0
                                                   >>  76         be      floppy_dosoftint
                                                   >>  77          nop
                                                   >>  78 
                                                   >>  79         /* Load fdc register base */
                                                   >>  80         sethi   %hi(fdc_status), %l3
                                                   >>  81         ld      [%l3 + %lo(fdc_status)], %l3
                                                   >>  82 
                                                   >>  83         /* Setup register addresses */
                                                   >>  84         sethi   %hi(pdma_vaddr), %l5    ! transfer buffer
                                                   >>  85         ld      [%l5 + %lo(pdma_vaddr)], %l4
                                                   >>  86         sethi   %hi(pdma_size), %l5     ! bytes to go
                                                   >>  87         ld      [%l5 + %lo(pdma_size)], %l6
                                                   >>  88 next_byte:
                                                   >>  89         ldub    [%l3], %l7
                                                   >>  90 
                                                   >>  91         andcc   %l7, 0x80, %g0          ! Does fifo still have data
                                                   >>  92         bz      floppy_fifo_emptied     ! fifo has been emptied...
                                                   >>  93          andcc  %l7, 0x20, %g0          ! in non-dma mode still?
                                                   >>  94         bz      floppy_overrun          ! nope, overrun
                                                   >>  95          andcc  %l7, 0x40, %g0          ! 0=write 1=read
                                                   >>  96         bz      floppy_write
                                                   >>  97          sub    %l6, 0x1, %l6
                                                   >>  98 
                                                   >>  99         /* Ok, actually read this byte */
                                                   >> 100         ldub    [%l3 + 1], %l7
                                                   >> 101         orcc    %g0, %l6, %g0
                                                   >> 102         stb     %l7, [%l4]
                                                   >> 103         bne     next_byte
                                                   >> 104          add    %l4, 0x1, %l4
                                                   >> 105 
                                                   >> 106         b       floppy_tdone
                                                   >> 107          nop
                                                   >> 108 
                                                   >> 109 floppy_write:
                                                   >> 110         /* Ok, actually write this byte */
                                                   >> 111         ldub    [%l4], %l7
                                                   >> 112         orcc    %g0, %l6, %g0
                                                   >> 113         stb     %l7, [%l3 + 1]
                                                   >> 114         bne     next_byte
                                                   >> 115          add    %l4, 0x1, %l4
                                                   >> 116 
                                                   >> 117         /* fall through... */
                                                   >> 118 floppy_tdone:
                                                   >> 119         sethi   %hi(pdma_vaddr), %l5
                                                   >> 120         st      %l4, [%l5 + %lo(pdma_vaddr)]
                                                   >> 121         sethi   %hi(pdma_size), %l5
                                                   >> 122         st      %l6, [%l5 + %lo(pdma_size)]
                                                   >> 123         /* Flip terminal count pin */
                                                   >> 124         set     auxio_register, %l7
                                                   >> 125         ld      [%l7], %l7
                                                   >> 126 
                                                   >> 127         ldub    [%l7], %l5
                                                   >> 128 
                                                   >> 129         or      %l5, 0xc2, %l5
                                                   >> 130         stb     %l5, [%l7]
                                                   >> 131         andn    %l5, 0x02, %l5
                                                   >> 132 
                                                   >> 133 2:
                                                   >> 134         /* Kill some time so the bits set */
                                                   >> 135         WRITE_PAUSE
                                                   >> 136         WRITE_PAUSE
                                                   >> 137 
                                                   >> 138         stb     %l5, [%l7]
                                                   >> 139 
                                                   >> 140         /* Prevent recursion */
                                                   >> 141         sethi   %hi(doing_pdma), %l7
                                                   >> 142         b       floppy_dosoftint
                                                   >> 143          st     %g0, [%l7 + %lo(doing_pdma)]
                                                   >> 144 
                                                   >> 145         /* We emptied the FIFO, but we haven't read everything
                                                   >> 146          * as of yet.  Store the current transfer address and
                                                   >> 147          * bytes left to read so we can continue when the next
                                                   >> 148          * fast IRQ comes in.
 53          */                                       149          */
 54         srli    a1, a2, 6                      !! 150 floppy_fifo_emptied:
 55         slli    a1, a1, 3                      !! 151         sethi   %hi(pdma_vaddr), %l5
 56         la      a0, new_vmalloc                !! 152         st      %l4, [%l5 + %lo(pdma_vaddr)]
 57         add     a0, a0, a1                     !! 153         sethi   %hi(pdma_size), %l7
 58         /*                                     !! 154         st      %l6, [%l7 + %lo(pdma_size)]
 59          * Compute the bit position in the new !! 155 
 60          * bit_pos = cpu % 64 = cpu - (cpu / 6 !! 156         /* Restore condition codes */
 61          *         = cpu - ((cpu >> 6) << 3) < !! 157         wr      %l0, 0x0, %psr
 62          */                                    !! 158         WRITE_PAUSE
 63         slli    a1, a1, 3                      !! 159 
 64         sub     a1, a2, a1                     !! 160         jmp     %l1
 65         /* Compute the "get mask": 1 << bit_po !! 161         rett    %l2
 66         li      a2, 1                          !! 162 
 67         sll     a1, a2, a1                     !! 163 floppy_overrun:
 68                                                !! 164         sethi   %hi(pdma_vaddr), %l5
 69         /* Check the value of new_vmalloc for  !! 165         st      %l4, [%l5 + %lo(pdma_vaddr)]
 70         REG_L   a2, 0(a0)                      !! 166         sethi   %hi(pdma_size), %l5
 71         and     a2, a2, a1                     !! 167         st      %l6, [%l5 + %lo(pdma_size)]
 72         beq     a2, zero, _new_vmalloc_restore !! 168         /* Prevent recursion */
 73                                                !! 169         sethi   %hi(doing_pdma), %l7
 74         /* Atomically reset the current cpu bi !! 170         st      %g0, [%l7 + %lo(doing_pdma)]
 75         amoxor.d        a0, a1, (a0)           !! 171 
 76                                                !! 172         /* fall through... */
 77         /* Only emit a sfence.vma if the uarch !! 173 floppy_dosoftint:
 78         ALTERNATIVE("sfence.vma", "nop", 0, RI !! 174         rd      %wim, %l3
 79                                                !! 175         SAVE_ALL
 80         REG_L   a0, TASK_TI_A0(tp)             !! 176 
 81         REG_L   a1, TASK_TI_A1(tp)             !! 177         /* Set all IRQs off. */
 82         REG_L   a2, TASK_TI_A2(tp)             !! 178         or      %l0, PSR_PIL, %l4
 83         csrw    CSR_SCRATCH, x0                !! 179         wr      %l4, 0x0, %psr
 84         sret                                   !! 180         WRITE_PAUSE
 85                                                !! 181         wr      %l4, PSR_ET, %psr
 86 _new_vmalloc_restore_context:                  !! 182         WRITE_PAUSE
 87         REG_L   a2, TASK_TI_A2(tp)             !! 183 
 88 _new_vmalloc_restore_context_a1:               !! 184         mov     11, %o0                 ! floppy irq level (unused anyway)
 89         REG_L   a1, TASK_TI_A1(tp)             !! 185         mov     %g0, %o1                ! devid is not used in fast interrupts
 90 _new_vmalloc_restore_context_a0:               !! 186         call    sparc_floppy_irq
 91         REG_L   a0, TASK_TI_A0(tp)             !! 187          add    %sp, STACKFRAME_SZ, %o2 ! struct pt_regs *regs
 92 .endm                                          !! 188 
                                                   >> 189         RESTORE_ALL
                                                   >> 190         
                                                   >> 191 #endif /* (CONFIG_BLK_DEV_FD) */
                                                   >> 192 
                                                   >> 193         /* Bad trap handler */
                                                   >> 194         .globl  bad_trap_handler
                                                   >> 195 bad_trap_handler:
                                                   >> 196         SAVE_ALL
                                                   >> 197 
                                                   >> 198         wr      %l0, PSR_ET, %psr
                                                   >> 199         WRITE_PAUSE
                                                   >> 200 
                                                   >> 201         add     %sp, STACKFRAME_SZ, %o0 ! pt_regs
                                                   >> 202         call    do_hw_interrupt
                                                   >> 203          mov    %l7, %o1                ! trap number
                                                   >> 204 
                                                   >> 205         RESTORE_ALL
                                                   >> 206         
                                                   >> 207 /* For now all IRQ's not registered get sent here. handler_irq() will
                                                   >> 208  * see if a routine is registered to handle this interrupt and if not
                                                   >> 209  * it will say so on the console.
                                                   >> 210  */
 93                                                   211 
                                                   >> 212         .align  4
                                                   >> 213         .globl  real_irq_entry, patch_handler_irq
                                                   >> 214 real_irq_entry:
                                                   >> 215         SAVE_ALL
                                                   >> 216 
                                                   >> 217 #ifdef CONFIG_SMP
                                                   >> 218         .globl  patchme_maybe_smp_msg
                                                   >> 219 
                                                   >> 220         cmp     %l7, 11
                                                   >> 221 patchme_maybe_smp_msg:
                                                   >> 222         bgu     maybe_smp4m_msg
                                                   >> 223          nop
                                                   >> 224 #endif
 94                                                   225 
 95 SYM_CODE_START(handle_exception)               !! 226 real_irq_continue:
 96         /*                                     !! 227         or      %l0, PSR_PIL, %g2
 97          * If coming from userspace, preserve  !! 228         wr      %g2, 0x0, %psr
 98          * the kernel thread pointer.  If we c !! 229         WRITE_PAUSE
 99          * register will contain 0, and we sho !! 230         wr      %g2, PSR_ET, %psr
                                                   >> 231         WRITE_PAUSE
                                                   >> 232         mov     %l7, %o0                ! irq level
                                                   >> 233 patch_handler_irq:
                                                   >> 234         call    handler_irq
                                                   >> 235          add    %sp, STACKFRAME_SZ, %o1 ! pt_regs ptr
                                                   >> 236         or      %l0, PSR_PIL, %g2       ! restore PIL after handler_irq
                                                   >> 237         wr      %g2, PSR_ET, %psr       ! keep ET up
                                                   >> 238         WRITE_PAUSE
                                                   >> 239 
                                                   >> 240         RESTORE_ALL
                                                   >> 241 
                                                   >> 242 #ifdef CONFIG_SMP
                                                   >> 243         /* SMP per-cpu ticker interrupts are handled specially. */
                                                   >> 244 smp4m_ticker:
                                                   >> 245         bne     real_irq_continue+4
                                                   >> 246          or     %l0, PSR_PIL, %g2
                                                   >> 247         wr      %g2, 0x0, %psr
                                                   >> 248         WRITE_PAUSE
                                                   >> 249         wr      %g2, PSR_ET, %psr
                                                   >> 250         WRITE_PAUSE
                                                   >> 251         call    smp4m_percpu_timer_interrupt
                                                   >> 252          add    %sp, STACKFRAME_SZ, %o0
                                                   >> 253         wr      %l0, PSR_ET, %psr
                                                   >> 254         WRITE_PAUSE
                                                   >> 255         RESTORE_ALL
                                                   >> 256 
                                                   >> 257 #define GET_PROCESSOR4M_ID(reg) \
                                                   >> 258         rd      %tbr, %reg;     \
                                                   >> 259         srl     %reg, 12, %reg; \
                                                   >> 260         and     %reg, 3, %reg;
                                                   >> 261 
                                                   >> 262         /* Here is where we check for possible SMP IPI passed to us
                                                   >> 263          * on some level other than 15 which is the NMI and only used
                                                   >> 264          * for cross calls.  That has a separate entry point below.
                                                   >> 265          *
                                                   >> 266          * IPIs are sent on Level 12, 13 and 14. See IRQ_IPI_*.
100          */                                       267          */
101         csrrw tp, CSR_SCRATCH, tp              !! 268 maybe_smp4m_msg:
102         bnez tp, .Lsave_context                !! 269         GET_PROCESSOR4M_ID(o3)
                                                   >> 270         sethi   %hi(sun4m_irq_percpu), %l5
                                                   >> 271         sll     %o3, 2, %o3
                                                   >> 272         or      %l5, %lo(sun4m_irq_percpu), %o5
                                                   >> 273         sethi   %hi(0x70000000), %o2    ! Check all soft-IRQs
                                                   >> 274         ld      [%o5 + %o3], %o1
                                                   >> 275         ld      [%o1 + 0x00], %o3       ! sun4m_irq_percpu[cpu]->pending
                                                   >> 276         andcc   %o3, %o2, %g0
                                                   >> 277         be,a    smp4m_ticker
                                                   >> 278          cmp    %l7, 14
                                                   >> 279         /* Soft-IRQ IPI */
                                                   >> 280         st      %o2, [%o1 + 0x04]       ! sun4m_irq_percpu[cpu]->clear=0x70000000
                                                   >> 281         WRITE_PAUSE
                                                   >> 282         ld      [%o1 + 0x00], %g0       ! sun4m_irq_percpu[cpu]->pending
                                                   >> 283         WRITE_PAUSE
                                                   >> 284         or      %l0, PSR_PIL, %l4
                                                   >> 285         wr      %l4, 0x0, %psr
                                                   >> 286         WRITE_PAUSE
                                                   >> 287         wr      %l4, PSR_ET, %psr
                                                   >> 288         WRITE_PAUSE
                                                   >> 289         srl     %o3, 28, %o2            ! shift for simpler checks below
                                                   >> 290 maybe_smp4m_msg_check_single:
                                                   >> 291         andcc   %o2, 0x1, %g0
                                                   >> 292         beq,a   maybe_smp4m_msg_check_mask
                                                   >> 293          andcc  %o2, 0x2, %g0
                                                   >> 294         call    smp_call_function_single_interrupt
                                                   >> 295          nop
                                                   >> 296         andcc   %o2, 0x2, %g0
                                                   >> 297 maybe_smp4m_msg_check_mask:
                                                   >> 298         beq,a   maybe_smp4m_msg_check_resched
                                                   >> 299          andcc  %o2, 0x4, %g0
                                                   >> 300         call    smp_call_function_interrupt
                                                   >> 301          nop
                                                   >> 302         andcc   %o2, 0x4, %g0
                                                   >> 303 maybe_smp4m_msg_check_resched:
                                                   >> 304         /* rescheduling is done in RESTORE_ALL regardless, but incr stats */
                                                   >> 305         beq,a   maybe_smp4m_msg_out
                                                   >> 306          nop
                                                   >> 307         call    smp_resched_interrupt
                                                   >> 308          nop
                                                   >> 309 maybe_smp4m_msg_out:
                                                   >> 310         RESTORE_ALL
                                                   >> 311 
                                                   >> 312         .align  4
                                                   >> 313         .globl  linux_trap_ipi15_sun4m
                                                   >> 314 linux_trap_ipi15_sun4m:
                                                   >> 315         SAVE_ALL
                                                   >> 316         sethi   %hi(0x80000000), %o2
                                                   >> 317         GET_PROCESSOR4M_ID(o0)
                                                   >> 318         sethi   %hi(sun4m_irq_percpu), %l5
                                                   >> 319         or      %l5, %lo(sun4m_irq_percpu), %o5
                                                   >> 320         sll     %o0, 2, %o0
                                                   >> 321         ld      [%o5 + %o0], %o5
                                                   >> 322         ld      [%o5 + 0x00], %o3       ! sun4m_irq_percpu[cpu]->pending
                                                   >> 323         andcc   %o3, %o2, %g0
                                                   >> 324         be      sun4m_nmi_error         ! Must be an NMI async memory error
                                                   >> 325          st     %o2, [%o5 + 0x04]       ! sun4m_irq_percpu[cpu]->clear=0x80000000
                                                   >> 326         WRITE_PAUSE
                                                   >> 327         ld      [%o5 + 0x00], %g0       ! sun4m_irq_percpu[cpu]->pending
                                                   >> 328         WRITE_PAUSE
                                                   >> 329         or      %l0, PSR_PIL, %l4
                                                   >> 330         wr      %l4, 0x0, %psr
                                                   >> 331         WRITE_PAUSE
                                                   >> 332         wr      %l4, PSR_ET, %psr
                                                   >> 333         WRITE_PAUSE
                                                   >> 334         call    smp4m_cross_call_irq
                                                   >> 335          nop
                                                   >> 336         b       ret_trap_lockless_ipi
                                                   >> 337          clr    %l6
                                                   >> 338 
                                                   >> 339         .globl  smp4d_ticker
                                                   >> 340         /* SMP per-cpu ticker interrupts are handled specially. */
                                                   >> 341 smp4d_ticker:
                                                   >> 342         SAVE_ALL
                                                   >> 343         or      %l0, PSR_PIL, %g2
                                                   >> 344         sethi   %hi(CC_ICLR), %o0
                                                   >> 345         sethi   %hi(1 << 14), %o1
                                                   >> 346         or      %o0, %lo(CC_ICLR), %o0
                                                   >> 347         stha    %o1, [%o0] ASI_M_MXCC   /* Clear PIL 14 in MXCC's ICLR */
                                                   >> 348         wr      %g2, 0x0, %psr
                                                   >> 349         WRITE_PAUSE
                                                   >> 350         wr      %g2, PSR_ET, %psr
                                                   >> 351         WRITE_PAUSE
                                                   >> 352         call    smp4d_percpu_timer_interrupt
                                                   >> 353          add    %sp, STACKFRAME_SZ, %o0
                                                   >> 354         wr      %l0, PSR_ET, %psr
                                                   >> 355         WRITE_PAUSE
                                                   >> 356         RESTORE_ALL
                                                   >> 357 
                                                   >> 358         .align  4
                                                   >> 359         .globl  linux_trap_ipi15_sun4d
                                                   >> 360 linux_trap_ipi15_sun4d:
                                                   >> 361         SAVE_ALL
                                                   >> 362         sethi   %hi(CC_BASE), %o4
                                                   >> 363         sethi   %hi(MXCC_ERR_ME|MXCC_ERR_PEW|MXCC_ERR_ASE|MXCC_ERR_PEE), %o2
                                                   >> 364         or      %o4, (CC_EREG - CC_BASE), %o0
                                                   >> 365         ldda    [%o0] ASI_M_MXCC, %o0
                                                   >> 366         andcc   %o0, %o2, %g0
                                                   >> 367         bne     1f
                                                   >> 368          sethi  %hi(BB_STAT2), %o2
                                                   >> 369         lduba   [%o2] ASI_M_CTL, %o2
                                                   >> 370         andcc   %o2, BB_STAT2_MASK, %g0
                                                   >> 371         bne     2f
                                                   >> 372          or     %o4, (CC_ICLR - CC_BASE), %o0
                                                   >> 373         sethi   %hi(1 << 15), %o1
                                                   >> 374         stha    %o1, [%o0] ASI_M_MXCC   /* Clear PIL 15 in MXCC's ICLR */
                                                   >> 375         or      %l0, PSR_PIL, %l4
                                                   >> 376         wr      %l4, 0x0, %psr
                                                   >> 377         WRITE_PAUSE
                                                   >> 378         wr      %l4, PSR_ET, %psr
                                                   >> 379         WRITE_PAUSE
                                                   >> 380         call    smp4d_cross_call_irq
                                                   >> 381          nop
                                                   >> 382         b       ret_trap_lockless_ipi
                                                   >> 383          clr    %l6
                                                   >> 384 
                                                   >> 385 1:      /* MXCC error */
                                                   >> 386 2:      /* BB error */
                                                   >> 387         /* Disable PIL 15 */
                                                   >> 388         set     CC_IMSK, %l4
                                                   >> 389         lduha   [%l4] ASI_M_MXCC, %l5
                                                   >> 390         sethi   %hi(1 << 15), %l7
                                                   >> 391         or      %l5, %l7, %l5
                                                   >> 392         stha    %l5, [%l4] ASI_M_MXCC
                                                   >> 393         /* FIXME */
                                                   >> 394 1:      b,a     1b
                                                   >> 395 
                                                   >> 396         .globl  smpleon_ipi
                                                   >> 397         .extern leon_ipi_interrupt
                                                   >> 398         /* SMP per-cpu IPI interrupts are handled specially. */
                                                   >> 399 smpleon_ipi:
                                                   >> 400         SAVE_ALL
                                                   >> 401         or      %l0, PSR_PIL, %g2
                                                   >> 402         wr      %g2, 0x0, %psr
                                                   >> 403         WRITE_PAUSE
                                                   >> 404         wr      %g2, PSR_ET, %psr
                                                   >> 405         WRITE_PAUSE
                                                   >> 406         call    leonsmp_ipi_interrupt
                                                   >> 407          add    %sp, STACKFRAME_SZ, %o1 ! pt_regs
                                                   >> 408         wr      %l0, PSR_ET, %psr
                                                   >> 409         WRITE_PAUSE
                                                   >> 410         RESTORE_ALL
                                                   >> 411 
                                                   >> 412         .align  4
                                                   >> 413         .globl  linux_trap_ipi15_leon
                                                   >> 414 linux_trap_ipi15_leon:
                                                   >> 415         SAVE_ALL
                                                   >> 416         or      %l0, PSR_PIL, %l4
                                                   >> 417         wr      %l4, 0x0, %psr
                                                   >> 418         WRITE_PAUSE
                                                   >> 419         wr      %l4, PSR_ET, %psr
                                                   >> 420         WRITE_PAUSE
                                                   >> 421         call    leon_cross_call_irq
                                                   >> 422          nop
                                                   >> 423         b       ret_trap_lockless_ipi
                                                   >> 424          clr    %l6
103                                                   425 
104 .Lrestore_kernel_tpsp:                         !! 426 #endif /* CONFIG_SMP */
105         csrr tp, CSR_SCRATCH                   << 
106                                                   427 
107 #ifdef CONFIG_64BIT                            !! 428         /* This routine handles illegal instructions and privileged
108         /*                                     !! 429          * instruction attempts from user code.
109          * The RISC-V kernel does not eagerly  << 
110          * new vmalloc mapping, which may resu << 
111          * - if the uarch caches invalid entri << 
112          *   observed by the page table walker << 
113          * - if the uarch does not cache inval << 
114          *   could "miss" the new mapping and  << 
115          *   to retry the access, no sfence.vm << 
116          */                                       430          */
117         new_vmalloc_check                      !! 431         .align  4
118 #endif                                         !! 432         .globl  bad_instruction
                                                   >> 433 bad_instruction:
                                                   >> 434         sethi   %hi(0xc1f80000), %l4
                                                   >> 435         ld      [%l1], %l5
                                                   >> 436         sethi   %hi(0x81d80000), %l7
                                                   >> 437         and     %l5, %l4, %l5
                                                   >> 438         cmp     %l5, %l7
                                                   >> 439         be      1f
                                                   >> 440         SAVE_ALL
                                                   >> 441 
                                                   >> 442         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 443         WRITE_PAUSE
                                                   >> 444 
                                                   >> 445         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 446         mov     %l1, %o1
                                                   >> 447         mov     %l2, %o2
                                                   >> 448         call    do_illegal_instruction
                                                   >> 449          mov    %l0, %o3
                                                   >> 450 
                                                   >> 451         RESTORE_ALL
                                                   >> 452 
                                                   >> 453 1:      /* unimplemented flush - just skip */
                                                   >> 454         jmpl    %l2, %g0
                                                   >> 455          rett   %l2 + 4
                                                   >> 456 
                                                   >> 457         .align  4
                                                   >> 458         .globl  priv_instruction
                                                   >> 459 priv_instruction:
                                                   >> 460         SAVE_ALL
                                                   >> 461 
                                                   >> 462         wr      %l0, PSR_ET, %psr
                                                   >> 463         WRITE_PAUSE
                                                   >> 464 
                                                   >> 465         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 466         mov     %l1, %o1
                                                   >> 467         mov     %l2, %o2
                                                   >> 468         call    do_priv_instruction
                                                   >> 469          mov    %l0, %o3
                                                   >> 470 
                                                   >> 471         RESTORE_ALL
                                                   >> 472 
                                                   >> 473         /* This routine handles unaligned data accesses. */
                                                   >> 474         .align  4
                                                   >> 475         .globl  mna_handler
                                                   >> 476 mna_handler:
                                                   >> 477         andcc   %l0, PSR_PS, %g0
                                                   >> 478         be      mna_fromuser
                                                   >> 479          nop
                                                   >> 480 
                                                   >> 481         SAVE_ALL
                                                   >> 482 
                                                   >> 483         wr      %l0, PSR_ET, %psr
                                                   >> 484         WRITE_PAUSE
                                                   >> 485 
                                                   >> 486         ld      [%l1], %o1
                                                   >> 487         call    kernel_unaligned_trap
                                                   >> 488          add    %sp, STACKFRAME_SZ, %o0
                                                   >> 489 
                                                   >> 490         RESTORE_ALL
                                                   >> 491 
                                                   >> 492 mna_fromuser:
                                                   >> 493         SAVE_ALL
                                                   >> 494 
                                                   >> 495         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 496         WRITE_PAUSE
                                                   >> 497 
                                                   >> 498         ld      [%l1], %o1
                                                   >> 499         call    user_unaligned_trap
                                                   >> 500          add    %sp, STACKFRAME_SZ, %o0
                                                   >> 501 
                                                   >> 502         RESTORE_ALL
                                                   >> 503 
                                                   >> 504         /* This routine handles floating point disabled traps. */
                                                   >> 505         .align  4
                                                   >> 506         .globl  fpd_trap_handler
                                                   >> 507 fpd_trap_handler:
                                                   >> 508         SAVE_ALL
                                                   >> 509 
                                                   >> 510         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 511         WRITE_PAUSE
                                                   >> 512 
                                                   >> 513         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 514         mov     %l1, %o1
                                                   >> 515         mov     %l2, %o2
                                                   >> 516         call    do_fpd_trap
                                                   >> 517          mov    %l0, %o3
                                                   >> 518 
                                                   >> 519         RESTORE_ALL
                                                   >> 520 
                                                   >> 521         /* This routine handles Floating Point Exceptions. */
                                                   >> 522         .align  4
                                                   >> 523         .globl  fpe_trap_handler
                                                   >> 524 fpe_trap_handler:
                                                   >> 525         set     fpsave_magic, %l5
                                                   >> 526         cmp     %l1, %l5
                                                   >> 527         be      1f
                                                   >> 528          sethi  %hi(fpsave), %l5
                                                   >> 529         or      %l5, %lo(fpsave), %l5
                                                   >> 530         cmp     %l1, %l5
                                                   >> 531         bne     2f
                                                   >> 532          sethi  %hi(fpsave_catch2), %l5
                                                   >> 533         or      %l5, %lo(fpsave_catch2), %l5
                                                   >> 534         wr      %l0, 0x0, %psr
                                                   >> 535         WRITE_PAUSE
                                                   >> 536         jmp     %l5
                                                   >> 537          rett   %l5 + 4
                                                   >> 538 1:      
                                                   >> 539         sethi   %hi(fpsave_catch), %l5
                                                   >> 540         or      %l5, %lo(fpsave_catch), %l5
                                                   >> 541         wr      %l0, 0x0, %psr
                                                   >> 542         WRITE_PAUSE
                                                   >> 543         jmp     %l5
                                                   >> 544          rett   %l5 + 4
                                                   >> 545 
                                                   >> 546 2:
                                                   >> 547         SAVE_ALL
                                                   >> 548 
                                                   >> 549         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 550         WRITE_PAUSE
                                                   >> 551 
                                                   >> 552         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 553         mov     %l1, %o1
                                                   >> 554         mov     %l2, %o2
                                                   >> 555         call    do_fpe_trap
                                                   >> 556          mov    %l0, %o3
                                                   >> 557 
                                                   >> 558         RESTORE_ALL
                                                   >> 559 
                                                   >> 560         /* This routine handles Tag Overflow Exceptions. */
                                                   >> 561         .align  4
                                                   >> 562         .globl  do_tag_overflow
                                                   >> 563 do_tag_overflow:
                                                   >> 564         SAVE_ALL
                                                   >> 565 
                                                   >> 566         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 567         WRITE_PAUSE
                                                   >> 568 
                                                   >> 569         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 570         mov     %l1, %o1
                                                   >> 571         mov     %l2, %o2
                                                   >> 572         call    handle_tag_overflow
                                                   >> 573          mov    %l0, %o3
                                                   >> 574 
                                                   >> 575         RESTORE_ALL
                                                   >> 576 
                                                   >> 577         /* This routine handles Watchpoint Exceptions. */
                                                   >> 578         .align  4
                                                   >> 579         .globl  do_watchpoint
                                                   >> 580 do_watchpoint:
                                                   >> 581         SAVE_ALL
                                                   >> 582 
                                                   >> 583         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 584         WRITE_PAUSE
                                                   >> 585 
                                                   >> 586         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 587         mov     %l1, %o1
                                                   >> 588         mov     %l2, %o2
                                                   >> 589         call    handle_watchpoint
                                                   >> 590          mov    %l0, %o3
                                                   >> 591 
                                                   >> 592         RESTORE_ALL
                                                   >> 593 
                                                   >> 594         /* This routine handles Register Access Exceptions. */
                                                   >> 595         .align  4
                                                   >> 596         .globl  do_reg_access
                                                   >> 597 do_reg_access:
                                                   >> 598         SAVE_ALL
                                                   >> 599 
                                                   >> 600         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 601         WRITE_PAUSE
                                                   >> 602 
                                                   >> 603         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 604         mov     %l1, %o1
                                                   >> 605         mov     %l2, %o2
                                                   >> 606         call    handle_reg_access
                                                   >> 607          mov    %l0, %o3
                                                   >> 608 
                                                   >> 609         RESTORE_ALL
                                                   >> 610 
                                                   >> 611         /* This routine handles Co-Processor Disabled Exceptions. */
                                                   >> 612         .align  4
                                                   >> 613         .globl  do_cp_disabled
                                                   >> 614 do_cp_disabled:
                                                   >> 615         SAVE_ALL
                                                   >> 616 
                                                   >> 617         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 618         WRITE_PAUSE
                                                   >> 619 
                                                   >> 620         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 621         mov     %l1, %o1
                                                   >> 622         mov     %l2, %o2
                                                   >> 623         call    handle_cp_disabled
                                                   >> 624          mov    %l0, %o3
                                                   >> 625 
                                                   >> 626         RESTORE_ALL
                                                   >> 627 
                                                   >> 628         /* This routine handles Co-Processor Exceptions. */
                                                   >> 629         .align  4
                                                   >> 630         .globl  do_cp_exception
                                                   >> 631 do_cp_exception:
                                                   >> 632         SAVE_ALL
                                                   >> 633 
                                                   >> 634         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 635         WRITE_PAUSE
                                                   >> 636 
                                                   >> 637         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 638         mov     %l1, %o1
                                                   >> 639         mov     %l2, %o2
                                                   >> 640         call    handle_cp_exception
                                                   >> 641          mov    %l0, %o3
                                                   >> 642 
                                                   >> 643         RESTORE_ALL
                                                   >> 644 
                                                   >> 645         /* This routine handles Hardware Divide By Zero Exceptions. */
                                                   >> 646         .align  4
                                                   >> 647         .globl  do_hw_divzero
                                                   >> 648 do_hw_divzero:
                                                   >> 649         SAVE_ALL
                                                   >> 650 
                                                   >> 651         wr      %l0, PSR_ET, %psr               ! re-enable traps
                                                   >> 652         WRITE_PAUSE
                                                   >> 653 
                                                   >> 654         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 655         mov     %l1, %o1
                                                   >> 656         mov     %l2, %o2
                                                   >> 657         call    handle_hw_divzero
                                                   >> 658          mov    %l0, %o3
                                                   >> 659 
                                                   >> 660         RESTORE_ALL
                                                   >> 661 
                                                   >> 662         .align  4
                                                   >> 663         .globl  do_flush_windows
                                                   >> 664 do_flush_windows:
                                                   >> 665         SAVE_ALL
                                                   >> 666 
                                                   >> 667         wr      %l0, PSR_ET, %psr
                                                   >> 668         WRITE_PAUSE
                                                   >> 669 
                                                   >> 670         andcc   %l0, PSR_PS, %g0
                                                   >> 671         bne     dfw_kernel
                                                   >> 672          nop
                                                   >> 673 
                                                   >> 674         call    flush_user_windows
                                                   >> 675          nop
                                                   >> 676 
                                                   >> 677         /* Advance over the trap instruction. */
                                                   >> 678         ld      [%sp + STACKFRAME_SZ + PT_NPC], %l1
                                                   >> 679         add     %l1, 0x4, %l2
                                                   >> 680         st      %l1, [%sp + STACKFRAME_SZ + PT_PC]
                                                   >> 681         st      %l2, [%sp + STACKFRAME_SZ + PT_NPC]
                                                   >> 682 
                                                   >> 683         RESTORE_ALL
                                                   >> 684 
                                                   >> 685         .globl  flush_patch_one
                                                   >> 686 
                                                   >> 687         /* We get these for debugging routines using __builtin_return_address() */
                                                   >> 688 dfw_kernel:
                                                   >> 689 flush_patch_one:
                                                   >> 690         FLUSH_ALL_KERNEL_WINDOWS
                                                   >> 691 
                                                   >> 692         /* Advance over the trap instruction. */
                                                   >> 693         ld      [%sp + STACKFRAME_SZ + PT_NPC], %l1
                                                   >> 694         add     %l1, 0x4, %l2
                                                   >> 695         st      %l1, [%sp + STACKFRAME_SZ + PT_PC]
                                                   >> 696         st      %l2, [%sp + STACKFRAME_SZ + PT_NPC]
119                                                   697 
120         REG_S sp, TASK_TI_KERNEL_SP(tp)        !! 698         RESTORE_ALL
121                                                   699 
122 #ifdef CONFIG_VMAP_STACK                       !! 700         /* The getcc software trap.  The user wants the condition codes from
123         addi sp, sp, -(PT_SIZE_ON_STACK)       !! 701          * the %psr in register %g1.
124         srli sp, sp, THREAD_SHIFT              !! 702          */
125         andi sp, sp, 0x1                       << 
126         bnez sp, handle_kernel_stack_overflow  << 
127         REG_L sp, TASK_TI_KERNEL_SP(tp)        << 
128 #endif                                         << 
129                                                   703 
130 .Lsave_context:                                !! 704         .align  4
131         REG_S sp, TASK_TI_USER_SP(tp)          !! 705         .globl  getcc_trap_handler
132         REG_L sp, TASK_TI_KERNEL_SP(tp)        !! 706 getcc_trap_handler:
133         addi sp, sp, -(PT_SIZE_ON_STACK)       !! 707         srl     %l0, 20, %g1    ! give user
134         REG_S x1,  PT_RA(sp)                   !! 708         and     %g1, 0xf, %g1   ! only ICC bits in %psr
135         REG_S x3,  PT_GP(sp)                   !! 709         jmp     %l2             ! advance over trap instruction
136         REG_S x5,  PT_T0(sp)                   !! 710         rett    %l2 + 0x4       ! like this...
137         save_from_x6_to_x31                    !! 711 
                                                   >> 712         /* The setcc software trap.  The user has condition codes in %g1
                                                   >> 713          * that it would like placed in the %psr.  Be careful not to flip
                                                   >> 714          * any unintentional bits!
                                                   >> 715          */
138                                                   716 
139         /*                                     !! 717         .align  4
140          * Disable user-mode memory access as  !! 718         .globl  setcc_trap_handler
141          * actual user copy routines.          !! 719 setcc_trap_handler:
                                                   >> 720         sll     %g1, 0x14, %l4
                                                   >> 721         set     PSR_ICC, %l5
                                                   >> 722         andn    %l0, %l5, %l0   ! clear ICC bits in %psr
                                                   >> 723         and     %l4, %l5, %l4   ! clear non-ICC bits in user value
                                                   >> 724         or      %l4, %l0, %l4   ! or them in... mix mix mix
                                                   >> 725 
                                                   >> 726         wr      %l4, 0x0, %psr  ! set new %psr
                                                   >> 727         WRITE_PAUSE             ! TI scumbags...
                                                   >> 728 
                                                   >> 729         jmp     %l2             ! advance over trap instruction
                                                   >> 730         rett    %l2 + 0x4       ! like this...
                                                   >> 731 
                                                   >> 732 sun4m_nmi_error:
                                                   >> 733         /* NMI async memory error handling. */
                                                   >> 734         sethi   %hi(0x80000000), %l4
                                                   >> 735         sethi   %hi(sun4m_irq_global), %o5
                                                   >> 736         ld      [%o5 + %lo(sun4m_irq_global)], %l5
                                                   >> 737         st      %l4, [%l5 + 0x0c]       ! sun4m_irq_global->mask_set=0x80000000
                                                   >> 738         WRITE_PAUSE
                                                   >> 739         ld      [%l5 + 0x00], %g0       ! sun4m_irq_global->pending
                                                   >> 740         WRITE_PAUSE
                                                   >> 741         or      %l0, PSR_PIL, %l4
                                                   >> 742         wr      %l4, 0x0, %psr
                                                   >> 743         WRITE_PAUSE
                                                   >> 744         wr      %l4, PSR_ET, %psr
                                                   >> 745         WRITE_PAUSE
                                                   >> 746         call    sun4m_nmi
                                                   >> 747          nop
                                                   >> 748         st      %l4, [%l5 + 0x08]       ! sun4m_irq_global->mask_clear=0x80000000
                                                   >> 749         WRITE_PAUSE
                                                   >> 750         ld      [%l5 + 0x00], %g0       ! sun4m_irq_global->pending
                                                   >> 751         WRITE_PAUSE
                                                   >> 752         RESTORE_ALL
                                                   >> 753 
                                                   >> 754 #ifndef CONFIG_SMP
                                                   >> 755         .align  4
                                                   >> 756         .globl  linux_trap_ipi15_sun4m
                                                   >> 757 linux_trap_ipi15_sun4m:
                                                   >> 758         SAVE_ALL
                                                   >> 759 
                                                   >> 760         ba      sun4m_nmi_error
                                                   >> 761          nop
                                                   >> 762 #endif /* CONFIG_SMP */
                                                   >> 763 
                                                   >> 764         .align  4
                                                   >> 765         .globl  srmmu_fault
                                                   >> 766 srmmu_fault:
                                                   >> 767         mov     0x400, %l5
                                                   >> 768         mov     0x300, %l4
                                                   >> 769 
                                                   >> 770 LEON_PI(lda     [%l5] ASI_LEON_MMUREGS, %l6)    ! read sfar first
                                                   >> 771 SUN_PI_(lda     [%l5] ASI_M_MMUREGS, %l6)       ! read sfar first
                                                   >> 772 
                                                   >> 773 LEON_PI(lda     [%l4] ASI_LEON_MMUREGS, %l5)    ! read sfsr last
                                                   >> 774 SUN_PI_(lda     [%l4] ASI_M_MMUREGS, %l5)       ! read sfsr last
                                                   >> 775 
                                                   >> 776         andn    %l6, 0xfff, %l6
                                                   >> 777         srl     %l5, 6, %l5                     ! and encode all info into l7
                                                   >> 778 
                                                   >> 779         and     %l5, 2, %l5
                                                   >> 780         or      %l5, %l6, %l6
                                                   >> 781 
                                                   >> 782         or      %l6, %l7, %l7                   ! l7 = [addr,write,txtfault]
                                                   >> 783 
                                                   >> 784         SAVE_ALL
                                                   >> 785 
                                                   >> 786         mov     %l7, %o1
                                                   >> 787         mov     %l7, %o2
                                                   >> 788         and     %o1, 1, %o1             ! arg2 = text_faultp
                                                   >> 789         mov     %l7, %o3
                                                   >> 790         and     %o2, 2, %o2             ! arg3 = writep
                                                   >> 791         andn    %o3, 0xfff, %o3         ! arg4 = faulting address
                                                   >> 792 
                                                   >> 793         wr      %l0, PSR_ET, %psr
                                                   >> 794         WRITE_PAUSE
                                                   >> 795 
                                                   >> 796         call    do_sparc_fault
                                                   >> 797          add    %sp, STACKFRAME_SZ, %o0 ! arg1 = pt_regs ptr
                                                   >> 798 
                                                   >> 799         RESTORE_ALL
                                                   >> 800 
                                                   >> 801         .align  4
                                                   >> 802         .globl  sys_nis_syscall
                                                   >> 803 sys_nis_syscall:
                                                   >> 804         mov     %o7, %l5
                                                   >> 805         add     %sp, STACKFRAME_SZ, %o0         ! pt_regs *regs arg
                                                   >> 806         call    c_sys_nis_syscall
                                                   >> 807          mov    %l5, %o7
                                                   >> 808 
                                                   >> 809 sunos_execv:
                                                   >> 810         .globl  sunos_execv
                                                   >> 811         b       sys_execve
                                                   >> 812          clr    %i2
                                                   >> 813 
                                                   >> 814         .align  4
                                                   >> 815         .globl  sys_sparc_pipe
                                                   >> 816 sys_sparc_pipe:
                                                   >> 817         mov     %o7, %l5
                                                   >> 818         add     %sp, STACKFRAME_SZ, %o0         ! pt_regs *regs arg
                                                   >> 819         call    sparc_pipe
                                                   >> 820          mov    %l5, %o7
                                                   >> 821 
                                                   >> 822         .align  4
                                                   >> 823         .globl  sys_sigstack
                                                   >> 824 sys_sigstack:
                                                   >> 825         mov     %o7, %l5
                                                   >> 826         mov     %fp, %o2
                                                   >> 827         call    do_sys_sigstack
                                                   >> 828          mov    %l5, %o7
                                                   >> 829 
                                                   >> 830         .align  4
                                                   >> 831         .globl  sys_sigreturn
                                                   >> 832 sys_sigreturn:
                                                   >> 833         call    do_sigreturn
                                                   >> 834          add    %sp, STACKFRAME_SZ, %o0
                                                   >> 835 
                                                   >> 836         ld      [%curptr + TI_FLAGS], %l5
                                                   >> 837         andcc   %l5, _TIF_SYSCALL_TRACE, %g0
                                                   >> 838         be      1f
                                                   >> 839          nop
                                                   >> 840 
                                                   >> 841         call    syscall_trace
                                                   >> 842          mov    1, %o1
                                                   >> 843 
                                                   >> 844 1:
                                                   >> 845         /* We don't want to muck with user registers like a
                                                   >> 846          * normal syscall, just return.
                                                   >> 847          */
                                                   >> 848         RESTORE_ALL
                                                   >> 849 
                                                   >> 850         .align  4
                                                   >> 851         .globl  sys_rt_sigreturn
                                                   >> 852 sys_rt_sigreturn:
                                                   >> 853         call    do_rt_sigreturn
                                                   >> 854          add    %sp, STACKFRAME_SZ, %o0
                                                   >> 855 
                                                   >> 856         ld      [%curptr + TI_FLAGS], %l5
                                                   >> 857         andcc   %l5, _TIF_SYSCALL_TRACE, %g0
                                                   >> 858         be      1f
                                                   >> 859          nop
                                                   >> 860 
                                                   >> 861         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 862         call    syscall_trace
                                                   >> 863          mov    1, %o1
                                                   >> 864 
                                                   >> 865 1:
                                                   >> 866         /* We are returning to a signal handler. */
                                                   >> 867         RESTORE_ALL
                                                   >> 868 
                                                   >> 869         /* Now that we have a real sys_clone, sys_fork() is
                                                   >> 870          * implemented in terms of it.  Our _real_ implementation
                                                   >> 871          * of SunOS vfork() will use sys_vfork().
142          *                                        872          *
143          * Disable the FPU/Vector to detect il !! 873          * XXX These three should be consolidated into mostly shared
144          * or vector in kernel space.          !! 874          * XXX code just like on sparc64... -DaveM
145          */                                       875          */
146         li t0, SR_SUM | SR_FS_VS               !! 876         .align  4
                                                   >> 877         .globl  sys_fork, flush_patch_two
                                                   >> 878 sys_fork:
                                                   >> 879         mov     %o7, %l5
                                                   >> 880 flush_patch_two:
                                                   >> 881         FLUSH_ALL_KERNEL_WINDOWS;
                                                   >> 882         ld      [%curptr + TI_TASK], %o4
                                                   >> 883         rd      %psr, %g4
                                                   >> 884         WRITE_PAUSE
                                                   >> 885         mov     SIGCHLD, %o0                    ! arg0: clone flags
                                                   >> 886         rd      %wim, %g5
                                                   >> 887         WRITE_PAUSE
                                                   >> 888         mov     %fp, %o1                        ! arg1: usp
                                                   >> 889         std     %g4, [%o4 + AOFF_task_thread + AOFF_thread_fork_kpsr]
                                                   >> 890         add     %sp, STACKFRAME_SZ, %o2         ! arg2: pt_regs ptr
                                                   >> 891         mov     0, %o3
                                                   >> 892         call    sparc_do_fork
                                                   >> 893          mov    %l5, %o7
                                                   >> 894 
                                                   >> 895         /* Whee, kernel threads! */
                                                   >> 896         .globl  sys_clone, flush_patch_three
                                                   >> 897 sys_clone:
                                                   >> 898         mov     %o7, %l5
                                                   >> 899 flush_patch_three:
                                                   >> 900         FLUSH_ALL_KERNEL_WINDOWS;
                                                   >> 901         ld      [%curptr + TI_TASK], %o4
                                                   >> 902         rd      %psr, %g4
                                                   >> 903         WRITE_PAUSE
                                                   >> 904 
                                                   >> 905         /* arg0,1: flags,usp  -- loaded already */
                                                   >> 906         cmp     %o1, 0x0                        ! Is new_usp NULL?
                                                   >> 907         rd      %wim, %g5
                                                   >> 908         WRITE_PAUSE
                                                   >> 909         be,a    1f
                                                   >> 910          mov    %fp, %o1                        ! yes, use callers usp
                                                   >> 911         andn    %o1, 7, %o1                     ! no, align to 8 bytes
                                                   >> 912 1:
                                                   >> 913         std     %g4, [%o4 + AOFF_task_thread + AOFF_thread_fork_kpsr]
                                                   >> 914         add     %sp, STACKFRAME_SZ, %o2         ! arg2: pt_regs ptr
                                                   >> 915         mov     0, %o3
                                                   >> 916         call    sparc_do_fork
                                                   >> 917          mov    %l5, %o7
                                                   >> 918 
                                                   >> 919         /* Whee, real vfork! */
                                                   >> 920         .globl  sys_vfork, flush_patch_four
                                                   >> 921 sys_vfork:
                                                   >> 922 flush_patch_four:
                                                   >> 923         FLUSH_ALL_KERNEL_WINDOWS;
                                                   >> 924         ld      [%curptr + TI_TASK], %o4
                                                   >> 925         rd      %psr, %g4
                                                   >> 926         WRITE_PAUSE
                                                   >> 927         rd      %wim, %g5
                                                   >> 928         WRITE_PAUSE
                                                   >> 929         std     %g4, [%o4 + AOFF_task_thread + AOFF_thread_fork_kpsr]
                                                   >> 930         sethi   %hi(0x4000 | 0x0100 | SIGCHLD), %o0
                                                   >> 931         mov     %fp, %o1
                                                   >> 932         or      %o0, %lo(0x4000 | 0x0100 | SIGCHLD), %o0
                                                   >> 933         sethi   %hi(sparc_do_fork), %l1
                                                   >> 934         mov     0, %o3
                                                   >> 935         jmpl    %l1 + %lo(sparc_do_fork), %g0
                                                   >> 936          add    %sp, STACKFRAME_SZ, %o2
                                                   >> 937 
                                                   >> 938         .align  4
                                                   >> 939 linux_sparc_ni_syscall:
                                                   >> 940         sethi   %hi(sys_ni_syscall), %l7
                                                   >> 941         b       do_syscall
                                                   >> 942          or     %l7, %lo(sys_ni_syscall), %l7
                                                   >> 943 
                                                   >> 944 linux_syscall_trace:
                                                   >> 945         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 946         call    syscall_trace
                                                   >> 947          mov    0, %o1
                                                   >> 948         cmp     %o0, 0
                                                   >> 949         bne     3f
                                                   >> 950          mov    -ENOSYS, %o0
                                                   >> 951 
                                                   >> 952         /* Syscall tracing can modify the registers.  */
                                                   >> 953         ld      [%sp + STACKFRAME_SZ + PT_G1], %g1
                                                   >> 954         sethi   %hi(sys_call_table), %l7
                                                   >> 955         ld      [%sp + STACKFRAME_SZ + PT_I0], %i0
                                                   >> 956         or      %l7, %lo(sys_call_table), %l7
                                                   >> 957         ld      [%sp + STACKFRAME_SZ + PT_I1], %i1
                                                   >> 958         ld      [%sp + STACKFRAME_SZ + PT_I2], %i2
                                                   >> 959         ld      [%sp + STACKFRAME_SZ + PT_I3], %i3
                                                   >> 960         ld      [%sp + STACKFRAME_SZ + PT_I4], %i4
                                                   >> 961         ld      [%sp + STACKFRAME_SZ + PT_I5], %i5
                                                   >> 962         cmp     %g1, NR_syscalls
                                                   >> 963         bgeu    3f
                                                   >> 964          mov    -ENOSYS, %o0
                                                   >> 965 
                                                   >> 966         sll     %g1, 2, %l4
                                                   >> 967         mov     %i0, %o0
                                                   >> 968         ld      [%l7 + %l4], %l7
                                                   >> 969         mov     %i1, %o1
                                                   >> 970         mov     %i2, %o2
                                                   >> 971         mov     %i3, %o3
                                                   >> 972         b       2f
                                                   >> 973          mov    %i4, %o4
                                                   >> 974 
                                                   >> 975         .globl  ret_from_fork
                                                   >> 976 ret_from_fork:
                                                   >> 977         call    schedule_tail
                                                   >> 978          ld     [%g3 + TI_TASK], %o0
                                                   >> 979         b       ret_sys_call
                                                   >> 980          ld     [%sp + STACKFRAME_SZ + PT_I0], %o0
                                                   >> 981 
                                                   >> 982         .globl  ret_from_kernel_thread
                                                   >> 983 ret_from_kernel_thread:
                                                   >> 984         call    schedule_tail
                                                   >> 985          ld     [%g3 + TI_TASK], %o0
                                                   >> 986         ld      [%sp + STACKFRAME_SZ + PT_G1], %l0
                                                   >> 987         call    %l0
                                                   >> 988          ld     [%sp + STACKFRAME_SZ + PT_G2], %o0
                                                   >> 989         rd      %psr, %l1
                                                   >> 990         ld      [%sp + STACKFRAME_SZ + PT_PSR], %l0
                                                   >> 991         andn    %l0, PSR_CWP, %l0
                                                   >> 992         nop
                                                   >> 993         and     %l1, PSR_CWP, %l1
                                                   >> 994         or      %l0, %l1, %l0
                                                   >> 995         st      %l0, [%sp + STACKFRAME_SZ + PT_PSR]
                                                   >> 996         b       ret_sys_call
                                                   >> 997          mov    0, %o0
                                                   >> 998 
                                                   >> 999         /* Linux native system calls enter here... */
                                                   >> 1000         .align  4
                                                   >> 1001         .globl  linux_sparc_syscall
                                                   >> 1002 linux_sparc_syscall:
                                                   >> 1003         sethi   %hi(PSR_SYSCALL), %l4
                                                   >> 1004         or      %l0, %l4, %l0
                                                   >> 1005         /* Direct access to user regs, must faster. */
                                                   >> 1006         cmp     %g1, NR_syscalls
                                                   >> 1007         bgeu    linux_sparc_ni_syscall
                                                   >> 1008          sll    %g1, 2, %l4
                                                   >> 1009         ld      [%l7 + %l4], %l7
                                                   >> 1010 
                                                   >> 1011 do_syscall:
                                                   >> 1012         SAVE_ALL_HEAD
                                                   >> 1013          rd     %wim, %l3
                                                   >> 1014 
                                                   >> 1015         wr      %l0, PSR_ET, %psr
                                                   >> 1016         mov     %i0, %o0
                                                   >> 1017         mov     %i1, %o1
                                                   >> 1018         mov     %i2, %o2
                                                   >> 1019 
                                                   >> 1020         ld      [%curptr + TI_FLAGS], %l5
                                                   >> 1021         mov     %i3, %o3
                                                   >> 1022         andcc   %l5, _TIF_SYSCALL_TRACE, %g0
                                                   >> 1023         mov     %i4, %o4
                                                   >> 1024         bne     linux_syscall_trace
                                                   >> 1025          mov    %i0, %l5
                                                   >> 1026 2:
                                                   >> 1027         call    %l7
                                                   >> 1028          mov    %i5, %o5
147                                                   1029 
148         REG_L s0, TASK_TI_USER_SP(tp)          !! 1030 3:
149         csrrc s1, CSR_STATUS, t0               !! 1031         st      %o0, [%sp + STACKFRAME_SZ + PT_I0]
150         csrr s2, CSR_EPC                       << 
151         csrr s3, CSR_TVAL                      << 
152         csrr s4, CSR_CAUSE                     << 
153         csrr s5, CSR_SCRATCH                   << 
154         REG_S s0, PT_SP(sp)                    << 
155         REG_S s1, PT_STATUS(sp)                << 
156         REG_S s2, PT_EPC(sp)                   << 
157         REG_S s3, PT_BADADDR(sp)               << 
158         REG_S s4, PT_CAUSE(sp)                 << 
159         REG_S s5, PT_TP(sp)                    << 
160                                                   1032 
161         /*                                     !! 1033 ret_sys_call:
162          * Set the scratch register to 0, so t !! 1034         ld      [%curptr + TI_FLAGS], %l6
163          * occurs, the exception vector knows  !! 1035         cmp     %o0, -ERESTART_RESTARTBLOCK
                                                   >> 1036         ld      [%sp + STACKFRAME_SZ + PT_PSR], %g3
                                                   >> 1037         set     PSR_C, %g2
                                                   >> 1038         bgeu    1f
                                                   >> 1039          andcc  %l6, _TIF_SYSCALL_TRACE, %g0
                                                   >> 1040 
                                                   >> 1041         /* System call success, clear Carry condition code. */
                                                   >> 1042         andn    %g3, %g2, %g3
                                                   >> 1043         clr     %l6
                                                   >> 1044         st      %g3, [%sp + STACKFRAME_SZ + PT_PSR]     
                                                   >> 1045         bne     linux_syscall_trace2
                                                   >> 1046          ld     [%sp + STACKFRAME_SZ + PT_NPC], %l1 /* pc = npc */
                                                   >> 1047         add     %l1, 0x4, %l2                   /* npc = npc+4 */
                                                   >> 1048         st      %l1, [%sp + STACKFRAME_SZ + PT_PC]
                                                   >> 1049         b       ret_trap_entry
                                                   >> 1050          st     %l2, [%sp + STACKFRAME_SZ + PT_NPC]
                                                   >> 1051 1:
                                                   >> 1052         /* System call failure, set Carry condition code.
                                                   >> 1053          * Also, get abs(errno) to return to the process.
164          */                                       1054          */
165         csrw CSR_SCRATCH, x0                   !! 1055         sub     %g0, %o0, %o0
                                                   >> 1056         or      %g3, %g2, %g3
                                                   >> 1057         st      %o0, [%sp + STACKFRAME_SZ + PT_I0]
                                                   >> 1058         mov     1, %l6
                                                   >> 1059         st      %g3, [%sp + STACKFRAME_SZ + PT_PSR]
                                                   >> 1060         bne     linux_syscall_trace2
                                                   >> 1061          ld     [%sp + STACKFRAME_SZ + PT_NPC], %l1 /* pc = npc */
                                                   >> 1062         add     %l1, 0x4, %l2                   /* npc = npc+4 */
                                                   >> 1063         st      %l1, [%sp + STACKFRAME_SZ + PT_PC]
                                                   >> 1064         b       ret_trap_entry
                                                   >> 1065          st     %l2, [%sp + STACKFRAME_SZ + PT_NPC]
                                                   >> 1066 
                                                   >> 1067 linux_syscall_trace2:
                                                   >> 1068         add     %sp, STACKFRAME_SZ, %o0
                                                   >> 1069         mov     1, %o1
                                                   >> 1070         call    syscall_trace
                                                   >> 1071          add    %l1, 0x4, %l2                   /* npc = npc+4 */
                                                   >> 1072         st      %l1, [%sp + STACKFRAME_SZ + PT_PC]
                                                   >> 1073         b       ret_trap_entry
                                                   >> 1074          st     %l2, [%sp + STACKFRAME_SZ + PT_NPC]
166                                                   1075 
167         /* Load the global pointer */          << 
168         load_global_pointer                    << 
169                                                   1076 
170         /* Load the kernel shadow call stack p !! 1077 /* Saving and restoring the FPU state is best done from lowlevel code.
171         scs_load_current_if_task_changed s5    !! 1078  *
                                                   >> 1079  * void fpsave(unsigned long *fpregs, unsigned long *fsr,
                                                   >> 1080  *             void *fpqueue, unsigned long *fpqdepth)
                                                   >> 1081  */
172                                                   1082 
173 #ifdef CONFIG_RISCV_ISA_V_PREEMPTIVE           !! 1083         .globl  fpsave
174         move a0, sp                            !! 1084 fpsave:
175         call riscv_v_context_nesting_start     !! 1085         st      %fsr, [%o1]     ! this can trap on us if fpu is in bogon state
176 #endif                                         !! 1086         ld      [%o1], %g1
177         move a0, sp /* pt_regs */              !! 1087         set     0x2000, %g4
                                                   >> 1088         andcc   %g1, %g4, %g0
                                                   >> 1089         be      2f
                                                   >> 1090          mov    0, %g2
178                                                   1091 
179         /*                                     !! 1092         /* We have an fpqueue to save. */
180          * MSB of cause differentiates between !! 1093 1:
181          * interrupts and exceptions           !! 1094         std     %fq, [%o2]
                                                   >> 1095 fpsave_magic:
                                                   >> 1096         st      %fsr, [%o1]
                                                   >> 1097         ld      [%o1], %g3
                                                   >> 1098         andcc   %g3, %g4, %g0
                                                   >> 1099         add     %g2, 1, %g2
                                                   >> 1100         bne     1b
                                                   >> 1101          add    %o2, 8, %o2
                                                   >> 1102 
                                                   >> 1103 2:
                                                   >> 1104         st      %g2, [%o3]
                                                   >> 1105 
                                                   >> 1106         std     %f0, [%o0 + 0x00]
                                                   >> 1107         std     %f2, [%o0 + 0x08]
                                                   >> 1108         std     %f4, [%o0 + 0x10]
                                                   >> 1109         std     %f6, [%o0 + 0x18]
                                                   >> 1110         std     %f8, [%o0 + 0x20]
                                                   >> 1111         std     %f10, [%o0 + 0x28]
                                                   >> 1112         std     %f12, [%o0 + 0x30]
                                                   >> 1113         std     %f14, [%o0 + 0x38]
                                                   >> 1114         std     %f16, [%o0 + 0x40]
                                                   >> 1115         std     %f18, [%o0 + 0x48]
                                                   >> 1116         std     %f20, [%o0 + 0x50]
                                                   >> 1117         std     %f22, [%o0 + 0x58]
                                                   >> 1118         std     %f24, [%o0 + 0x60]
                                                   >> 1119         std     %f26, [%o0 + 0x68]
                                                   >> 1120         std     %f28, [%o0 + 0x70]
                                                   >> 1121         retl
                                                   >> 1122          std    %f30, [%o0 + 0x78]
                                                   >> 1123 
                                                   >> 1124         /* Thanks for Theo Deraadt and the authors of the Sprite/netbsd/openbsd
                                                   >> 1125          * code for pointing out this possible deadlock, while we save state
                                                   >> 1126          * above we could trap on the fsr store so our low level fpu trap
                                                   >> 1127          * code has to know how to deal with this.
                                                   >> 1128          */
                                                   >> 1129 fpsave_catch:
                                                   >> 1130         b       fpsave_magic + 4
                                                   >> 1131          st     %fsr, [%o1]
                                                   >> 1132 
                                                   >> 1133 fpsave_catch2:
                                                   >> 1134         b       fpsave + 4
                                                   >> 1135          st     %fsr, [%o1]
                                                   >> 1136 
                                                   >> 1137         /* void fpload(unsigned long *fpregs, unsigned long *fsr); */
                                                   >> 1138 
                                                   >> 1139         .globl  fpload
                                                   >> 1140 fpload:
                                                   >> 1141         ldd     [%o0 + 0x00], %f0
                                                   >> 1142         ldd     [%o0 + 0x08], %f2
                                                   >> 1143         ldd     [%o0 + 0x10], %f4
                                                   >> 1144         ldd     [%o0 + 0x18], %f6
                                                   >> 1145         ldd     [%o0 + 0x20], %f8
                                                   >> 1146         ldd     [%o0 + 0x28], %f10
                                                   >> 1147         ldd     [%o0 + 0x30], %f12
                                                   >> 1148         ldd     [%o0 + 0x38], %f14
                                                   >> 1149         ldd     [%o0 + 0x40], %f16
                                                   >> 1150         ldd     [%o0 + 0x48], %f18
                                                   >> 1151         ldd     [%o0 + 0x50], %f20
                                                   >> 1152         ldd     [%o0 + 0x58], %f22
                                                   >> 1153         ldd     [%o0 + 0x60], %f24
                                                   >> 1154         ldd     [%o0 + 0x68], %f26
                                                   >> 1155         ldd     [%o0 + 0x70], %f28
                                                   >> 1156         ldd     [%o0 + 0x78], %f30
                                                   >> 1157         ld      [%o1], %fsr
                                                   >> 1158         retl
                                                   >> 1159          nop
                                                   >> 1160 
                                                   >> 1161         /* __ndelay and __udelay take two arguments:
                                                   >> 1162          * 0 - nsecs or usecs to delay
                                                   >> 1163          * 1 - per_cpu udelay_val (loops per jiffy)
                                                   >> 1164          *
                                                   >> 1165          * Note that ndelay gives HZ times higher resolution but has a 10ms
                                                   >> 1166          * limit.  udelay can handle up to 1s.
182          */                                       1167          */
183         bge s4, zero, 1f                       !! 1168         .globl  __ndelay
                                                   >> 1169 __ndelay:
                                                   >> 1170         save    %sp, -STACKFRAME_SZ, %sp
                                                   >> 1171         mov     %i0, %o0                ! round multiplier up so large ns ok
                                                   >> 1172         mov     0x1ae, %o1              ! 2**32 / (1 000 000 000 / HZ)
                                                   >> 1173         umul    %o0, %o1, %o0
                                                   >> 1174         rd      %y, %o1
                                                   >> 1175         mov     %i1, %o1                ! udelay_val
                                                   >> 1176         umul    %o0, %o1, %o0
                                                   >> 1177         rd      %y, %o1
                                                   >> 1178         ba      delay_continue
                                                   >> 1179          mov    %o1, %o0                ! >>32 later for better resolution
                                                   >> 1180 
                                                   >> 1181         .globl  __udelay
                                                   >> 1182 __udelay:
                                                   >> 1183         save    %sp, -STACKFRAME_SZ, %sp
                                                   >> 1184         mov     %i0, %o0
                                                   >> 1185         sethi   %hi(0x10c7), %o1        ! round multiplier up so large us ok
                                                   >> 1186         or      %o1, %lo(0x10c7), %o1   ! 2**32 / 1 000 000
                                                   >> 1187         umul    %o0, %o1, %o0
                                                   >> 1188         rd      %y, %o1
                                                   >> 1189         mov     %i1, %o1                ! udelay_val
                                                   >> 1190         umul    %o0, %o1, %o0
                                                   >> 1191         rd      %y, %o1
                                                   >> 1192         sethi   %hi(0x028f4b62), %l0    ! Add in rounding constant * 2**32,
                                                   >> 1193         or      %g0, %lo(0x028f4b62), %l0
                                                   >> 1194         addcc   %o0, %l0, %o0           ! 2**32 * 0.009 999
                                                   >> 1195         bcs,a   3f
                                                   >> 1196          add    %o1, 0x01, %o1
                                                   >> 1197 3:
                                                   >> 1198         mov     HZ, %o0                 ! >>32 earlier for wider range
                                                   >> 1199         umul    %o0, %o1, %o0
                                                   >> 1200         rd      %y, %o1
184                                                   1201 
185         /* Handle interrupts */                !! 1202 delay_continue:
186         call do_irq                            !! 1203         cmp     %o0, 0x0
187         j ret_from_exception                   << 
188 1:                                                1204 1:
189         /* Handle other exceptions */          !! 1205         bne     1b
190         slli t0, s4, RISCV_LGPTR               !! 1206          subcc  %o0, 1, %o0
191         la t1, excp_vect_table                 !! 1207         
192         la t2, excp_vect_table_end             !! 1208         ret
193         add t0, t1, t0                         !! 1209         restore
194         /* Check if exception code lies within << 
195         bgeu t0, t2, 3f                        << 
196         REG_L t1, 0(t0)                        << 
197 2:      jalr t1                                << 
198         j ret_from_exception                   << 
199 3:                                             << 
200                                                   1210 
201         la t1, do_trap_unknown                 !! 1211         /* Handle a software breakpoint */
202         j 2b                                   !! 1212         /* We have to inform parent that child has stopped */
203 SYM_CODE_END(handle_exception)                 !! 1213         .align 4
204 ASM_NOKPROBE(handle_exception)                 !! 1214         .globl breakpoint_trap
205                                                !! 1215 breakpoint_trap:
206 /*                                             !! 1216         rd      %wim,%l3
207  * The ret_from_exception must be called with  !! 1217         SAVE_ALL
208  * caller list:                                !! 1218         wr      %l0, PSR_ET, %psr
209  *  - handle_exception                         !! 1219         WRITE_PAUSE
210  *  - ret_from_fork                            !! 1220 
211  */                                            !! 1221         st      %i0, [%sp + STACKFRAME_SZ + PT_G0] ! for restarting syscalls
212 SYM_CODE_START_NOALIGN(ret_from_exception)     !! 1222         call    sparc_breakpoint
213         REG_L s0, PT_STATUS(sp)                !! 1223          add    %sp, STACKFRAME_SZ, %o0
214 #ifdef CONFIG_RISCV_M_MODE                     !! 1224 
215         /* the MPP value is too large to be us !! 1225         RESTORE_ALL
216         li t0, SR_MPP                          !! 1226 
217         and s0, s0, t0                         !! 1227 #ifdef CONFIG_KGDB
218 #else                                          !! 1228         .align  4
219         andi s0, s0, SR_SPP                    !! 1229         .globl  kgdb_trap_low
220 #endif                                         !! 1230         .type   kgdb_trap_low,#function
221         bnez s0, 1f                            !! 1231 kgdb_trap_low:
                                                   >> 1232         rd      %wim,%l3
                                                   >> 1233         SAVE_ALL
                                                   >> 1234         wr      %l0, PSR_ET, %psr
                                                   >> 1235         WRITE_PAUSE
222                                                   1236 
223 #ifdef CONFIG_GCC_PLUGIN_STACKLEAK             !! 1237         call    kgdb_trap
224         call    stackleak_erase_on_task_stack  !! 1238          add    %sp, STACKFRAME_SZ, %o0
                                                   >> 1239 
                                                   >> 1240         RESTORE_ALL
                                                   >> 1241         .size   kgdb_trap_low,.-kgdb_trap_low
225 #endif                                            1242 #endif
226                                                   1243 
227         /* Save unwound kernel stack pointer i !! 1244         .align  4
228         addi s0, sp, PT_SIZE_ON_STACK          !! 1245         .globl  flush_patch_exception
229         REG_S s0, TASK_TI_KERNEL_SP(tp)        !! 1246 flush_patch_exception:
                                                   >> 1247         FLUSH_ALL_KERNEL_WINDOWS;
                                                   >> 1248         ldd     [%o0], %o6
                                                   >> 1249         jmpl    %o7 + 0xc, %g0                  ! see asm-sparc/processor.h
                                                   >> 1250          mov    1, %g1                          ! signal EFAULT condition
                                                   >> 1251 
                                                   >> 1252         .align  4
                                                   >> 1253         .globl  kill_user_windows, kuw_patch1_7win
                                                   >> 1254         .globl  kuw_patch1
                                                   >> 1255 kuw_patch1_7win:        sll     %o3, 6, %o3
                                                   >> 1256 
                                                   >> 1257         /* No matter how much overhead this routine has in the worst
                                                   >> 1258          * case scenerio, it is several times better than taking the
                                                   >> 1259          * traps with the old method of just doing flush_user_windows().
                                                   >> 1260          */
                                                   >> 1261 kill_user_windows:
                                                   >> 1262         ld      [%g6 + TI_UWINMASK], %o0        ! get current umask
                                                   >> 1263         orcc    %g0, %o0, %g0                   ! if no bits set, we are done
                                                   >> 1264         be      3f                              ! nothing to do
                                                   >> 1265          rd     %psr, %o5                       ! must clear interrupts
                                                   >> 1266         or      %o5, PSR_PIL, %o4               ! or else that could change
                                                   >> 1267         wr      %o4, 0x0, %psr                  ! the uwinmask state
                                                   >> 1268         WRITE_PAUSE                             ! burn them cycles
                                                   >> 1269 1:
                                                   >> 1270         ld      [%g6 + TI_UWINMASK], %o0        ! get consistent state
                                                   >> 1271         orcc    %g0, %o0, %g0                   ! did an interrupt come in?
                                                   >> 1272         be      4f                              ! yep, we are done
                                                   >> 1273          rd     %wim, %o3                       ! get current wim
                                                   >> 1274         srl     %o3, 1, %o4                     ! simulate a save
                                                   >> 1275 kuw_patch1:
                                                   >> 1276         sll     %o3, 7, %o3                     ! compute next wim
                                                   >> 1277         or      %o4, %o3, %o3                   ! result
                                                   >> 1278         andncc  %o0, %o3, %o0                   ! clean this bit in umask
                                                   >> 1279         bne     kuw_patch1                      ! not done yet
                                                   >> 1280          srl    %o3, 1, %o4                     ! begin another save simulation
                                                   >> 1281         wr      %o3, 0x0, %wim                  ! set the new wim
                                                   >> 1282         st      %g0, [%g6 + TI_UWINMASK]        ! clear uwinmask
                                                   >> 1283 4:
                                                   >> 1284         wr      %o5, 0x0, %psr                  ! re-enable interrupts
                                                   >> 1285         WRITE_PAUSE                             ! burn baby burn
                                                   >> 1286 3:
                                                   >> 1287         retl                                    ! return
                                                   >> 1288          st     %g0, [%g6 + TI_W_SAVED]         ! no windows saved
230                                                   1289 
231         /* Save the kernel shadow call stack p !! 1290         .align  4
232         scs_save_current                       !! 1291         .globl  restore_current
                                                   >> 1292 restore_current:
                                                   >> 1293         LOAD_CURRENT(g6, o0)
                                                   >> 1294         retl
                                                   >> 1295          nop
                                                   >> 1296 
                                                   >> 1297 #ifdef CONFIG_PCIC_PCI
                                                   >> 1298 #include <asm/pcic.h>
                                                   >> 1299 
                                                   >> 1300         .align  4
                                                   >> 1301         .globl  linux_trap_ipi15_pcic
                                                   >> 1302 linux_trap_ipi15_pcic:
                                                   >> 1303         rd      %wim, %l3
                                                   >> 1304         SAVE_ALL
233                                                   1305 
234         /*                                        1306         /*
235          * Save TP into the scratch register , !! 1307          * First deactivate NMI
236          * structures again.                   !! 1308          * or we cannot drop ET, cannot get window spill traps.
                                                   >> 1309          * The busy loop is necessary because the PIO error
                                                   >> 1310          * sometimes does not go away quickly and we trap again.
237          */                                       1311          */
238         csrw CSR_SCRATCH, tp                   !! 1312         sethi   %hi(pcic_regs), %o1
239 1:                                             !! 1313         ld      [%o1 + %lo(pcic_regs)], %o2
240 #ifdef CONFIG_RISCV_ISA_V_PREEMPTIVE           << 
241         move a0, sp                            << 
242         call riscv_v_context_nesting_end       << 
243 #endif                                         << 
244         REG_L a0, PT_STATUS(sp)                << 
245         /*                                     << 
246          * The current load reservation is eff << 
247          * state, in the sense that load reser << 
248          * different hart contexts.  We can't  << 
249          * reservation, so instead here we cle << 
250          * it's always legal for implementatio << 
251          * any point (as long as the forward p << 
252          * we'll ignore that here).            << 
253          *                                     << 
254          * Dangling load reservations can be t << 
255          * middle of an LR/SC sequence, but ca << 
256          * forward branch around an SC -- whic << 
257          * result we need to clear reservation << 
258          * jump back to the new context.  Whil << 
259          * completes, implementations are allo << 
260          * arbitrarily large.                  << 
261          */                                    << 
262         REG_L  a2, PT_EPC(sp)                  << 
263         REG_SC x0, a2, PT_EPC(sp)              << 
264                                                << 
265         csrw CSR_STATUS, a0                    << 
266         csrw CSR_EPC, a2                       << 
267                                                << 
268         REG_L x1,  PT_RA(sp)                   << 
269         REG_L x3,  PT_GP(sp)                   << 
270         REG_L x4,  PT_TP(sp)                   << 
271         REG_L x5,  PT_T0(sp)                   << 
272         restore_from_x6_to_x31                 << 
273                                                << 
274         REG_L x2,  PT_SP(sp)                   << 
275                                                << 
276 #ifdef CONFIG_RISCV_M_MODE                     << 
277         mret                                   << 
278 #else                                          << 
279         sret                                   << 
280 #endif                                         << 
281 SYM_CODE_END(ret_from_exception)               << 
282 ASM_NOKPROBE(ret_from_exception)               << 
283                                                   1314 
284 #ifdef CONFIG_VMAP_STACK                       !! 1315         ! Get pending status for printouts later.
285 SYM_CODE_START_LOCAL(handle_kernel_stack_overf !! 1316         ld      [%o2 + PCI_SYS_INT_PENDING], %o0
286         /* we reach here from kernel context,  << 
287         csrrw x31, CSR_SCRATCH, x31            << 
288         asm_per_cpu sp, overflow_stack, x31    << 
289         li x31, OVERFLOW_STACK_SIZE            << 
290         add sp, sp, x31                        << 
291         /* zero out x31 again and restore x31  << 
292         xor x31, x31, x31                      << 
293         csrrw x31, CSR_SCRATCH, x31            << 
294                                                << 
295         addi sp, sp, -(PT_SIZE_ON_STACK)       << 
296                                                << 
297         //save context to overflow stack       << 
298         REG_S x1,  PT_RA(sp)                   << 
299         REG_S x3,  PT_GP(sp)                   << 
300         REG_S x5,  PT_T0(sp)                   << 
301         save_from_x6_to_x31                    << 
302                                                << 
303         REG_L s0, TASK_TI_KERNEL_SP(tp)        << 
304         csrr s1, CSR_STATUS                    << 
305         csrr s2, CSR_EPC                       << 
306         csrr s3, CSR_TVAL                      << 
307         csrr s4, CSR_CAUSE                     << 
308         csrr s5, CSR_SCRATCH                   << 
309         REG_S s0, PT_SP(sp)                    << 
310         REG_S s1, PT_STATUS(sp)                << 
311         REG_S s2, PT_EPC(sp)                   << 
312         REG_S s3, PT_BADADDR(sp)               << 
313         REG_S s4, PT_CAUSE(sp)                 << 
314         REG_S s5, PT_TP(sp)                    << 
315         move a0, sp                            << 
316         tail handle_bad_stack                  << 
317 SYM_CODE_END(handle_kernel_stack_overflow)     << 
318 ASM_NOKPROBE(handle_kernel_stack_overflow)     << 
319 #endif                                         << 
320                                                   1317 
321 SYM_CODE_START(ret_from_fork)                  !! 1318         mov     PCI_SYS_INT_PENDING_CLEAR_ALL, %o1
322         call schedule_tail                     !! 1319         stb     %o1, [%o2 + PCI_SYS_INT_PENDING_CLEAR]
323         beqz s0, 1f     /* not from kernel thr << 
324         /* Call fn(arg) */                     << 
325         move a0, s1                            << 
326         jalr s0                                << 
327 1:                                                1320 1:
328         move a0, sp /* pt_regs */              !! 1321         ld      [%o2 + PCI_SYS_INT_PENDING], %o1
329         call syscall_exit_to_user_mode         !! 1322         andcc   %o1, ((PCI_SYS_INT_PENDING_PIO|PCI_SYS_INT_PENDING_PCI)>>24), %g0
330         j ret_from_exception                   !! 1323         bne     1b
331 SYM_CODE_END(ret_from_fork)                    !! 1324          nop
332                                                !! 1325 
333 #ifdef CONFIG_IRQ_STACKS                       !! 1326         or      %l0, PSR_PIL, %l4
334 /*                                             !! 1327         wr      %l4, 0x0, %psr
335  * void call_on_irq_stack(struct pt_regs *regs !! 1328         WRITE_PAUSE
336  *                        void (*func)(struct  !! 1329         wr      %l4, PSR_ET, %psr
337  *                                             !! 1330         WRITE_PAUSE
338  * Calls func(regs) using the per-CPU IRQ stac !! 1331 
339  */                                            !! 1332         call    pcic_nmi
340 SYM_FUNC_START(call_on_irq_stack)              !! 1333          add    %sp, STACKFRAME_SZ, %o1 ! struct pt_regs *regs
341         /* Create a frame record to save ra an !! 1334         RESTORE_ALL
342         addi    sp, sp, -STACKFRAME_SIZE_ON_ST !! 1335 
343         REG_S   ra, STACKFRAME_RA(sp)          !! 1336         .globl  pcic_nmi_trap_patch
344         REG_S   s0, STACKFRAME_FP(sp)          !! 1337 pcic_nmi_trap_patch:
345         addi    s0, sp, STACKFRAME_SIZE_ON_STA !! 1338         sethi   %hi(linux_trap_ipi15_pcic), %l3
346                                                !! 1339         jmpl    %l3 + %lo(linux_trap_ipi15_pcic), %g0
347         /* Switch to the per-CPU shadow call s !! 1340          rd     %psr, %l0
348         scs_save_current                       !! 1341         .word   0
349         scs_load_irq_stack t0                  !! 1342 
350                                                !! 1343 #endif /* CONFIG_PCIC_PCI */
351         /* Switch to the per-CPU IRQ stack and !! 1344 
352         load_per_cpu t0, irq_stack_ptr, t1     !! 1345         .globl  flushw_all
353         li      t1, IRQ_STACK_SIZE             !! 1346 flushw_all:
354         add     sp, t0, t1                     !! 1347         save    %sp, -0x40, %sp
355         jalr    a1                             !! 1348         save    %sp, -0x40, %sp
356                                                !! 1349         save    %sp, -0x40, %sp
357         /* Switch back to the thread shadow ca !! 1350         save    %sp, -0x40, %sp
358         scs_load_current                       !! 1351         save    %sp, -0x40, %sp
359                                                !! 1352         save    %sp, -0x40, %sp
360         /* Switch back to the thread stack and !! 1353         save    %sp, -0x40, %sp
361         addi    sp, s0, -STACKFRAME_SIZE_ON_ST !! 1354         restore
362         REG_L   ra, STACKFRAME_RA(sp)          !! 1355         restore
363         REG_L   s0, STACKFRAME_FP(sp)          !! 1356         restore
364         addi    sp, sp, STACKFRAME_SIZE_ON_STA !! 1357         restore
365                                                !! 1358         restore
                                                   >> 1359         restore
366         ret                                       1360         ret
367 SYM_FUNC_END(call_on_irq_stack)                !! 1361          restore
368 #endif /* CONFIG_IRQ_STACKS */                 << 
369                                                   1362 
370 /*                                             !! 1363 #ifdef CONFIG_SMP
371  * Integer register context switch             !! 1364 ENTRY(hard_smp_processor_id)
372  * The callee-saved registers must be saved an !! 1365 661:    rd              %tbr, %g1
373  *                                             !! 1366         srl             %g1, 12, %o0
374  *   a0: previous task_struct (must be preserv !! 1367         and             %o0, 3, %o0
375  *   a1: next task_struct                      !! 1368         .section        .cpuid_patch, "ax"
376  *                                             !! 1369         /* Instruction location. */
377  * The value of a0 and a1 must be preserved by !! 1370         .word           661b
378  * arguments are passed to schedule_tail.      !! 1371         /* SUN4D implementation. */
379  */                                            !! 1372         lda             [%g0] ASI_M_VIKING_TMP1, %o0
380 SYM_FUNC_START(__switch_to)                    !! 1373         nop
381         /* Save context into prev->thread */   !! 1374         nop
382         li    a4,  TASK_THREAD_RA              !! 1375         /* LEON implementation. */
383         add   a3, a0, a4                       !! 1376         rd              %asr17, %o0
384         add   a4, a1, a4                       !! 1377         srl             %o0, 0x1c, %o0
385         REG_S ra,  TASK_THREAD_RA_RA(a3)       !! 1378         nop
386         REG_S sp,  TASK_THREAD_SP_RA(a3)       !! 1379         .previous
387         REG_S s0,  TASK_THREAD_S0_RA(a3)       !! 1380         retl
388         REG_S s1,  TASK_THREAD_S1_RA(a3)       !! 1381          nop
389         REG_S s2,  TASK_THREAD_S2_RA(a3)       !! 1382 ENDPROC(hard_smp_processor_id)
390         REG_S s3,  TASK_THREAD_S3_RA(a3)       << 
391         REG_S s4,  TASK_THREAD_S4_RA(a3)       << 
392         REG_S s5,  TASK_THREAD_S5_RA(a3)       << 
393         REG_S s6,  TASK_THREAD_S6_RA(a3)       << 
394         REG_S s7,  TASK_THREAD_S7_RA(a3)       << 
395         REG_S s8,  TASK_THREAD_S8_RA(a3)       << 
396         REG_S s9,  TASK_THREAD_S9_RA(a3)       << 
397         REG_S s10, TASK_THREAD_S10_RA(a3)      << 
398         REG_S s11, TASK_THREAD_S11_RA(a3)      << 
399         /* Save the kernel shadow call stack p << 
400         scs_save_current                       << 
401         /* Restore context from next->thread * << 
402         REG_L ra,  TASK_THREAD_RA_RA(a4)       << 
403         REG_L sp,  TASK_THREAD_SP_RA(a4)       << 
404         REG_L s0,  TASK_THREAD_S0_RA(a4)       << 
405         REG_L s1,  TASK_THREAD_S1_RA(a4)       << 
406         REG_L s2,  TASK_THREAD_S2_RA(a4)       << 
407         REG_L s3,  TASK_THREAD_S3_RA(a4)       << 
408         REG_L s4,  TASK_THREAD_S4_RA(a4)       << 
409         REG_L s5,  TASK_THREAD_S5_RA(a4)       << 
410         REG_L s6,  TASK_THREAD_S6_RA(a4)       << 
411         REG_L s7,  TASK_THREAD_S7_RA(a4)       << 
412         REG_L s8,  TASK_THREAD_S8_RA(a4)       << 
413         REG_L s9,  TASK_THREAD_S9_RA(a4)       << 
414         REG_L s10, TASK_THREAD_S10_RA(a4)      << 
415         REG_L s11, TASK_THREAD_S11_RA(a4)      << 
416         /* The offset of thread_info in task_s << 
417         move tp, a1                            << 
418         /* Switch to the next shadow call stac << 
419         scs_load_current                       << 
420         ret                                    << 
421 SYM_FUNC_END(__switch_to)                      << 
422                                                << 
423 #ifndef CONFIG_MMU                             << 
424 #define do_page_fault do_trap_unknown          << 
425 #endif                                            1383 #endif
426                                                   1384 
427         .section ".rodata"                     !! 1385 /* End of entry.S */
428         .align LGREG                           << 
429         /* Exception vector table */           << 
430 SYM_DATA_START_LOCAL(excp_vect_table)          << 
431         RISCV_PTR do_trap_insn_misaligned      << 
432         ALT_INSN_FAULT(RISCV_PTR do_trap_insn_ << 
433         RISCV_PTR do_trap_insn_illegal         << 
434         RISCV_PTR do_trap_break                << 
435         RISCV_PTR do_trap_load_misaligned      << 
436         RISCV_PTR do_trap_load_fault           << 
437         RISCV_PTR do_trap_store_misaligned     << 
438         RISCV_PTR do_trap_store_fault          << 
439         RISCV_PTR do_trap_ecall_u /* system ca << 
440         RISCV_PTR do_trap_ecall_s              << 
441         RISCV_PTR do_trap_unknown              << 
442         RISCV_PTR do_trap_ecall_m              << 
443         /* instruciton page fault */           << 
444         ALT_PAGE_FAULT(RISCV_PTR do_page_fault << 
445         RISCV_PTR do_page_fault   /* load page << 
446         RISCV_PTR do_trap_unknown              << 
447         RISCV_PTR do_page_fault   /* store pag << 
448 SYM_DATA_END_LABEL(excp_vect_table, SYM_L_LOCA << 
449                                                << 
450 #ifndef CONFIG_MMU                             << 
451 SYM_DATA_START(__user_rt_sigreturn)            << 
452         li a7, __NR_rt_sigreturn               << 
453         ecall                                  << 
454 SYM_DATA_END(__user_rt_sigreturn)              << 
455 #endif                                         << 
                                                      

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php