~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/m68k/kernel/entry.S

Version: ~ [ linux-6.12-rc7 ] ~ [ linux-6.11.7 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.60 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.116 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.171 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.229 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.285 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.323 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.12 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 /* SPDX-License-Identifier: GPL-2.0-or-later
  2  * -*- mode: asm -*-
  3  *
  4  *  linux/arch/m68k/kernel/entry.S
  5  *
  6  *  Copyright (C) 1991, 1992  Linus Torvalds
  7  *
  8  * Linux/m68k support by Hamish Macdonald
  9  *
 10  * 68060 fixes by Jesper Skov
 11  *
 12  */
 13 
 14 /*
 15  * entry.S  contains the system-call and fault low-level handling routines.
 16  * This also contains the timer-interrupt handler, as well as all interrupts
 17  * and faults that can result in a task-switch.
 18  *
 19  * NOTE: This code handles signal-recognition, which happens every time
 20  * after a timer-interrupt and after each system call.
 21  *
 22  */
 23 
 24 /*
 25  * 12/03/96 Jes: Currently we only support m68k single-cpu systems, so
 26  *               all pointers that used to be 'current' are now entry
 27  *               number 0 in the 'current_set' list.
 28  *
 29  *  6/05/00 RZ:  addedd writeback completion after return from sighandler
 30  *               for 68040
 31  */
 32 
 33 #include <linux/linkage.h>
 34 #include <asm/errno.h>
 35 #include <asm/setup.h>
 36 #include <asm/traps.h>
 37 #include <asm/unistd.h>
 38 #include <asm/asm-offsets.h>
 39 #include <asm/entry.h>
 40 
 41 .globl system_call, buserr, trap, resume
 42 .globl sys_call_table
 43 .globl __sys_fork, __sys_clone, __sys_vfork
 44 .globl bad_interrupt
 45 .globl auto_irqhandler_fixup
 46 .globl user_irqvec_fixup
 47 
 48 .text
 49 ENTRY(__sys_fork)
 50         SAVE_SWITCH_STACK
 51         jbsr    sys_fork
 52         lea     %sp@(24),%sp
 53         rts
 54 
 55 ENTRY(__sys_clone)
 56         SAVE_SWITCH_STACK
 57         pea     %sp@(SWITCH_STACK_SIZE)
 58         jbsr    m68k_clone
 59         lea     %sp@(28),%sp
 60         rts
 61 
 62 ENTRY(__sys_vfork)
 63         SAVE_SWITCH_STACK
 64         jbsr    sys_vfork
 65         lea     %sp@(24),%sp
 66         rts
 67 
 68 ENTRY(__sys_clone3)
 69         SAVE_SWITCH_STACK
 70         pea     %sp@(SWITCH_STACK_SIZE)
 71         jbsr    m68k_clone3
 72         lea     %sp@(28),%sp
 73         rts
 74 
 75 ENTRY(sys_sigreturn)
 76         SAVE_SWITCH_STACK
 77         movel   %sp,%a1                         | switch_stack pointer
 78         lea     %sp@(SWITCH_STACK_SIZE),%a0     | pt_regs pointer
 79         lea     %sp@(-84),%sp                   | leave a gap
 80         movel   %a1,%sp@-
 81         movel   %a0,%sp@-
 82         jbsr    do_sigreturn
 83         jra     1f                              | shared with rt_sigreturn()
 84 
 85 ENTRY(sys_rt_sigreturn)
 86         SAVE_SWITCH_STACK
 87         movel   %sp,%a1                         | switch_stack pointer
 88         lea     %sp@(SWITCH_STACK_SIZE),%a0     | pt_regs pointer
 89         lea     %sp@(-84),%sp                   | leave a gap
 90         movel   %a1,%sp@-
 91         movel   %a0,%sp@-
 92         | stack contents:
 93         |   [original pt_regs address] [original switch_stack address]
 94         |   [gap] [switch_stack] [pt_regs] [exception frame]
 95         jbsr    do_rt_sigreturn
 96 
 97 1:
 98         | stack contents now:
 99         |   [original pt_regs address] [original switch_stack address]
100         |   [unused part of the gap] [moved switch_stack] [moved pt_regs]
101         |   [replacement exception frame]
102         | return value of do_{rt_,}sigreturn() points to moved switch_stack.
103 
104         movel   %d0,%sp                         | discard the leftover junk
105         RESTORE_SWITCH_STACK
106         | stack contents now is just [syscall return address] [pt_regs] [frame]
107         | return pt_regs.d0
108         movel   %sp@(PT_OFF_D0+4),%d0
109         rts
110 
111 ENTRY(buserr)
112         SAVE_ALL_INT
113         GET_CURRENT(%d0)
114         movel   %sp,%sp@-               | stack frame pointer argument
115         jbsr    buserr_c
116         addql   #4,%sp
117         jra     ret_from_exception
118 
119 ENTRY(trap)
120         SAVE_ALL_INT
121         GET_CURRENT(%d0)
122         movel   %sp,%sp@-               | stack frame pointer argument
123         jbsr    trap_c
124         addql   #4,%sp
125         jra     ret_from_exception
126 
127         | After a fork we jump here directly from resume,
128         | so that %d1 contains the previous task
129         | schedule_tail now used regardless of CONFIG_SMP
130 ENTRY(ret_from_fork)
131         movel   %d1,%sp@-
132         jsr     schedule_tail
133         addql   #4,%sp
134         jra     ret_from_exception
135 
136 ENTRY(ret_from_kernel_thread)
137         | a3 contains the kernel thread payload, d7 - its argument
138         movel   %d1,%sp@-
139         jsr     schedule_tail
140         movel   %d7,(%sp)
141         jsr     %a3@
142         addql   #4,%sp
143         jra     ret_from_exception
144 
145 #if defined(CONFIG_COLDFIRE) || !defined(CONFIG_MMU)
146 
147 #ifdef TRAP_DBG_INTERRUPT
148 
149 .globl dbginterrupt
150 ENTRY(dbginterrupt)
151         SAVE_ALL_INT
152         GET_CURRENT(%d0)
153         movel   %sp,%sp@-               /* stack frame pointer argument */
154         jsr     dbginterrupt_c
155         addql   #4,%sp
156         jra     ret_from_exception
157 #endif
158 
159 ENTRY(reschedule)
160         /* save top of frame */
161         pea     %sp@
162         jbsr    set_esp0
163         addql   #4,%sp
164         pea     ret_from_exception
165         jmp     schedule
166 
167 ENTRY(ret_from_user_signal)
168         moveq #__NR_sigreturn,%d0
169         trap #0
170 
171 ENTRY(ret_from_user_rt_signal)
172         movel #__NR_rt_sigreturn,%d0
173         trap #0
174 
175 #else
176 
177 do_trace_entry:
178         movel   #-ENOSYS,%sp@(PT_OFF_D0)| needed for strace
179         subql   #4,%sp
180         SAVE_SWITCH_STACK
181         jbsr    syscall_trace_enter
182         RESTORE_SWITCH_STACK
183         addql   #4,%sp
184         addql   #1,%d0                  | optimization for cmpil #-1,%d0
185         jeq     ret_from_syscall
186         movel   %sp@(PT_OFF_ORIG_D0),%d0
187         cmpl    #NR_syscalls,%d0
188         jcs     syscall
189         jra     ret_from_syscall
190 badsys:
191         movel   #-ENOSYS,%sp@(PT_OFF_D0)
192         jra     ret_from_syscall
193 
194 do_trace_exit:
195         subql   #4,%sp
196         SAVE_SWITCH_STACK
197         jbsr    syscall_trace_leave
198         RESTORE_SWITCH_STACK
199         addql   #4,%sp
200         jra     .Lret_from_exception
201 
202 ENTRY(system_call)
203         SAVE_ALL_SYS
204 
205         GET_CURRENT(%d1)
206         movel   %d1,%a1
207 
208         | save top of frame
209         movel   %sp,%curptr@(TASK_THREAD+THREAD_ESP0)
210 
211         | syscall trace?
212         tstb    %a1@(TINFO_FLAGS+2)
213         jmi     do_trace_entry
214         | seccomp filter active?
215         btst    #5,%a1@(TINFO_FLAGS+2)
216         bnes    do_trace_entry
217         cmpl    #NR_syscalls,%d0
218         jcc     badsys
219 syscall:
220         jbsr    @(sys_call_table,%d0:l:4)@(0)
221         movel   %d0,%sp@(PT_OFF_D0)     | save the return value
222 ret_from_syscall:
223         |oriw   #0x0700,%sr
224         movel   %curptr@(TASK_STACK),%a1
225         movew   %a1@(TINFO_FLAGS+2),%d0
226         jne     syscall_exit_work
227 1:      RESTORE_ALL
228 
229 syscall_exit_work:
230         btst    #5,%sp@(PT_OFF_SR)      | check if returning to kernel
231         bnes    1b                      | if so, skip resched, signals
232         lslw    #1,%d0
233         jcs     do_trace_exit
234         jmi     do_delayed_trace
235         lslw    #8,%d0
236         jne     do_signal_return
237         pea     resume_userspace
238         jra     schedule
239 
240 
241 ENTRY(ret_from_exception)
242 .Lret_from_exception:
243         btst    #5,%sp@(PT_OFF_SR)      | check if returning to kernel
244         bnes    1f                      | if so, skip resched, signals
245         | only allow interrupts when we are really the last one on the
246         | kernel stack, otherwise stack overflow can occur during
247         | heavy interrupt load
248         andw    #ALLOWINT,%sr
249 
250 resume_userspace:
251         movel   %curptr@(TASK_STACK),%a1
252         moveb   %a1@(TINFO_FLAGS+3),%d0
253         jne     exit_work
254 1:      RESTORE_ALL
255 
256 exit_work:
257         | save top of frame
258         movel   %sp,%curptr@(TASK_THREAD+THREAD_ESP0)
259         lslb    #1,%d0
260         jne     do_signal_return
261         pea     resume_userspace
262         jra     schedule
263 
264 
265 do_signal_return:
266         |andw   #ALLOWINT,%sr
267         subql   #4,%sp                  | dummy return address
268         SAVE_SWITCH_STACK
269         pea     %sp@(SWITCH_STACK_SIZE)
270         bsrl    do_notify_resume
271         addql   #4,%sp
272         RESTORE_SWITCH_STACK
273         addql   #4,%sp
274         jbra    resume_userspace
275 
276 do_delayed_trace:
277         bclr    #7,%sp@(PT_OFF_SR)      | clear trace bit in SR
278         pea     1                       | send SIGTRAP
279         movel   %curptr,%sp@-
280         pea     LSIGTRAP
281         jbsr    send_sig
282         addql   #8,%sp
283         addql   #4,%sp
284         jbra    resume_userspace
285 
286 
287 /* This is the main interrupt handler for autovector interrupts */
288 
289 ENTRY(auto_inthandler)
290         SAVE_ALL_INT
291         GET_CURRENT(%d0)
292                                         |  put exception # in d0
293         bfextu  %sp@(PT_OFF_FORMATVEC){#4,#10},%d0
294         subw    #VEC_SPUR,%d0
295 
296         movel   %sp,%sp@-
297         movel   %d0,%sp@-               |  put vector # on stack
298 auto_irqhandler_fixup = . + 2
299         jsr     do_IRQ                  |  process the IRQ
300         addql   #8,%sp                  |  pop parameters off stack
301         jra     ret_from_exception
302 
303 /* Handler for user defined interrupt vectors */
304 
305 ENTRY(user_inthandler)
306         SAVE_ALL_INT
307         GET_CURRENT(%d0)
308                                         |  put exception # in d0
309         bfextu  %sp@(PT_OFF_FORMATVEC){#4,#10},%d0
310 user_irqvec_fixup = . + 2
311         subw    #VEC_USER,%d0
312 
313         movel   %sp,%sp@-
314         movel   %d0,%sp@-               |  put vector # on stack
315         jsr     do_IRQ                  |  process the IRQ
316         addql   #8,%sp                  |  pop parameters off stack
317         jra     ret_from_exception
318 
319 /* Handler for uninitialized and spurious interrupts */
320 
321 ENTRY(bad_inthandler)
322         SAVE_ALL_INT
323         GET_CURRENT(%d0)
324 
325         movel   %sp,%sp@-
326         jsr     handle_badint
327         addql   #4,%sp
328         jra     ret_from_exception
329 
330 resume:
331         /*
332          * Beware - when entering resume, prev (the current task) is
333          * in a0, next (the new task) is in a1,so don't change these
334          * registers until their contents are no longer needed.
335          */
336 
337         /* save sr */
338         movew   %sr,%a0@(TASK_THREAD+THREAD_SR)
339 
340         /* save fs (sfc,%dfc) (may be pointing to kernel memory) */
341         movec   %sfc,%d0
342         movew   %d0,%a0@(TASK_THREAD+THREAD_FC)
343 
344         /* save usp */
345         /* it is better to use a movel here instead of a movew 8*) */
346         movec   %usp,%d0
347         movel   %d0,%a0@(TASK_THREAD+THREAD_USP)
348 
349         /* save non-scratch registers on stack */
350         SAVE_SWITCH_STACK
351 
352         /* save current kernel stack pointer */
353         movel   %sp,%a0@(TASK_THREAD+THREAD_KSP)
354 
355         /* save floating point context */
356 #ifndef CONFIG_M68KFPU_EMU_ONLY
357 #ifdef CONFIG_M68KFPU_EMU
358         tstl    m68k_fputype
359         jeq     3f
360 #endif
361         fsave   %a0@(TASK_THREAD+THREAD_FPSTATE)
362 
363 #if defined(CONFIG_M68060)
364 #if !defined(CPU_M68060_ONLY)
365         btst    #3,m68k_cputype+3
366         beqs    1f
367 #endif
368         /* The 060 FPU keeps status in bits 15-8 of the first longword */
369         tstb    %a0@(TASK_THREAD+THREAD_FPSTATE+2)
370         jeq     3f
371 #if !defined(CPU_M68060_ONLY)
372         jra     2f
373 #endif
374 #endif /* CONFIG_M68060 */
375 #if !defined(CPU_M68060_ONLY)
376 1:      tstb    %a0@(TASK_THREAD+THREAD_FPSTATE)
377         jeq     3f
378 #endif
379 2:      fmovemx %fp0-%fp7,%a0@(TASK_THREAD+THREAD_FPREG)
380         fmoveml %fpcr/%fpsr/%fpiar,%a0@(TASK_THREAD+THREAD_FPCNTL)
381 3:
382 #endif  /* CONFIG_M68KFPU_EMU_ONLY */
383         /* Return previous task in %d1 */
384         movel   %curptr,%d1
385 
386         /* switch to new task (a1 contains new task) */
387         movel   %a1,%curptr
388 
389         /* restore floating point context */
390 #ifndef CONFIG_M68KFPU_EMU_ONLY
391 #ifdef CONFIG_M68KFPU_EMU
392         tstl    m68k_fputype
393         jeq     4f
394 #endif
395 #if defined(CONFIG_M68060)
396 #if !defined(CPU_M68060_ONLY)
397         btst    #3,m68k_cputype+3
398         beqs    1f
399 #endif
400         /* The 060 FPU keeps status in bits 15-8 of the first longword */
401         tstb    %a1@(TASK_THREAD+THREAD_FPSTATE+2)
402         jeq     3f
403 #if !defined(CPU_M68060_ONLY)
404         jra     2f
405 #endif
406 #endif /* CONFIG_M68060 */
407 #if !defined(CPU_M68060_ONLY)
408 1:      tstb    %a1@(TASK_THREAD+THREAD_FPSTATE)
409         jeq     3f
410 #endif
411 2:      fmovemx %a1@(TASK_THREAD+THREAD_FPREG),%fp0-%fp7
412         fmoveml %a1@(TASK_THREAD+THREAD_FPCNTL),%fpcr/%fpsr/%fpiar
413 3:      frestore %a1@(TASK_THREAD+THREAD_FPSTATE)
414 4:
415 #endif  /* CONFIG_M68KFPU_EMU_ONLY */
416 
417         /* restore the kernel stack pointer */
418         movel   %a1@(TASK_THREAD+THREAD_KSP),%sp
419 
420         /* restore non-scratch registers */
421         RESTORE_SWITCH_STACK
422 
423         /* restore user stack pointer */
424         movel   %a1@(TASK_THREAD+THREAD_USP),%a0
425         movel   %a0,%usp
426 
427         /* restore fs (sfc,%dfc) */
428         movew   %a1@(TASK_THREAD+THREAD_FC),%a0
429         movec   %a0,%sfc
430         movec   %a0,%dfc
431 
432         /* restore status register */
433         movew   %a1@(TASK_THREAD+THREAD_SR),%d0
434         oriw    #0x0700,%d0
435         movew   %d0,%sr
436 
437         rts
438 
439 #endif /* CONFIG_MMU && !CONFIG_COLDFIRE */

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php