1 /* SPDX-License-Identifier: GPL-2.0-or-later * !! 1 /* SPDX-License-Identifier: GPL-2.0-or-later 2 /* !! 2 * -*- mode: asm -*- 3 * Linux/PA-RISC Project (http://www.parisc-li !! 3 * >> 4 * linux/arch/m68k/kernel/entry.S >> 5 * >> 6 * Copyright (C) 1991, 1992 Linus Torvalds >> 7 * >> 8 * Linux/m68k support by Hamish Macdonald >> 9 * >> 10 * 68060 fixes by Jesper Skov 4 * 11 * 5 * kernel entry points (interruptions, system << 6 * Copyright (C) 1999,2000 Philipp Rumpf << 7 * Copyright (C) 1999 SuSE GmbH Nuernberg << 8 * Copyright (C) 2000 Hewlett-Packard (John M << 9 * Copyright (C) 1999 Hewlett-Packard (Frank << 10 */ 12 */ 11 13 12 #include <asm/asm-offsets.h> << 13 << 14 /* we have the following possibilities to act << 15 * - handle in assembly and use shadowed regi << 16 * - save registers to kernel stack and handl << 17 << 18 << 19 #include <asm/psw.h> << 20 #include <asm/cache.h> /* for L1_CACH << 21 #include <asm/assembly.h> /* for LDREG/S << 22 #include <asm/signal.h> << 23 #include <asm/unistd.h> << 24 #include <asm/ldcw.h> << 25 #include <asm/traps.h> << 26 #include <asm/thread_info.h> << 27 #include <asm/alternative.h> << 28 #include <asm/spinlock_types.h> << 29 << 30 #include <linux/linkage.h> << 31 #include <linux/pgtable.h> << 32 << 33 #ifdef CONFIG_64BIT << 34 .level 2.0w << 35 #else << 36 .level 2.0 << 37 #endif << 38 << 39 /* 14 /* 40 * We need seven instructions after a TLB inse !! 15 * entry.S contains the system-call and fault low-level handling routines. 41 * The PA8800/PA8900 processors are an excepti !! 16 * This also contains the timer-interrupt handler, as well as all interrupts 42 * The RFI changes both IAOQ_Back and IAOQ_Fro !! 17 * and faults that can result in a task-switch. >> 18 * >> 19 * NOTE: This code handles signal-recognition, which happens every time >> 20 * after a timer-interrupt and after each system call. >> 21 * 43 */ 22 */ 44 #ifdef CONFIG_64BIT << 45 #define NUM_PIPELINE_INSNS 12 << 46 #else << 47 #define NUM_PIPELINE_INSNS 7 << 48 #endif << 49 << 50 /* Insert num nops */ << 51 .macro insert_nops num << 52 .rept \num << 53 nop << 54 .endr << 55 .endm << 56 << 57 /* Get aligned page_table_lock address << 58 .macro get_ptl reg << 59 mfctl %cr28,\reg << 60 .endm << 61 << 62 /* space_to_prot macro creates a prot << 63 << 64 #if (SPACEID_SHIFT) == 0 << 65 .macro space_to_prot spc prot << 66 depd,z \spc,62,31,\prot << 67 .endm << 68 #else << 69 .macro space_to_prot spc prot << 70 extrd,u \spc,(64 - (SPACEID_SHIFT)),32 << 71 .endm << 72 #endif << 73 /* << 74 * The "get_stack" macros are responsi << 75 * kernel stack value. << 76 * << 77 * If sr7 == 0 << 78 * Already using a kernel sta << 79 * get_stack_use_r30 macro to << 80 * on the stack, and store re << 81 * else << 82 * Need to set up a kernel st << 83 * get_stack_use_cr30 macro t << 84 * to the pt_regs structure c << 85 * task pointer pointed to by << 86 * pointer from the task stru << 87 * << 88 * Note that we use shadowed registers << 89 * we can save %r26 and %r29. %r26 is << 90 * %r8 (a shadowed register) which tem << 91 * either the fault type ("code") or t << 92 * to use a non-shadowed register to c << 93 * the rfir in virt_map. We use %r26 s << 94 * up being passed as the argument to << 95 * or handle_interruption. %r29 is use << 96 * the register save area, and once ag << 97 * be a non-shadowed register so that << 98 */ << 99 << 100 .macro get_stack_use_cr30 << 101 << 102 /* we save the registers in the task s << 103 << 104 copy %r30, %r17 << 105 mfctl %cr30, %r1 << 106 tophys %r1,%r9 /* task_struct << 107 LDREG TASK_STACK(%r9),%r30 << 108 ldo PT_SZ_ALGN(%r30),%r30 << 109 mtsp %r0,%sr7 /* clear sr7 a << 110 mtsp %r16,%sr3 << 111 ldo TASK_REGS(%r9),%r9 << 112 STREG %r17,PT_GR30(%r9) << 113 STREG %r29,PT_GR29(%r9) << 114 STREG %r26,PT_GR26(%r9) << 115 STREG %r16,PT_SR7(%r9) << 116 copy %r9,%r29 << 117 .endm << 118 << 119 .macro get_stack_use_r30 << 120 << 121 /* we put a struct pt_regs on the stac << 122 << 123 tophys %r30,%r9 << 124 copy %r30,%r1 << 125 ldo PT_SZ_ALGN(%r30),%r30 << 126 STREG %r1,PT_GR30(%r9) << 127 STREG %r29,PT_GR29(%r9) << 128 STREG %r26,PT_GR26(%r9) << 129 STREG %r16,PT_SR7(%r9) << 130 copy %r9,%r29 << 131 .endm << 132 << 133 .macro rest_stack << 134 LDREG PT_GR1(%r29), %r1 << 135 LDREG PT_GR30(%r29),%r30 << 136 LDREG PT_GR29(%r29),%r29 << 137 .endm << 138 << 139 /* default interruption handler << 140 * (calls traps.c:handle_interruption) << 141 .macro def code << 142 b intr_save << 143 ldi \code, %r8 << 144 .align 32 << 145 .endm << 146 << 147 /* Interrupt interruption handler << 148 * (calls irq.c:do_cpu_irq_mask) */ << 149 .macro extint code << 150 b intr_extint << 151 mfsp %sr7,%r16 << 152 .align 32 << 153 .endm << 154 << 155 .import os_hpmc, code << 156 << 157 /* HPMC handler */ << 158 .macro hpmc code << 159 nop /* must be a N << 160 load32 PA(os_hpmc), %r3 << 161 bv,n 0(%r3) << 162 nop << 163 .word 0 /* checksum (w << 164 .word 0 /* address of << 165 .word 0 /* length of h << 166 .endm << 167 << 168 /* << 169 * Performance Note: Instructions will << 170 * this part of the code later on, onc << 171 * that the tlb miss handlers are clos << 172 */ << 173 << 174 /* Register definitions for tlb miss h << 175 << 176 va = r8 /* virtual address for << 177 spc = r24 /* space for which the << 178 << 179 #ifndef CONFIG_64BIT << 180 << 181 /* << 182 * itlb miss interruption handler (par << 183 */ << 184 << 185 .macro itlb_11 code << 186 << 187 mfctl %pcsq, spc << 188 b itlb_miss_11 << 189 mfctl %pcoq, va << 190 << 191 .align 32 << 192 .endm << 193 #endif << 194 << 195 /* << 196 * itlb miss interruption handler (par << 197 */ << 198 << 199 .macro itlb_20 code << 200 mfctl %pcsq, spc << 201 #ifdef CONFIG_64BIT << 202 b itlb_miss_20w << 203 #else << 204 b itlb_miss_20 << 205 #endif << 206 mfctl %pcoq, va << 207 << 208 .align 32 << 209 .endm << 210 << 211 #ifndef CONFIG_64BIT << 212 /* << 213 * naitlb miss interruption handler (p << 214 */ << 215 << 216 .macro naitlb_11 code << 217 << 218 mfctl %isr,spc << 219 b naitlb_miss_11 << 220 mfctl %ior,va << 221 << 222 .align 32 << 223 .endm << 224 #endif << 225 << 226 /* << 227 * naitlb miss interruption handler (p << 228 */ << 229 23 230 .macro naitlb_20 code !! 24 /* 231 !! 25 * 12/03/96 Jes: Currently we only support m68k single-cpu systems, so 232 mfctl %isr,spc !! 26 * all pointers that used to be 'current' are now entry 233 #ifdef CONFIG_64BIT !! 27 * number 0 in the 'current_set' list. 234 b naitlb_miss_20w !! 28 * 235 #else !! 29 * 6/05/00 RZ: addedd writeback completion after return from sighandler 236 b naitlb_miss_20 !! 30 * for 68040 237 #endif !! 31 */ 238 mfctl %ior,va << 239 << 240 .align 32 << 241 .endm << 242 << 243 #ifndef CONFIG_64BIT << 244 /* << 245 * dtlb miss interruption handler (par << 246 */ << 247 << 248 .macro dtlb_11 code << 249 << 250 mfctl %isr, spc << 251 b dtlb_miss_11 << 252 mfctl %ior, va << 253 << 254 .align 32 << 255 .endm << 256 #endif << 257 << 258 /* << 259 * dtlb miss interruption handler (par << 260 */ << 261 << 262 .macro dtlb_20 code << 263 << 264 mfctl %isr, spc << 265 #ifdef CONFIG_64BIT << 266 b dtlb_miss_20w << 267 #else << 268 b dtlb_miss_20 << 269 #endif << 270 mfctl %ior, va << 271 << 272 .align 32 << 273 .endm << 274 << 275 #ifndef CONFIG_64BIT << 276 /* nadtlb miss interruption handler (p << 277 << 278 .macro nadtlb_11 code << 279 << 280 mfctl %isr,spc << 281 b nadtlb_miss_11 << 282 mfctl %ior,va << 283 << 284 .align 32 << 285 .endm << 286 #endif << 287 << 288 /* nadtlb miss interruption handler (p << 289 << 290 .macro nadtlb_20 code << 291 << 292 mfctl %isr,spc << 293 #ifdef CONFIG_64BIT << 294 b nadtlb_miss_20w << 295 #else << 296 b nadtlb_miss_20 << 297 #endif << 298 mfctl %ior,va << 299 << 300 .align 32 << 301 .endm << 302 << 303 #ifndef CONFIG_64BIT << 304 /* << 305 * dirty bit trap interruption handler << 306 */ << 307 << 308 .macro dbit_11 code << 309 << 310 mfctl %isr,spc << 311 b dbit_trap_11 << 312 mfctl %ior,va << 313 << 314 .align 32 << 315 .endm << 316 #endif << 317 << 318 /* << 319 * dirty bit trap interruption handler << 320 */ << 321 << 322 .macro dbit_20 code << 323 << 324 mfctl %isr,spc << 325 #ifdef CONFIG_64BIT << 326 b dbit_trap_20w << 327 #else << 328 b dbit_trap_20 << 329 #endif << 330 mfctl %ior,va << 331 << 332 .align 32 << 333 .endm << 334 << 335 /* In LP64, the space contains part of << 336 * fault. We have to extract this and << 337 * zeroing the corresponding bits in t << 338 .macro space_adjust spc,va << 339 #ifdef CONFIG_64BIT << 340 extrd,u \spc,63,SPACEID_SHIFT, << 341 depd %r0,63,SPACEID_SHIFT,\ << 342 depd \tmp,31,SPACEID_SHIFT, << 343 #endif << 344 .endm << 345 << 346 .import swapper_pg_dir,code << 347 << 348 /* Get the pgd. For faults on space z << 349 * is simply swapper_pg_dir. For user << 350 * pgd is stored in %cr25 */ << 351 .macro get_pgd spc,re << 352 ldil L%PA(swapper_pg_dir),\ << 353 ldo R%PA(swapper_pg_dir)(\ << 354 or,COND(=) %r0,\spc,%r0 << 355 mfctl %cr25,\reg << 356 .endm << 357 << 358 /* << 359 space_check(spc,tmp,fault) << 360 << 361 spc - The space we saw the fau << 362 tmp - The place to store the c << 363 fault - Function to call on fa << 364 << 365 Only allow faults on different << 366 currently active one if we're << 367 << 368 */ << 369 .macro space_check spc,tm << 370 mfsp %sr7,\tmp << 371 /* check against %r0 which is same val << 372 or,COND(<>) %r0,\spc,%r0 /* use << 373 * as << 374 * che << 375 copy \spc,\tmp << 376 or,COND(=) %r0,\tmp,%r0 /* nul << 377 cmpb,COND(<>),n \tmp,\spc,\fault << 378 .endm << 379 << 380 /* Look up a PTE in a 2-Level scheme ( << 381 * level if the entry isn't present << 382 * << 383 * NOTE: we use ldw even for LP64, sin << 384 * can address up to 1TB << 385 */ << 386 .macro L2_ptep pmd,pte,index, << 387 #if CONFIG_PGTABLE_LEVELS == 3 << 388 extru_safe \va,31-ASM_PMD_SHIFT,A << 389 #else << 390 extru_safe \va,31-ASM_PGDIR_SHIFT << 391 #endif << 392 dep %r0,31,PAGE_SHIFT,\pmd << 393 #if CONFIG_PGTABLE_LEVELS < 3 << 394 copy %r0,\pte << 395 #endif << 396 ldw,s \index(\pmd),\pmd << 397 bb,>=,n \pmd,_PxD_PRESENT_BIT, << 398 dep %r0,31,PxD_FLAG_SHIFT, << 399 SHLREG \pmd,PxD_VALUE_SHIFT,\ << 400 extru_safe \va,31-PAGE_SHIFT,ASM_ << 401 dep %r0,31,PAGE_SHIFT,\pmd << 402 shladd \index,BITS_PER_PTE_EN << 403 .endm << 404 << 405 /* Look up PTE in a 3-Level scheme. */ << 406 .macro L3_ptep pgd,pte,index, << 407 #if CONFIG_PGTABLE_LEVELS == 3 << 408 copy %r0,\pte << 409 extrd,u \va,63-ASM_PGDIR_SHIFT << 410 ldw,s \index(\pgd),\pgd << 411 bb,>=,n \pgd,_PxD_PRESENT_BIT, << 412 shld \pgd,PxD_VALUE_SHIFT,\ << 413 #endif << 414 L2_ptep \pgd,\pte,\index,\va,\ << 415 .endm << 416 << 417 /* Acquire page_table_lock and check p << 418 .macro ptl_lock spc,pt << 419 #ifdef CONFIG_TLB_PTLOCK << 420 98: cmpib,COND(=),n 0,\spc,2f << 421 get_ptl \tmp << 422 1: LDCW 0(\tmp),\tmp1 << 423 cmpib,COND(=) 0,\tmp1,1b << 424 nop << 425 LDREG 0(\ptp),\pte << 426 bb,<,n \pte,_PAGE_PRESENT_BIT << 427 b \fault << 428 stw \tmp1,0(\tmp) << 429 99: ALTERNATIVE(98b, 99b, ALT_COND_NO_SMP, << 430 #endif << 431 2: LDREG 0(\ptp),\pte << 432 bb,>=,n \pte,_PAGE_PRESENT_BIT << 433 3: << 434 .endm << 435 << 436 /* Release page_table_lock if for user << 437 store to ensure all prior accesses << 438 releasing the lock. Note stw may no << 439 provide one extra nop when CONFIG_T << 440 .macro ptl_unlock spc,tm << 441 #ifdef CONFIG_TLB_PTLOCK << 442 98: get_ptl \tmp << 443 ldi __ARCH_SPIN_LOCK_UNLOC << 444 or,COND(=) %r0,\spc,%r0 << 445 stw,ma \tmp2,0(\tmp) << 446 99: ALTERNATIVE(98b, 99b, ALT_COND_NO_SMP, << 447 insert_nops NUM_PIPELINE_INSNS - 4 << 448 #else << 449 insert_nops NUM_PIPELINE_INSNS - 1 << 450 #endif << 451 .endm << 452 << 453 /* Set the _PAGE_ACCESSED bit of the P << 454 * don't needlessly dirty the cache li << 455 .macro update_accessed ptp,pt << 456 ldi _PAGE_ACCESSED,\tmp1 << 457 or \tmp1,\pte,\tmp << 458 and,COND(<>) \tmp1,\pte,%r0 << 459 STREG \tmp,0(\ptp) << 460 .endm << 461 << 462 /* Set the dirty bit (and accessed bit << 463 * clever, this is only used from the << 464 .macro update_dirty ptp,pt << 465 ldi _PAGE_ACCESSED|_PAGE_D << 466 or \tmp,\pte,\pte << 467 STREG \pte,0(\ptp) << 468 .endm << 469 << 470 /* We have (depending on the page size << 471 * - 38 to 52-bit Physical Page Number << 472 * - 12 to 26-bit page offset << 473 */ << 474 /* bitshift difference between a PFN ( << 475 * to a CPU TLB 4k PFN (4k => 12 bits << 476 #define PAGE_ADD_SHIFT (PAGE_ << 477 #define PAGE_ADD_HUGE_SHIFT (REAL_ << 478 #define PFN_START_BIT (63-ASM_PFN_PT << 479 << 480 /* Drop prot bits and convert to page << 481 .macro convert_for_tlb_insert << 482 #ifdef CONFIG_HUGETLB_PAGE << 483 copy \pte,\tmp << 484 extrd,u \tmp,PFN_START_BIT,PFN << 485 << 486 depdi _PAGE_SIZE_ENCODING_DE << 487 (63-58)+PAGE_A << 488 extrd,u,*= \tmp,_PAGE_HPAGE_BIT+3 << 489 depdi _HUGE_PAGE_SIZE_ENCODI << 490 (63-58)+PAGE_A << 491 #else /* Huge pages disabled */ << 492 extrd,u \pte,PFN_START_BIT,PFN << 493 depdi _PAGE_SIZE_ENCODING_DE << 494 (63-58)+PAGE_A << 495 #endif << 496 .endm << 497 << 498 /* Convert the pte and prot to tlb ins << 499 * this happens is quite subtle, read << 500 .macro make_insert_tlb spc,pt << 501 space_to_prot \spc \prot /* c << 502 /* The following is the real subtlety. << 503 * T <-> _PAGE_REFTRAP << 504 * D <-> _PAGE_DIRTY << 505 * B <-> _PAGE_DMB (memory break) << 506 * << 507 * Then incredible subtlety: The acces << 508 * _PAGE_GATEWAY, _PAGE_EXEC and _PAGE << 509 * See 3-14 of the parisc 2.0 manual << 510 * << 511 * Finally, _PAGE_READ goes in the top << 512 * trigger an access rights trap in us << 513 * tries to read an unreadable page */ << 514 #if _PAGE_SPECIAL_BIT == _PAGE_DMB_BIT << 515 /* need to drop DMB bit, as it's used << 516 depi 0,_PAGE_SPECIAL_BIT,1, << 517 #endif << 518 depd \pte,8,7,\prot << 519 << 520 /* PAGE_USER indicates the page can be << 521 * so deposit X1|11 to PL1|PL2 (rememb << 522 * contains _PAGE_READ) */ << 523 extrd,u,*= \pte,_PAGE_USER_BIT+32 << 524 depdi 7,11,3,\prot << 525 /* If we're a gateway page, drop PL2 b << 526 * to kernel privilege (so we can exec << 527 * Any privilege promotion page always << 528 extrd,u,*= \pte,_PAGE_GATEWAY_BIT << 529 depd %r0,11,2,\prot /* If << 530 << 531 /* Enforce uncacheable pages. << 532 * This should ONLY be use for MMIO on << 533 * Memory/DMA is cache coherent on all << 534 * (that means T-class is NOT supporte << 535 * on most of those machines only hand << 536 */ << 537 extrd,u,*= \pte,_PAGE_NO_CACHE_BI << 538 depdi 1,12,1,\prot << 539 << 540 /* Drop prot bits and convert to page << 541 convert_for_tlb_insert20 \pte \tmp << 542 .endm << 543 << 544 /* Identical macro to make_insert_tlb << 545 * makes the tlb entry for the differe << 546 * insertion instructions */ << 547 .macro make_insert_tlb_11 << 548 #if _PAGE_SPECIAL_BIT == _PAGE_DMB_BIT << 549 /* need to drop DMB bit, as it's used << 550 depi 0,_PAGE_SPECIAL_BIT,1, << 551 #endif << 552 zdep \spc,30,15,\prot << 553 dep \pte,8,7,\prot << 554 extru,= \pte,_PAGE_NO_CACHE_BI << 555 depi 1,12,1,\prot << 556 extru,= \pte,_PAGE_USER_BIT,1, << 557 depi 7,11,3,\prot /* Set << 558 extru,= \pte,_PAGE_GATEWAY_BIT << 559 depi 0,11,2,\prot /* If << 560 << 561 /* Get rid of prot bits and convert to << 562 << 563 depi 0,31,ASM_PFN_PTE_SHIFT << 564 SHRREG \pte,(ASM_PFN_PTE_SHIF << 565 .endm << 566 << 567 /* This is for ILP32 PA2.0 only. The << 568 * to extend into I/O space if the add << 569 * so we extend the f's into the top w << 570 * this case */ << 571 .macro f_extend pte,tm << 572 extrd,s \pte,42,4,\tmp << 573 addi,<> 1,\tmp,%r0 << 574 extrd,s \pte,63,25,\pte << 575 .endm << 576 << 577 /* The alias region is comprised of a << 578 * aligned to 8 MB. It is used to clea << 579 * using kernel virtual addresses cong << 580 * virtual address. << 581 * << 582 * To use the alias page, you set %r26 << 583 * entry (identifying the physical pag << 584 * the from tlb entry (or nothing if o << 585 * clear_user_page_asm) */ << 586 .macro do_alias spc,tm << 587 cmpib,COND(<>),n 0,\spc,\fault << 588 ldil L%(TMPALIAS_MAP_START) << 589 copy \va,\tmp1 << 590 depi_safe 0,31,TMPALIAS_SIZE_BIT << 591 cmpb,COND(<>),n \tmp,\tmp1,\fault << 592 mfctl %cr19,\tmp /* iir << 593 /* get the opcode (first six bits) int << 594 extrw,u \tmp,5,6,\tmp << 595 /* << 596 * Only setting the T bit prevents dat << 597 * Setting access rights to zero preve << 598 * << 599 * Note subtlety here: _PAGE_GATEWAY, << 600 * to type field and _PAGE_READ goes t << 601 */ << 602 ldi (_PAGE_REFTRAP|_PAGE_R << 603 /* << 604 * so if the opcode is one (i.e. this << 605 * instruction) nullify the next load << 606 * Otherwise this is a normal data ope << 607 */ << 608 cmpiclr,= 0x01,\tmp,%r0 << 609 ldi (_PAGE_DIRTY|_PAGE_REA << 610 .ifc \patype,20 << 611 depd,z \prot,8,7,\prot << 612 .else << 613 .ifc \patype,11 << 614 depw,z \prot,8,7,\prot << 615 .else << 616 .error "undefined PA type to do_alias" << 617 .endif << 618 .endif << 619 /* << 620 * OK, it is in the temp alias region, << 621 * Check "subtle" note in pacache.S re << 622 */ << 623 extrw,u,= \va,31-TMPALIAS_SIZE_B << 624 or,COND(tr) %r23,%r0,\pte << 625 or %r26,%r0,\pte << 626 << 627 /* convert phys addr in \pte (from r23 << 628 SHRREG \pte,PAGE_SHIFT+PAGE_A << 629 depi_safe _PAGE_SIZE_ENCODING_DE << 630 .endm << 631 << 632 << 633 /* << 634 * Fault_vectors are architecturally r << 635 * boundary << 636 */ << 637 << 638 .section .text.hot << 639 .align 2048 << 640 << 641 ENTRY(fault_vector_20) << 642 /* First vector is invalid (0) */ << 643 .ascii "cows can fly" << 644 .byte 0 << 645 .align 32 << 646 << 647 hpmc 1 << 648 def 2 << 649 def 3 << 650 extint 4 << 651 def 5 << 652 itlb_20 PARISC_ITLB_TRAP << 653 def 7 << 654 def 8 << 655 def 9 << 656 def 10 << 657 def 11 << 658 def 12 << 659 def 13 << 660 def 14 << 661 dtlb_20 15 << 662 naitlb_20 16 << 663 nadtlb_20 17 << 664 def 18 << 665 def 19 << 666 dbit_20 20 << 667 def 21 << 668 def 22 << 669 def 23 << 670 def 24 << 671 def 25 << 672 def 26 << 673 def 27 << 674 def 28 << 675 def 29 << 676 def 30 << 677 def 31 << 678 END(fault_vector_20) << 679 << 680 #ifndef CONFIG_64BIT << 681 << 682 .align 2048 << 683 << 684 ENTRY(fault_vector_11) << 685 /* First vector is invalid (0) */ << 686 .ascii "cows can fly" << 687 .byte 0 << 688 .align 32 << 689 << 690 hpmc 1 << 691 def 2 << 692 def 3 << 693 extint 4 << 694 def 5 << 695 itlb_11 PARISC_ITLB_TRAP << 696 def 7 << 697 def 8 << 698 def 9 << 699 def 10 << 700 def 11 << 701 def 12 << 702 def 13 << 703 def 14 << 704 dtlb_11 15 << 705 naitlb_11 16 << 706 nadtlb_11 17 << 707 def 18 << 708 def 19 << 709 dbit_11 20 << 710 def 21 << 711 def 22 << 712 def 23 << 713 def 24 << 714 def 25 << 715 def 26 << 716 def 27 << 717 def 28 << 718 def 29 << 719 def 30 << 720 def 31 << 721 END(fault_vector_11) << 722 << 723 #endif << 724 /* Fault vector is separately protecte << 725 .align PAGE_SIZE << 726 << 727 .import handle_interruption,co << 728 .import do_cpu_irq_mask,code << 729 << 730 /* << 731 * Child Returns here << 732 * << 733 * copy_thread moved args into task sa << 734 */ << 735 << 736 ENTRY(ret_from_kernel_thread) << 737 /* Call schedule_tail first though */ << 738 BL schedule_tail, %r2 << 739 nop << 740 << 741 mfctl %cr30,%r1 /* task_struct << 742 LDREG TASK_PT_GR25(%r1), %r26 << 743 #ifdef CONFIG_64BIT << 744 LDREG TASK_PT_GR27(%r1), %r27 << 745 #endif << 746 LDREG TASK_PT_GR26(%r1), %r1 << 747 ble 0(%sr7, %r1) << 748 copy %r31, %r2 << 749 b finish_child_return << 750 nop << 751 END(ret_from_kernel_thread) << 752 << 753 << 754 /* << 755 * struct task_struct *_switch_to(stru << 756 * struct task_struct *next) << 757 * << 758 * switch kernel stacks and return pre << 759 ENTRY_CFI(_switch_to) << 760 STREG %r2, -RP_OFFSET(%r30) << 761 << 762 callee_save_float << 763 callee_save << 764 << 765 load32 _switch_to_ret, %r2 << 766 << 767 STREG %r2, TASK_PT_KPC(%r26) << 768 LDREG TASK_PT_KPC(%r25), %r2 << 769 << 770 STREG %r30, TASK_PT_KSP(%r26) << 771 LDREG TASK_PT_KSP(%r25), %r30 << 772 bv %r0(%r2) << 773 mtctl %r25,%cr30 << 774 << 775 ENTRY(_switch_to_ret) << 776 mtctl %r0, %cr0 /* Nee << 777 callee_rest << 778 callee_rest_float << 779 << 780 LDREG -RP_OFFSET(%r30), %r2 << 781 bv %r0(%r2) << 782 copy %r26, %r28 << 783 ENDPROC_CFI(_switch_to) << 784 << 785 /* << 786 * Common rfi return path for interrup << 787 * sys_rt_sigreturn (sometimes). The << 788 * return via this path if the signal << 789 * was running; if the process was blo << 790 * normal syscall_exit path is used. << 791 * proceses exit via intr_restore. << 792 * << 793 * XXX If any syscalls that change a p << 794 * this way, then we will need to copy << 795 * adjust IASQ[0..1]. << 796 * << 797 */ << 798 << 799 .align PAGE_SIZE << 800 << 801 ENTRY_CFI(syscall_exit_rfi) << 802 mfctl %cr30,%r16 /* tas << 803 ldo TASK_REGS(%r16),%r16 << 804 /* Force iaoq to userspace, as the use << 805 * context via sigcontext. Also Filter << 806 */ << 807 LDREG PT_IAOQ0(%r16),%r19 << 808 depi PRIV_USER,31,2,%r19 << 809 STREG %r19,PT_IAOQ0(%r16) << 810 LDREG PT_IAOQ1(%r16),%r19 << 811 depi PRIV_USER,31,2,%r19 << 812 STREG %r19,PT_IAOQ1(%r16) << 813 LDREG PT_PSW(%r16),%r19 << 814 load32 USER_PSW_MASK,%r1 << 815 #ifdef CONFIG_64BIT << 816 load32 USER_PSW_HI_MASK,%r20 << 817 depd %r20,31,32,%r1 << 818 #endif << 819 and %r19,%r1,%r19 /* Mask out bits << 820 load32 USER_PSW,%r1 << 821 or %r19,%r1,%r19 /* Make sure def << 822 STREG %r19,PT_PSW(%r16) << 823 << 824 /* << 825 * If we aren't being traced, we never << 826 * (we don't store them in the sigcont << 827 * to "proper" values now (otherwise w << 828 * whatever was last stored in the tas << 829 * be inconsistent if an interrupt occ << 830 * page). Note that we may be "trashin << 831 * them, but we don't support the user << 832 */ << 833 << 834 STREG %r0,PT_SR2(%r16) << 835 mfsp %sr3,%r19 << 836 STREG %r19,PT_SR0(%r16) << 837 STREG %r19,PT_SR1(%r16) << 838 STREG %r19,PT_SR3(%r16) << 839 STREG %r19,PT_SR4(%r16) << 840 STREG %r19,PT_SR5(%r16) << 841 STREG %r19,PT_SR6(%r16) << 842 STREG %r19,PT_SR7(%r16) << 843 << 844 ENTRY(intr_return) << 845 /* check for reschedule */ << 846 mfctl %cr30,%r1 << 847 LDREG TASK_TI_FLAGS(%r1),%r19 /* sch << 848 bb,<,n %r19,31-TIF_NEED_RESCHED,intr_ << 849 << 850 .import do_notify_resume,code << 851 intr_check_sig: << 852 /* As above */ << 853 mfctl %cr30,%r1 << 854 LDREG TASK_TI_FLAGS(%r1),%r19 << 855 ldi (_TIF_USER_WORK_MASK & ~_TIF_N << 856 and,COND(<>) %r19, %r20, %r0 << 857 b,n intr_restore /* skip past i << 858 << 859 /* This check is critical to having LW << 860 * working. The IASQ is zero on the ga << 861 * page and we cannot deliver any sign << 862 * we get off the gateway page. << 863 * << 864 * Only do signals if we are returning << 865 */ << 866 LDREG PT_IASQ0(%r16), %r20 << 867 cmpib,COND(=),n LINUX_GATEWAY_SPACE, % << 868 LDREG PT_IASQ1(%r16), %r20 << 869 cmpib,COND(=),n LINUX_GATEWAY_SPACE, % << 870 << 871 copy %r0, %r25 << 872 #ifdef CONFIG_64BIT << 873 ldo -16(%r30),%r29 << 874 #endif << 875 << 876 /* NOTE: We need to enable interrupts << 877 * signals. We used to do this earlier << 878 * stack overflows. */ << 879 ssm PSW_SM_I, %r0 << 880 << 881 BL do_notify_resume,%r2 << 882 copy %r16, %r26 << 883 << 884 b,n intr_check_sig << 885 << 886 intr_restore: << 887 copy %r16,%r29 << 888 ldo PT_FR31(%r29),%r1 << 889 rest_fp %r1 << 890 rest_general %r29 << 891 << 892 /* inverse of virt_map */ << 893 pcxt_ssm_bug << 894 rsm PSW_SM_QUIET,%r0 << 895 tophys_r1 %r29 << 896 << 897 /* Restore space id's and special cr's << 898 * structure pointed to by r29 << 899 */ << 900 rest_specials %r29 << 901 << 902 /* IMPORTANT: rest_stack restores r29 << 903 * It also restores r1 and r30. << 904 */ << 905 rest_stack << 906 << 907 rfi << 908 nop << 909 << 910 #ifndef CONFIG_PREEMPTION << 911 # define intr_do_preempt intr_restore << 912 #endif /* !CONFIG_PREEMPTION */ << 913 << 914 .import schedule,code << 915 intr_do_resched: << 916 /* Only call schedule on return to use << 917 * to kernel space, we may schedule if << 918 * we jump back to intr_restore. << 919 */ << 920 LDREG PT_IASQ0(%r16), %r20 << 921 cmpib,COND(=) 0, %r20, intr_do_preem << 922 nop << 923 LDREG PT_IASQ1(%r16), %r20 << 924 cmpib,COND(=) 0, %r20, intr_do_preem << 925 nop << 926 << 927 /* NOTE: We need to enable interrupts << 928 * to do this earlier but it caused ke << 929 ssm PSW_SM_I, %r0 << 930 << 931 #ifdef CONFIG_64BIT << 932 ldo -16(%r30),%r29 /* Ref << 933 #endif << 934 << 935 ldil L%intr_check_sig, %r2 << 936 #ifndef CONFIG_64BIT << 937 b schedule << 938 #else << 939 load32 schedule, %r20 << 940 bv %r0(%r20) << 941 #endif << 942 ldo R%intr_check_sig(%r2), %r2 << 943 << 944 /* preempt the current task on returni << 945 * mode from an interrupt, iff need_re << 946 * and preempt_count is 0. otherwise, << 947 * our merry way back to the current r << 948 */ << 949 #ifdef CONFIG_PREEMPTION << 950 .import preempt_schedule_irq,code << 951 intr_do_preempt: << 952 rsm PSW_SM_I, %r0 /* dis << 953 << 954 /* current_thread_info()->preempt_coun << 955 mfctl %cr30, %r1 << 956 ldw TI_PRE_COUNT(%r1), %r19 << 957 cmpib,<> 0, %r19, intr_restore << 958 nop /* pre << 959 << 960 /* check if we interrupted a critical << 961 LDREG PT_PSW(%r16), %r20 << 962 bb,<,n %r20, 31 - PSW_SM_I, intr_rest << 963 nop << 964 << 965 /* ssm PSW_SM_I done later in intr_res << 966 #ifdef CONFIG_MLONGCALLS << 967 ldil L%intr_restore, %r2 << 968 load32 preempt_schedule_irq, %r1 << 969 bv %r0(%r1) << 970 ldo R%intr_restore(%r2), %r2 << 971 #else << 972 ldil L%intr_restore, %r1 << 973 BL preempt_schedule_irq, %r2 << 974 ldo R%intr_restore(%r1), %r2 << 975 #endif << 976 #endif /* CONFIG_PREEMPTION */ << 977 << 978 /* << 979 * External interrupts. << 980 */ << 981 << 982 intr_extint: << 983 cmpib,COND(=),n 0,%r16,1f << 984 << 985 get_stack_use_cr30 << 986 b,n 2f << 987 << 988 1: << 989 get_stack_use_r30 << 990 2: << 991 save_specials %r29 << 992 virt_map << 993 save_general %r29 << 994 << 995 ldo PT_FR0(%r29), %r24 << 996 save_fp %r24 << 997 << 998 loadgp << 999 << 1000 copy %r29, %r26 /* arg0 is pt << 1001 copy %r29, %r16 /* save pt_re << 1002 << 1003 ldil L%intr_return, %r2 << 1004 << 1005 #ifdef CONFIG_64BIT << 1006 ldo -16(%r30),%r29 /* Reference << 1007 #endif << 1008 << 1009 b do_cpu_irq_mask << 1010 ldo R%intr_return(%r2), %r2 /* re << 1011 ENDPROC_CFI(syscall_exit_rfi) << 1012 32 >> 33 #include <linux/linkage.h> >> 34 #include <asm/errno.h> >> 35 #include <asm/setup.h> >> 36 #include <asm/traps.h> >> 37 #include <asm/unistd.h> >> 38 #include <asm/asm-offsets.h> >> 39 #include <asm/entry.h> 1013 40 1014 /* Generic interruptions (illegal ins !! 41 .globl system_call, buserr, trap, resume 1015 !! 42 .globl sys_call_table 1016 ENTRY_CFI(intr_save) /* for os_hpm !! 43 .globl __sys_fork, __sys_clone, __sys_vfork 1017 mfsp %sr7,%r16 !! 44 .globl bad_interrupt 1018 cmpib,COND(=),n 0,%r16,1f !! 45 .globl auto_irqhandler_fixup 1019 get_stack_use_cr30 !! 46 .globl user_irqvec_fixup 1020 b 2f !! 47 1021 copy %r8,%r26 !! 48 .text >> 49 ENTRY(__sys_fork) >> 50 SAVE_SWITCH_STACK >> 51 jbsr sys_fork >> 52 lea %sp@(24),%sp >> 53 rts >> 54 >> 55 ENTRY(__sys_clone) >> 56 SAVE_SWITCH_STACK >> 57 pea %sp@(SWITCH_STACK_SIZE) >> 58 jbsr m68k_clone >> 59 lea %sp@(28),%sp >> 60 rts >> 61 >> 62 ENTRY(__sys_vfork) >> 63 SAVE_SWITCH_STACK >> 64 jbsr sys_vfork >> 65 lea %sp@(24),%sp >> 66 rts >> 67 >> 68 ENTRY(__sys_clone3) >> 69 SAVE_SWITCH_STACK >> 70 pea %sp@(SWITCH_STACK_SIZE) >> 71 jbsr m68k_clone3 >> 72 lea %sp@(28),%sp >> 73 rts >> 74 >> 75 ENTRY(sys_sigreturn) >> 76 SAVE_SWITCH_STACK >> 77 movel %sp,%a1 | switch_stack pointer >> 78 lea %sp@(SWITCH_STACK_SIZE),%a0 | pt_regs pointer >> 79 lea %sp@(-84),%sp | leave a gap >> 80 movel %a1,%sp@- >> 81 movel %a0,%sp@- >> 82 jbsr do_sigreturn >> 83 jra 1f | shared with rt_sigreturn() >> 84 >> 85 ENTRY(sys_rt_sigreturn) >> 86 SAVE_SWITCH_STACK >> 87 movel %sp,%a1 | switch_stack pointer >> 88 lea %sp@(SWITCH_STACK_SIZE),%a0 | pt_regs pointer >> 89 lea %sp@(-84),%sp | leave a gap >> 90 movel %a1,%sp@- >> 91 movel %a0,%sp@- >> 92 | stack contents: >> 93 | [original pt_regs address] [original switch_stack address] >> 94 | [gap] [switch_stack] [pt_regs] [exception frame] >> 95 jbsr do_rt_sigreturn 1022 96 1023 1: 97 1: 1024 get_stack_use_r30 !! 98 | stack contents now: 1025 copy %r8,%r26 !! 99 | [original pt_regs address] [original switch_stack address] 1026 !! 100 | [unused part of the gap] [moved switch_stack] [moved pt_regs] 1027 2: !! 101 | [replacement exception frame] 1028 save_specials %r29 !! 102 | return value of do_{rt_,}sigreturn() points to moved switch_stack. 1029 !! 103 1030 /* If this trap is a itlb miss, skip !! 104 movel %d0,%sp | discard the leftover junk 1031 cmpib,COND(=),n PARISC_ITLB_TR !! 105 RESTORE_SWITCH_STACK 1032 !! 106 | stack contents now is just [syscall return address] [pt_regs] [frame] 1033 !! 107 | return pt_regs.d0 1034 mfctl %isr, %r16 !! 108 movel %sp@(PT_OFF_D0+4),%d0 1035 nop /* serialize mfctl on !! 109 rts 1036 mfctl %ior, %r17 !! 110 1037 !! 111 ENTRY(buserr) 1038 !! 112 SAVE_ALL_INT 1039 #ifdef CONFIG_64BIT !! 113 GET_CURRENT(%d0) 1040 /* !! 114 movel %sp,%sp@- | stack frame pointer argument 1041 * If the interrupted code was runnin !! 115 jbsr buserr_c 1042 * clear the b bits (bits 0 & 1) in t !! 116 addql #4,%sp 1043 * save_specials left ipsw value in r !! 117 jra ret_from_exception 1044 */ !! 118 1045 extrd,u,*<> %r8,PSW_W_BIT,1,%r0 !! 119 ENTRY(trap) 1046 depdi 0,1,2,%r17 !! 120 SAVE_ALL_INT 1047 !! 121 GET_CURRENT(%d0) 1048 /* adjust isr/ior: get high bits from !! 122 movel %sp,%sp@- | stack frame pointer argument 1049 space_adjust %r16,%r17,%r1 !! 123 jbsr trap_c 1050 #endif !! 124 addql #4,%sp 1051 STREG %r16, PT_ISR(%r29) !! 125 jra ret_from_exception 1052 STREG %r17, PT_IOR(%r29) !! 126 1053 !! 127 | After a fork we jump here directly from resume, 1054 #if defined(CONFIG_64BIT) !! 128 | so that %d1 contains the previous task 1055 b,n intr_save2 !! 129 | schedule_tail now used regardless of CONFIG_SMP 1056 !! 130 ENTRY(ret_from_fork) 1057 skip_save_ior: !! 131 movel %d1,%sp@- 1058 /* We have a itlb miss, and when exec !! 132 jsr schedule_tail 1059 * need to adjust iasq/iaoq here in t !! 133 addql #4,%sp 1060 * above. !! 134 jra ret_from_exception 1061 */ << 1062 bb,COND(>=),n %r8,PSW_W_BIT,intr_sa << 1063 LDREG PT_IASQ0(%r29), %r16 << 1064 LDREG PT_IAOQ0(%r29), %r17 << 1065 /* adjust iasq/iaoq */ << 1066 space_adjust %r16,%r17,%r1 << 1067 STREG %r16, PT_IASQ0(%r29) << 1068 STREG %r17, PT_IAOQ0(%r29) << 1069 #else << 1070 skip_save_ior: << 1071 #endif << 1072 << 1073 intr_save2: << 1074 virt_map << 1075 save_general %r29 << 1076 << 1077 ldo PT_FR0(%r29), %r25 << 1078 save_fp %r25 << 1079 << 1080 loadgp << 1081 << 1082 copy %r29, %r25 /* ar << 1083 #ifdef CONFIG_64BIT << 1084 ldo -16(%r30),%r29 /* Re << 1085 #endif << 1086 << 1087 ldil L%intr_check_sig, %r2 << 1088 copy %r25, %r16 /* sa << 1089 << 1090 b handle_interruption << 1091 ldo R%intr_check_sig(%r2) << 1092 ENDPROC_CFI(intr_save) << 1093 << 1094 << 1095 /* << 1096 * Note for all tlb miss handlers: << 1097 * << 1098 * cr24 contains a pointer to the ker << 1099 * page directory. << 1100 * << 1101 * cr25 contains a pointer to the cur << 1102 * space page directory. << 1103 * << 1104 * sr3 will contain the space id of t << 1105 * of the current running thread whil << 1106 * running in the kernel. << 1107 */ << 1108 << 1109 /* << 1110 * register number allocations. Note << 1111 * in the shadowed registers << 1112 */ << 1113 << 1114 t0 = r1 /* temporary register << 1115 va = r8 /* virtual address fo << 1116 t1 = r9 /* temporary register << 1117 pte = r16 /* pte/phys page # */ << 1118 prot = r17 /* prot bits */ << 1119 spc = r24 /* space for which th << 1120 ptp = r25 /* page directory/pag << 1121 << 1122 #ifdef CONFIG_64BIT << 1123 << 1124 dtlb_miss_20w: << 1125 space_adjust spc,va,t0 << 1126 get_pgd spc,ptp << 1127 space_check spc,t0,dtlb_fault << 1128 << 1129 L3_ptep ptp,pte,t0,va,dtlb_ch << 1130 << 1131 ptl_lock spc,ptp,pte,t0,t1,dtl << 1132 update_accessed ptp,pte,t0,t1 << 1133 135 1134 make_insert_tlb spc,pte,prot,t1 !! 136 ENTRY(ret_from_kernel_thread) 1135 !! 137 | a3 contains the kernel thread payload, d7 - its argument 1136 idtlbt pte,prot !! 138 movel %d1,%sp@- 1137 !! 139 jsr schedule_tail 1138 ptl_unlock spc,t0,t1 !! 140 movel %d7,(%sp) 1139 rfir !! 141 jsr %a3@ 1140 nop !! 142 addql #4,%sp 1141 !! 143 jra ret_from_exception 1142 dtlb_check_alias_20w: !! 144 1143 do_alias spc,t0,t1,va,pte,prot !! 145 #if defined(CONFIG_COLDFIRE) || !defined(CONFIG_MMU) 1144 !! 146 1145 idtlbt pte,prot !! 147 #ifdef TRAP_DBG_INTERRUPT 1146 !! 148 1147 insert_nops NUM_PIPELINE_INSNS - !! 149 .globl dbginterrupt 1148 rfir !! 150 ENTRY(dbginterrupt) 1149 nop !! 151 SAVE_ALL_INT 1150 !! 152 GET_CURRENT(%d0) 1151 nadtlb_miss_20w: !! 153 movel %sp,%sp@- /* stack frame pointer argument */ 1152 space_adjust spc,va,t0 !! 154 jsr dbginterrupt_c 1153 get_pgd spc,ptp !! 155 addql #4,%sp 1154 space_check spc,t0,nadtlb_fault !! 156 jra ret_from_exception 1155 !! 157 #endif 1156 L3_ptep ptp,pte,t0,va,nadtlb_ !! 158 1157 !! 159 ENTRY(reschedule) 1158 ptl_lock spc,ptp,pte,t0,t1,nad !! 160 /* save top of frame */ 1159 update_accessed ptp,pte,t0,t1 !! 161 pea %sp@ 1160 !! 162 jbsr set_esp0 1161 make_insert_tlb spc,pte,prot,t1 !! 163 addql #4,%sp 1162 !! 164 pea ret_from_exception 1163 idtlbt pte,prot !! 165 jmp schedule 1164 !! 166 1165 ptl_unlock spc,t0,t1 !! 167 ENTRY(ret_from_user_signal) 1166 rfir !! 168 moveq #__NR_sigreturn,%d0 1167 nop !! 169 trap #0 1168 !! 170 1169 nadtlb_check_alias_20w: !! 171 ENTRY(ret_from_user_rt_signal) 1170 do_alias spc,t0,t1,va,pte,prot !! 172 movel #__NR_rt_sigreturn,%d0 1171 !! 173 trap #0 1172 idtlbt pte,prot !! 174 1173 !! 175 #else 1174 insert_nops NUM_PIPELINE_INSNS - !! 176 1175 rfir !! 177 do_trace_entry: 1176 nop !! 178 movel #-ENOSYS,%sp@(PT_OFF_D0)| needed for strace 1177 !! 179 subql #4,%sp 1178 #else !! 180 SAVE_SWITCH_STACK 1179 !! 181 jbsr syscall_trace_enter 1180 dtlb_miss_11: !! 182 RESTORE_SWITCH_STACK 1181 get_pgd spc,ptp !! 183 addql #4,%sp 1182 !! 184 addql #1,%d0 | optimization for cmpil #-1,%d0 1183 space_check spc,t0,dtlb_fault !! 185 jeq ret_from_syscall 1184 !! 186 movel %sp@(PT_OFF_ORIG_D0),%d0 1185 L2_ptep ptp,pte,t0,va,dtlb_ch !! 187 cmpl #NR_syscalls,%d0 1186 !! 188 jcs syscall 1187 ptl_lock spc,ptp,pte,t0,t1,dtl !! 189 jra ret_from_syscall 1188 update_accessed ptp,pte,t0,t1 !! 190 badsys: 1189 !! 191 movel #-ENOSYS,%sp@(PT_OFF_D0) 1190 make_insert_tlb_11 spc,pte,prot !! 192 jra ret_from_syscall 1191 !! 193 1192 mfsp %sr1,t1 /* Save sr1 !! 194 do_trace_exit: 1193 mtsp spc,%sr1 !! 195 subql #4,%sp 1194 !! 196 SAVE_SWITCH_STACK 1195 idtlba pte,(%sr1,va) !! 197 jbsr syscall_trace_leave 1196 idtlbp prot,(%sr1,va) !! 198 RESTORE_SWITCH_STACK 1197 !! 199 addql #4,%sp 1198 mtsp t1, %sr1 /* Re !! 200 jra .Lret_from_exception 1199 !! 201 1200 ptl_unlock spc,t0,t1 !! 202 ENTRY(system_call) 1201 rfir !! 203 SAVE_ALL_SYS 1202 nop !! 204 1203 !! 205 GET_CURRENT(%d1) 1204 dtlb_check_alias_11: !! 206 movel %d1,%a1 1205 do_alias spc,t0,t1,va,pte,prot !! 207 1206 !! 208 | save top of frame 1207 idtlba pte,(va) !! 209 movel %sp,%curptr@(TASK_THREAD+THREAD_ESP0) 1208 idtlbp prot,(va) !! 210 1209 !! 211 | syscall trace? 1210 insert_nops NUM_PIPELINE_INSNS - !! 212 tstb %a1@(TINFO_FLAGS+2) 1211 rfir !! 213 jmi do_trace_entry 1212 nop !! 214 | seccomp filter active? 1213 !! 215 btst #5,%a1@(TINFO_FLAGS+2) 1214 nadtlb_miss_11: !! 216 bnes do_trace_entry 1215 get_pgd spc,ptp !! 217 cmpl #NR_syscalls,%d0 1216 !! 218 jcc badsys 1217 space_check spc,t0,nadtlb_fault !! 219 syscall: 1218 !! 220 jbsr @(sys_call_table,%d0:l:4)@(0) 1219 L2_ptep ptp,pte,t0,va,nadtlb_ !! 221 movel %d0,%sp@(PT_OFF_D0) | save the return value 1220 !! 222 ret_from_syscall: 1221 ptl_lock spc,ptp,pte,t0,t1,nad !! 223 |oriw #0x0700,%sr 1222 update_accessed ptp,pte,t0,t1 !! 224 movel %curptr@(TASK_STACK),%a1 1223 !! 225 movew %a1@(TINFO_FLAGS+2),%d0 1224 make_insert_tlb_11 spc,pte,prot !! 226 jne syscall_exit_work 1225 !! 227 1: RESTORE_ALL 1226 mfsp %sr1,t1 /* Save sr1 !! 228 1227 mtsp spc,%sr1 !! 229 syscall_exit_work: 1228 !! 230 btst #5,%sp@(PT_OFF_SR) | check if returning to kernel 1229 idtlba pte,(%sr1,va) !! 231 bnes 1b | if so, skip resched, signals 1230 idtlbp prot,(%sr1,va) !! 232 lslw #1,%d0 1231 !! 233 jcs do_trace_exit 1232 mtsp t1, %sr1 /* Re !! 234 jmi do_delayed_trace 1233 !! 235 lslw #8,%d0 1234 ptl_unlock spc,t0,t1 !! 236 jne do_signal_return 1235 rfir !! 237 pea resume_userspace 1236 nop !! 238 jra schedule 1237 !! 239 1238 nadtlb_check_alias_11: !! 240 1239 do_alias spc,t0,t1,va,pte,prot !! 241 ENTRY(ret_from_exception) 1240 !! 242 .Lret_from_exception: 1241 idtlba pte,(va) !! 243 btst #5,%sp@(PT_OFF_SR) | check if returning to kernel 1242 idtlbp prot,(va) !! 244 bnes 1f | if so, skip resched, signals 1243 !! 245 | only allow interrupts when we are really the last one on the 1244 insert_nops NUM_PIPELINE_INSNS - !! 246 | kernel stack, otherwise stack overflow can occur during 1245 rfir !! 247 | heavy interrupt load 1246 nop !! 248 andw #ALLOWINT,%sr 1247 !! 249 1248 dtlb_miss_20: !! 250 resume_userspace: 1249 space_adjust spc,va,t0 !! 251 movel %curptr@(TASK_STACK),%a1 1250 get_pgd spc,ptp !! 252 moveb %a1@(TINFO_FLAGS+3),%d0 1251 space_check spc,t0,dtlb_fault !! 253 jne exit_work 1252 !! 254 1: RESTORE_ALL 1253 L2_ptep ptp,pte,t0,va,dtlb_ch !! 255 1254 !! 256 exit_work: 1255 ptl_lock spc,ptp,pte,t0,t1,dtl !! 257 | save top of frame 1256 update_accessed ptp,pte,t0,t1 !! 258 movel %sp,%curptr@(TASK_THREAD+THREAD_ESP0) 1257 !! 259 lslb #1,%d0 1258 make_insert_tlb spc,pte,prot,t1 !! 260 jne do_signal_return 1259 !! 261 pea resume_userspace 1260 f_extend pte,t1 !! 262 jra schedule 1261 !! 263 1262 idtlbt pte,prot !! 264 1263 !! 265 do_signal_return: 1264 ptl_unlock spc,t0,t1 !! 266 |andw #ALLOWINT,%sr 1265 rfir !! 267 subql #4,%sp | dummy return address 1266 nop !! 268 SAVE_SWITCH_STACK 1267 !! 269 pea %sp@(SWITCH_STACK_SIZE) 1268 dtlb_check_alias_20: !! 270 bsrl do_notify_resume 1269 do_alias spc,t0,t1,va,pte,prot !! 271 addql #4,%sp 1270 !! 272 RESTORE_SWITCH_STACK 1271 idtlbt pte,prot !! 273 addql #4,%sp 1272 !! 274 jbra resume_userspace 1273 insert_nops NUM_PIPELINE_INSNS - !! 275 1274 rfir !! 276 do_delayed_trace: 1275 nop !! 277 bclr #7,%sp@(PT_OFF_SR) | clear trace bit in SR 1276 !! 278 pea 1 | send SIGTRAP 1277 nadtlb_miss_20: !! 279 movel %curptr,%sp@- 1278 get_pgd spc,ptp !! 280 pea LSIGTRAP 1279 !! 281 jbsr send_sig 1280 space_check spc,t0,nadtlb_fault !! 282 addql #8,%sp 1281 !! 283 addql #4,%sp 1282 L2_ptep ptp,pte,t0,va,nadtlb_ !! 284 jbra resume_userspace 1283 !! 285 1284 ptl_lock spc,ptp,pte,t0,t1,nad !! 286 1285 update_accessed ptp,pte,t0,t1 !! 287 /* This is the main interrupt handler for autovector interrupts */ 1286 !! 288 1287 make_insert_tlb spc,pte,prot,t1 !! 289 ENTRY(auto_inthandler) 1288 !! 290 SAVE_ALL_INT 1289 f_extend pte,t1 !! 291 GET_CURRENT(%d0) 1290 !! 292 | put exception # in d0 1291 idtlbt pte,prot !! 293 bfextu %sp@(PT_OFF_FORMATVEC){#4,#10},%d0 1292 !! 294 subw #VEC_SPUR,%d0 1293 ptl_unlock spc,t0,t1 !! 295 1294 rfir !! 296 movel %sp,%sp@- 1295 nop !! 297 movel %d0,%sp@- | put vector # on stack 1296 !! 298 auto_irqhandler_fixup = . + 2 1297 nadtlb_check_alias_20: !! 299 jsr do_IRQ | process the IRQ 1298 do_alias spc,t0,t1,va,pte,prot !! 300 addql #8,%sp | pop parameters off stack 1299 !! 301 jra ret_from_exception 1300 idtlbt pte,prot !! 302 1301 !! 303 /* Handler for user defined interrupt vectors */ 1302 insert_nops NUM_PIPELINE_INSNS - !! 304 1303 rfir !! 305 ENTRY(user_inthandler) 1304 nop !! 306 SAVE_ALL_INT 1305 !! 307 GET_CURRENT(%d0) 1306 #endif !! 308 | put exception # in d0 1307 !! 309 bfextu %sp@(PT_OFF_FORMATVEC){#4,#10},%d0 1308 nadtlb_emulate: !! 310 user_irqvec_fixup = . + 2 1309 !! 311 subw #VEC_USER,%d0 1310 /* !! 312 1311 * Non-access misses can be caused by !! 313 movel %sp,%sp@- 1312 * probei instructions. The kernel no !! 314 movel %d0,%sp@- | put vector # on stack 1313 * Use of lpa and probe instructions !! 315 jsr do_IRQ | process the IRQ 1314 * with shadow registers, we defer ev !! 316 addql #8,%sp | pop parameters off stack 1315 */ !! 317 jra ret_from_exception 1316 b,n nadtlb_fault !! 318 1317 !! 319 /* Handler for uninitialized and spurious interrupts */ 1318 #ifdef CONFIG_64BIT !! 320 1319 itlb_miss_20w: !! 321 ENTRY(bad_inthandler) 1320 !! 322 SAVE_ALL_INT 1321 /* !! 323 GET_CURRENT(%d0) 1322 * I miss is a little different, sinc !! 324 1323 * on the gateway page which is in th !! 325 movel %sp,%sp@- 1324 */ !! 326 jsr handle_badint 1325 !! 327 addql #4,%sp 1326 space_adjust spc,va,t0 !! 328 jra ret_from_exception 1327 get_pgd spc,ptp !! 329 1328 space_check spc,t0,itlb_fault !! 330 resume: 1329 !! 331 /* 1330 L3_ptep ptp,pte,t0,va,itlb_fa !! 332 * Beware - when entering resume, prev (the current task) is 1331 !! 333 * in a0, next (the new task) is in a1,so don't change these 1332 ptl_lock spc,ptp,pte,t0,t1,itl !! 334 * registers until their contents are no longer needed. 1333 update_accessed ptp,pte,t0,t1 !! 335 */ 1334 !! 336 1335 make_insert_tlb spc,pte,prot,t1 !! 337 /* save sr */ 1336 !! 338 movew %sr,%a0@(TASK_THREAD+THREAD_SR) 1337 iitlbt pte,prot !! 339 1338 !! 340 /* save fs (sfc,%dfc) (may be pointing to kernel memory) */ 1339 ptl_unlock spc,t0,t1 !! 341 movec %sfc,%d0 1340 rfir !! 342 movew %d0,%a0@(TASK_THREAD+THREAD_FC) 1341 nop !! 343 1342 !! 344 /* save usp */ 1343 naitlb_miss_20w: !! 345 /* it is better to use a movel here instead of a movew 8*) */ 1344 !! 346 movec %usp,%d0 1345 /* !! 347 movel %d0,%a0@(TASK_THREAD+THREAD_USP) 1346 * I miss is a little different, sinc !! 348 1347 * on the gateway page which is in th !! 349 /* save non-scratch registers on stack */ 1348 */ !! 350 SAVE_SWITCH_STACK 1349 !! 351 1350 space_adjust spc,va,t0 !! 352 /* save current kernel stack pointer */ 1351 get_pgd spc,ptp !! 353 movel %sp,%a0@(TASK_THREAD+THREAD_KSP) 1352 space_check spc,t0,naitlb_fault !! 354 1353 !! 355 /* save floating point context */ 1354 L3_ptep ptp,pte,t0,va,naitlb_ !! 356 #ifndef CONFIG_M68KFPU_EMU_ONLY 1355 !! 357 #ifdef CONFIG_M68KFPU_EMU 1356 ptl_lock spc,ptp,pte,t0,t1,nai !! 358 tstl m68k_fputype 1357 update_accessed ptp,pte,t0,t1 !! 359 jeq 3f 1358 !! 360 #endif 1359 make_insert_tlb spc,pte,prot,t1 !! 361 fsave %a0@(TASK_THREAD+THREAD_FPSTATE) 1360 !! 362 1361 iitlbt pte,prot !! 363 #if defined(CONFIG_M68060) 1362 !! 364 #if !defined(CPU_M68060_ONLY) 1363 ptl_unlock spc,t0,t1 !! 365 btst #3,m68k_cputype+3 1364 rfir !! 366 beqs 1f 1365 nop !! 367 #endif 1366 !! 368 /* The 060 FPU keeps status in bits 15-8 of the first longword */ 1367 naitlb_check_alias_20w: !! 369 tstb %a0@(TASK_THREAD+THREAD_FPSTATE+2) 1368 do_alias spc,t0,t1,va,pte,prot !! 370 jeq 3f 1369 !! 371 #if !defined(CPU_M68060_ONLY) 1370 iitlbt pte,prot !! 372 jra 2f 1371 !! 373 #endif 1372 insert_nops NUM_PIPELINE_INSNS - !! 374 #endif /* CONFIG_M68060 */ 1373 rfir !! 375 #if !defined(CPU_M68060_ONLY) 1374 nop !! 376 1: tstb %a0@(TASK_THREAD+THREAD_FPSTATE) 1375 !! 377 jeq 3f 1376 #else << 1377 << 1378 itlb_miss_11: << 1379 get_pgd spc,ptp << 1380 << 1381 space_check spc,t0,itlb_fault << 1382 << 1383 L2_ptep ptp,pte,t0,va,itlb_fa << 1384 << 1385 ptl_lock spc,ptp,pte,t0,t1,itl << 1386 update_accessed ptp,pte,t0,t1 << 1387 << 1388 make_insert_tlb_11 spc,pte,prot << 1389 << 1390 mfsp %sr1,t1 /* Save sr1 << 1391 mtsp spc,%sr1 << 1392 << 1393 iitlba pte,(%sr1,va) << 1394 iitlbp prot,(%sr1,va) << 1395 << 1396 mtsp t1, %sr1 /* Re << 1397 << 1398 ptl_unlock spc,t0,t1 << 1399 rfir << 1400 nop << 1401 << 1402 naitlb_miss_11: << 1403 get_pgd spc,ptp << 1404 << 1405 space_check spc,t0,naitlb_fault << 1406 << 1407 L2_ptep ptp,pte,t0,va,naitlb_ << 1408 << 1409 ptl_lock spc,ptp,pte,t0,t1,nai << 1410 update_accessed ptp,pte,t0,t1 << 1411 << 1412 make_insert_tlb_11 spc,pte,prot << 1413 << 1414 mfsp %sr1,t1 /* Save sr1 << 1415 mtsp spc,%sr1 << 1416 << 1417 iitlba pte,(%sr1,va) << 1418 iitlbp prot,(%sr1,va) << 1419 << 1420 mtsp t1, %sr1 /* Re << 1421 << 1422 ptl_unlock spc,t0,t1 << 1423 rfir << 1424 nop << 1425 << 1426 naitlb_check_alias_11: << 1427 do_alias spc,t0,t1,va,pte,prot << 1428 << 1429 iitlba pte,(%sr0, va) << 1430 iitlbp prot,(%sr0, va) << 1431 << 1432 insert_nops NUM_PIPELINE_INSNS - << 1433 rfir << 1434 nop << 1435 << 1436 << 1437 itlb_miss_20: << 1438 get_pgd spc,ptp << 1439 << 1440 space_check spc,t0,itlb_fault << 1441 << 1442 L2_ptep ptp,pte,t0,va,itlb_fa << 1443 << 1444 ptl_lock spc,ptp,pte,t0,t1,itl << 1445 update_accessed ptp,pte,t0,t1 << 1446 << 1447 make_insert_tlb spc,pte,prot,t1 << 1448 << 1449 f_extend pte,t1 << 1450 << 1451 iitlbt pte,prot << 1452 << 1453 ptl_unlock spc,t0,t1 << 1454 rfir << 1455 nop << 1456 << 1457 naitlb_miss_20: << 1458 get_pgd spc,ptp << 1459 << 1460 space_check spc,t0,naitlb_fault << 1461 << 1462 L2_ptep ptp,pte,t0,va,naitlb_ << 1463 << 1464 ptl_lock spc,ptp,pte,t0,t1,nai << 1465 update_accessed ptp,pte,t0,t1 << 1466 << 1467 make_insert_tlb spc,pte,prot,t1 << 1468 << 1469 f_extend pte,t1 << 1470 << 1471 iitlbt pte,prot << 1472 << 1473 ptl_unlock spc,t0,t1 << 1474 rfir << 1475 nop << 1476 << 1477 naitlb_check_alias_20: << 1478 do_alias spc,t0,t1,va,pte,prot << 1479 << 1480 iitlbt pte,prot << 1481 << 1482 insert_nops NUM_PIPELINE_INSNS - << 1483 rfir << 1484 nop << 1485 << 1486 #endif << 1487 << 1488 #ifdef CONFIG_64BIT << 1489 << 1490 dbit_trap_20w: << 1491 space_adjust spc,va,t0 << 1492 get_pgd spc,ptp << 1493 space_check spc,t0,dbit_fault << 1494 << 1495 L3_ptep ptp,pte,t0,va,dbit_fa << 1496 << 1497 ptl_lock spc,ptp,pte,t0,t1,dbi << 1498 update_dirty ptp,pte,t1 << 1499 << 1500 make_insert_tlb spc,pte,prot,t1 << 1501 << 1502 idtlbt pte,prot << 1503 << 1504 ptl_unlock spc,t0,t1 << 1505 rfir << 1506 nop << 1507 #else << 1508 << 1509 dbit_trap_11: << 1510 << 1511 get_pgd spc,ptp << 1512 << 1513 space_check spc,t0,dbit_fault << 1514 << 1515 L2_ptep ptp,pte,t0,va,dbit_fa << 1516 << 1517 ptl_lock spc,ptp,pte,t0,t1,dbi << 1518 update_dirty ptp,pte,t1 << 1519 << 1520 make_insert_tlb_11 spc,pte,prot << 1521 << 1522 mfsp %sr1,t1 /* Save sr1 << 1523 mtsp spc,%sr1 << 1524 << 1525 idtlba pte,(%sr1,va) << 1526 idtlbp prot,(%sr1,va) << 1527 << 1528 mtsp t1, %sr1 /* Resto << 1529 << 1530 ptl_unlock spc,t0,t1 << 1531 rfir << 1532 nop << 1533 << 1534 dbit_trap_20: << 1535 get_pgd spc,ptp << 1536 << 1537 space_check spc,t0,dbit_fault << 1538 << 1539 L2_ptep ptp,pte,t0,va,dbit_fa << 1540 << 1541 ptl_lock spc,ptp,pte,t0,t1,dbi << 1542 update_dirty ptp,pte,t1 << 1543 << 1544 make_insert_tlb spc,pte,prot,t1 << 1545 << 1546 f_extend pte,t1 << 1547 << 1548 idtlbt pte,prot << 1549 << 1550 ptl_unlock spc,t0,t1 << 1551 rfir << 1552 nop << 1553 #endif << 1554 << 1555 .import handle_interruption,code << 1556 << 1557 kernel_bad_space: << 1558 b intr_save << 1559 ldi 31,%r8 /* Use an unu << 1560 << 1561 dbit_fault: << 1562 b intr_save << 1563 ldi 20,%r8 << 1564 << 1565 itlb_fault: << 1566 b intr_save << 1567 ldi PARISC_ITLB_TRAP,%r8 << 1568 << 1569 nadtlb_fault: << 1570 b intr_save << 1571 ldi 17,%r8 << 1572 << 1573 naitlb_fault: << 1574 b intr_save << 1575 ldi 16,%r8 << 1576 << 1577 dtlb_fault: << 1578 b intr_save << 1579 ldi 15,%r8 << 1580 << 1581 /* Register saving semantics for syst << 1582 << 1583 %r1 clobbered by syste << 1584 %r2 saved in PT_REGS b << 1585 %r3 - %r18 preserved by C cod << 1586 %r19 - %r20 saved in PT_REGS b << 1587 %r21 - %r22 non-standard sysca << 1588 stored in kernel s << 1589 %r23 - %r26 arg3-arg0, saved i << 1590 %r27 - %r30 saved in PT_REGS b << 1591 %r31 syscall return poi << 1592 */ << 1593 << 1594 /* Floating point registers (FIXME: w << 1595 << 1596 %fr0 - %fr3 status/exception, << 1597 %fr4 - %fr7 arguments << 1598 %fr8 - %fr11 not preserved by C << 1599 %fr12 - %fr21 preserved by C cod << 1600 %fr22 - %fr31 not preserved by C << 1601 */ << 1602 << 1603 .macro reg_save regs << 1604 STREG %r3, PT_GR3(\regs) << 1605 STREG %r4, PT_GR4(\regs) << 1606 STREG %r5, PT_GR5(\regs) << 1607 STREG %r6, PT_GR6(\regs) << 1608 STREG %r7, PT_GR7(\regs) << 1609 STREG %r8, PT_GR8(\regs) << 1610 STREG %r9, PT_GR9(\regs) << 1611 STREG %r10,PT_GR10(\regs) << 1612 STREG %r11,PT_GR11(\regs) << 1613 STREG %r12,PT_GR12(\regs) << 1614 STREG %r13,PT_GR13(\regs) << 1615 STREG %r14,PT_GR14(\regs) << 1616 STREG %r15,PT_GR15(\regs) << 1617 STREG %r16,PT_GR16(\regs) << 1618 STREG %r17,PT_GR17(\regs) << 1619 STREG %r18,PT_GR18(\regs) << 1620 .endm << 1621 << 1622 .macro reg_restore regs << 1623 LDREG PT_GR3(\regs), %r3 << 1624 LDREG PT_GR4(\regs), %r4 << 1625 LDREG PT_GR5(\regs), %r5 << 1626 LDREG PT_GR6(\regs), %r6 << 1627 LDREG PT_GR7(\regs), %r7 << 1628 LDREG PT_GR8(\regs), %r8 << 1629 LDREG PT_GR9(\regs), %r9 << 1630 LDREG PT_GR10(\regs),%r10 << 1631 LDREG PT_GR11(\regs),%r11 << 1632 LDREG PT_GR12(\regs),%r12 << 1633 LDREG PT_GR13(\regs),%r13 << 1634 LDREG PT_GR14(\regs),%r14 << 1635 LDREG PT_GR15(\regs),%r15 << 1636 LDREG PT_GR16(\regs),%r16 << 1637 LDREG PT_GR17(\regs),%r17 << 1638 LDREG PT_GR18(\regs),%r18 << 1639 .endm << 1640 << 1641 .macro fork_like name << 1642 ENTRY_CFI(sys_\name\()_wrapper) << 1643 mfctl %cr30,%r1 << 1644 ldo TASK_REGS(%r1),%r1 << 1645 reg_save %r1 << 1646 mfctl %cr27, %r28 << 1647 ldil L%sys_\name, %r31 << 1648 be R%sys_\name(%sr4,%r31) << 1649 STREG %r28, PT_CR27(%r1) << 1650 ENDPROC_CFI(sys_\name\()_wrapper) << 1651 .endm << 1652 << 1653 fork_like clone << 1654 fork_like clone3 << 1655 fork_like fork << 1656 fork_like vfork << 1657 << 1658 /* Set the return value for the child << 1659 ENTRY(child_return) << 1660 BL schedule_tail, %r2 << 1661 nop << 1662 finish_child_return: << 1663 mfctl %cr30,%r1 << 1664 ldo TASK_REGS(%r1),%r1 /* g << 1665 << 1666 LDREG PT_CR27(%r1), %r3 << 1667 mtctl %r3, %cr27 << 1668 reg_restore %r1 << 1669 b syscall_exit << 1670 copy %r0,%r28 << 1671 END(child_return) << 1672 << 1673 ENTRY_CFI(sys_rt_sigreturn_wrapper) << 1674 mfctl %cr30,%r26 << 1675 ldo TASK_REGS(%r26),%r26 /* ge << 1676 /* Don't save regs, we are going to r << 1677 STREG %r2, -RP_OFFSET(%r30) << 1678 #ifdef CONFIG_64BIT << 1679 ldo FRAME_SIZE(%r30), %r30 << 1680 BL sys_rt_sigreturn,%r2 << 1681 ldo -16(%r30),%r29 /* Re << 1682 #else << 1683 BL sys_rt_sigreturn,%r2 << 1684 ldo FRAME_SIZE(%r30), %r30 << 1685 #endif << 1686 << 1687 ldo -FRAME_SIZE(%r30), %r30 << 1688 LDREG -RP_OFFSET(%r30), %r2 << 1689 << 1690 /* FIXME: I think we need to restore << 1691 mfctl %cr30,%r1 << 1692 ldo TASK_REGS(%r1),%r1 /* ge << 1693 reg_restore %r1 << 1694 << 1695 /* If the signal was received while t << 1696 * syscall, then r2 will take us to s << 1697 * take us to syscall_exit_rfi and on << 1698 */ << 1699 bv %r0(%r2) << 1700 LDREG PT_GR28(%r1),%r28 /* reload << 1701 ENDPROC_CFI(sys_rt_sigreturn_wrapper) << 1702 << 1703 ENTRY(syscall_exit) << 1704 /* NOTE: Not all syscalls exit this w << 1705 * via syscall_exit_rfi if the signal << 1706 * was running. << 1707 */ << 1708 << 1709 /* save return value now */ << 1710 mfctl %cr30, %r1 << 1711 STREG %r28,TASK_PT_GR28(%r1) << 1712 << 1713 /* Seems to me that dp could be wrong << 1714 * calling a module, and nothing got << 1715 */ << 1716 loadgp << 1717 << 1718 syscall_check_resched: << 1719 << 1720 /* check for reschedule */ << 1721 mfctl %cr30,%r19 << 1722 LDREG TASK_TI_FLAGS(%r19),%r19 << 1723 bb,<,n %r19, 31-TIF_NEED_RESCHED, sy << 1724 << 1725 .import do_signal,code << 1726 syscall_check_sig: << 1727 mfctl %cr30,%r19 << 1728 LDREG TASK_TI_FLAGS(%r19),%r19 << 1729 ldi (_TIF_USER_WORK_MASK & ~_TIF_ << 1730 and,COND(<>) %r19, %r26, %r0 << 1731 b,n syscall_restore /* skip past << 1732 << 1733 syscall_do_signal: << 1734 /* Save callee-save registers (for si << 1735 * FIXME: After this point the proces << 1736 * consistent with all the relevant s << 1737 * before the syscall. We need to ve << 1738 */ << 1739 mfctl %cr30,%r1 << 1740 ldo TASK_REGS(%r1), %r26 << 1741 reg_save %r26 << 1742 << 1743 #ifdef CONFIG_64BIT << 1744 ldo -16(%r30),%r29 << 1745 #endif << 1746 << 1747 BL do_notify_resume,%r2 << 1748 ldi 1, %r25 << 1749 << 1750 mfctl %cr30,%r1 << 1751 ldo TASK_REGS(%r1), %r20 << 1752 reg_restore %r20 << 1753 << 1754 b,n syscall_check_sig << 1755 << 1756 syscall_restore: << 1757 mfctl %cr30,%r1 << 1758 << 1759 /* Are we being ptraced? */ << 1760 LDREG TASK_TI_FLAGS(%r1),%r19 << 1761 ldi _TIF_SINGLESTEP|_TIF_BLOCKSTE << 1762 and,COND(=) %r19,%r2,%r0 << 1763 b,n syscall_restore_rfi << 1764 << 1765 ldo TASK_PT_FR31(%r1),%r19 << 1766 rest_fp %r19 << 1767 << 1768 LDREG TASK_PT_SAR(%r1),%r19 << 1769 mtsar %r19 << 1770 << 1771 LDREG TASK_PT_GR2(%r1),%r2 << 1772 LDREG TASK_PT_GR19(%r1),%r19 << 1773 LDREG TASK_PT_GR20(%r1),%r20 << 1774 LDREG TASK_PT_GR21(%r1),%r21 << 1775 LDREG TASK_PT_GR22(%r1),%r22 << 1776 LDREG TASK_PT_GR23(%r1),%r23 << 1777 LDREG TASK_PT_GR24(%r1),%r24 << 1778 LDREG TASK_PT_GR25(%r1),%r25 << 1779 LDREG TASK_PT_GR26(%r1),%r26 << 1780 LDREG TASK_PT_GR27(%r1),%r27 /* << 1781 LDREG TASK_PT_GR28(%r1),%r28 /* << 1782 LDREG TASK_PT_GR29(%r1),%r29 << 1783 LDREG TASK_PT_GR31(%r1),%r31 /* << 1784 << 1785 /* NOTE: We use rsm/ssm pair to make << 1786 LDREG TASK_PT_GR30(%r1),%r1 << 1787 rsm PSW_SM_I, %r0 << 1788 copy %r1,%r30 << 1789 mfsp %sr3,%r1 << 1790 mtsp %r1,%sr7 << 1791 ssm PSW_SM_I, %r0 << 1792 << 1793 /* Set sr2 to zero for userspace sysc << 1794 mtsp %r0,%sr2 << 1795 mtsp %r1,%sr4 << 1796 mtsp %r1,%sr5 << 1797 mtsp %r1,%sr6 << 1798 << 1799 depi PRIV_USER,31,2,%r31 /* en << 1800 << 1801 #ifdef CONFIG_64BIT << 1802 /* decide whether to reset the wide m << 1803 * << 1804 * For a syscall, the W bit is stored << 1805 * of sp. Extract it and reset W if << 1806 extrd,u,*<> %r30,63,1,%r1 << 1807 rsm PSW_SM_W, %r0 << 1808 /* now reset the lowest bit of sp if << 1809 xor %r30,%r1,%r30 << 1810 #endif << 1811 be,n 0(%sr3,%r31) << 1812 << 1813 /* We have to return via an RFI, so t << 1814 * appropriately. << 1815 * This sets up pt_regs so we can ret << 1816 * the most efficient way of doing th << 1817 */ << 1818 syscall_restore_rfi: << 1819 ldo -1(%r0),%r2 << 1820 mtctl %r2,%cr0 << 1821 LDREG TASK_PT_PSW(%r1),%r2 << 1822 ldi 0x0b,%r20 << 1823 depi -1,13,1,%r20 << 1824 << 1825 /* The values of SINGLESTEP_BIT and B << 1826 * set in thread_info.h and converted << 1827 * numbers in asm-offsets.c */ << 1828 << 1829 /* if ((%r19.SINGLESTEP_BIT)) { %r20. << 1830 extru,= %r19,TIF_SINGLESTEP_PA_BIT,1, << 1831 depi -1,27,1,%r20 << 1832 << 1833 /* if ((%r19.BLOCKSTEP_BIT)) { %r20.7 << 1834 extru,= %r19,TIF_BLOCKSTEP_PA_BIT,1,% << 1835 depi -1,7,1,%r20 << 1836 << 1837 STREG %r20,TASK_PT_PSW(%r1) << 1838 << 1839 /* Always store space registers, sinc << 1840 << 1841 mfsp %sr3,%r25 << 1842 STREG %r25,TASK_PT_SR3(%r1) << 1843 STREG %r25,TASK_PT_SR4(%r1) << 1844 STREG %r25,TASK_PT_SR5(%r1) << 1845 STREG %r25,TASK_PT_SR6(%r1) << 1846 STREG %r25,TASK_PT_SR7(%r1) << 1847 STREG %r25,TASK_PT_IASQ0(%r1) << 1848 STREG %r25,TASK_PT_IASQ1(%r1) << 1849 << 1850 /* XXX W bit??? */ << 1851 /* Now if old D bit is clear, it mean << 1852 * on syscall entry, so do that now. << 1853 * calls, or if someone attached to u << 1854 * We could make this more efficient << 1855 * then we wouldn't be able to use th << 1856 * It is only for traced processes an << 1857 * an issue. << 1858 */ << 1859 bb,< %r2,30,pt_regs_ok << 1860 ldo TASK_REGS(%r1),%r25 << 1861 reg_save %r25 << 1862 << 1863 /* Save the current sr */ << 1864 mfsp %sr0,%r2 << 1865 STREG %r2,TASK_PT_SR0(%r1) << 1866 << 1867 /* Save the scratch sr */ << 1868 mfsp %sr1,%r2 << 1869 STREG %r2,TASK_PT_SR1(%r1) << 1870 << 1871 /* sr2 should be set to zero for user << 1872 STREG %r0,TASK_PT_SR2(%r1) << 1873 << 1874 LDREG TASK_PT_GR31(%r1),%r2 << 1875 depi PRIV_USER,31,2,%r2 /* en << 1876 STREG %r2,TASK_PT_IAOQ0(%r1) << 1877 ldo 4(%r2),%r2 << 1878 STREG %r2,TASK_PT_IAOQ1(%r1) << 1879 b intr_restore << 1880 copy %r25,%r16 << 1881 << 1882 pt_regs_ok: << 1883 LDREG TASK_PT_IAOQ0(%r1),%r2 << 1884 depi PRIV_USER,31,2,%r2 /* en << 1885 STREG %r2,TASK_PT_IAOQ0(%r1) << 1886 LDREG TASK_PT_IAOQ1(%r1),%r2 << 1887 depi PRIV_USER,31,2,%r2 << 1888 STREG %r2,TASK_PT_IAOQ1(%r1) << 1889 b intr_restore << 1890 copy %r25,%r16 << 1891 << 1892 syscall_do_resched: << 1893 load32 syscall_check_resched,%r2 /* << 1894 load32 schedule,%r19 << 1895 bv %r0(%r19) /* ju << 1896 #ifdef CONFIG_64BIT << 1897 ldo -16(%r30),%r29 /* Re << 1898 #else << 1899 nop << 1900 #endif << 1901 END(syscall_exit) << 1902 << 1903 << 1904 #ifdef CONFIG_FUNCTION_TRACER << 1905 << 1906 .import ftrace_function_trampoline,co << 1907 .align L1_CACHE_BYTES << 1908 ENTRY_CFI(mcount, caller) << 1909 _mcount: << 1910 .export _mcount,data << 1911 /* << 1912 * The 64bit mcount() function pointe << 1913 * first two are free. We optimize i << 1914 * calling mcount(), and 2 instructio << 1915 * have all on one L1 cacheline. << 1916 */ << 1917 ldi 0, %arg3 << 1918 b ftrace_function_trampoline << 1919 copy %r3, %arg2 /* caller ori << 1920 ftrace_stub: << 1921 .globl ftrace_stub << 1922 .type ftrace_stub, @function << 1923 #ifdef CONFIG_64BIT << 1924 bve (%rp) << 1925 #else << 1926 bv %r0(%rp) << 1927 #endif << 1928 nop << 1929 #ifdef CONFIG_64BIT << 1930 .dword mcount << 1931 .dword 0 /* code in head.S puts value << 1932 #endif << 1933 ENDPROC_CFI(mcount) << 1934 << 1935 #ifdef CONFIG_DYNAMIC_FTRACE << 1936 << 1937 #ifdef CONFIG_64BIT << 1938 #define FTRACE_FRAME_SIZE (2*FRAME_SIZE) << 1939 #else << 1940 #define FTRACE_FRAME_SIZE FRAME_SIZE << 1941 #endif << 1942 ENTRY_CFI(ftrace_caller, caller,frame=FTRACE_ << 1943 ftrace_caller: << 1944 .global ftrace_caller << 1945 << 1946 STREG %r3, -FTRACE_FRAME_SIZE+1*REG << 1947 ldo -FTRACE_FRAME_SIZE(%sp), %r3 << 1948 STREG %rp, -RP_OFFSET(%r3) << 1949 << 1950 /* Offset 0 is already allocated for << 1951 STREG %r23, 2*REG_SZ(%r3) << 1952 STREG %r24, 3*REG_SZ(%r3) << 1953 STREG %r25, 4*REG_SZ(%r3) << 1954 STREG %r26, 5*REG_SZ(%r3) << 1955 STREG %r28, 6*REG_SZ(%r3) << 1956 STREG %r29, 7*REG_SZ(%r3) << 1957 #ifdef CONFIG_64BIT << 1958 STREG %r19, 8*REG_SZ(%r3) << 1959 STREG %r20, 9*REG_SZ(%r3) << 1960 STREG %r21, 10*REG_SZ(%r3) << 1961 STREG %r22, 11*REG_SZ(%r3) << 1962 STREG %r27, 12*REG_SZ(%r3) << 1963 STREG %r31, 13*REG_SZ(%r3) << 1964 loadgp << 1965 ldo -16(%sp),%r29 << 1966 #endif << 1967 LDREG 0(%r3), %r25 << 1968 copy %rp, %r26 << 1969 ldo -8(%r25), %r25 << 1970 ldi 0, %r23 /* no pt_regs << 1971 b,l ftrace_function_trampoline, % << 1972 copy %r3, %r24 << 1973 << 1974 LDREG -RP_OFFSET(%r3), %rp << 1975 LDREG 2*REG_SZ(%r3), %r23 << 1976 LDREG 3*REG_SZ(%r3), %r24 << 1977 LDREG 4*REG_SZ(%r3), %r25 << 1978 LDREG 5*REG_SZ(%r3), %r26 << 1979 LDREG 6*REG_SZ(%r3), %r28 << 1980 LDREG 7*REG_SZ(%r3), %r29 << 1981 #ifdef CONFIG_64BIT << 1982 LDREG 8*REG_SZ(%r3), %r19 << 1983 LDREG 9*REG_SZ(%r3), %r20 << 1984 LDREG 10*REG_SZ(%r3), %r21 << 1985 LDREG 11*REG_SZ(%r3), %r22 << 1986 LDREG 12*REG_SZ(%r3), %r27 << 1987 LDREG 13*REG_SZ(%r3), %r31 << 1988 #endif << 1989 LDREG 1*REG_SZ(%r3), %r3 << 1990 << 1991 LDREGM -FTRACE_FRAME_SIZE(%sp), %r1 << 1992 /* Adjust return point to jump back t << 1993 ldo -4(%r1), %r1 << 1994 bv,n (%r1) << 1995 << 1996 ENDPROC_CFI(ftrace_caller) << 1997 << 1998 #ifdef CONFIG_HAVE_DYNAMIC_FTRACE_WITH_REGS << 1999 ENTRY_CFI(ftrace_regs_caller,caller,frame=FTR << 2000 CALLS,SAVE_RP,SAVE_SP) << 2001 ftrace_regs_caller: << 2002 .global ftrace_regs_caller << 2003 << 2004 ldo -FTRACE_FRAME_SIZE(%sp), %r1 << 2005 STREG %rp, -RP_OFFSET(%r1) << 2006 << 2007 copy %sp, %r1 << 2008 ldo PT_SZ_ALGN(%sp), %sp << 2009 << 2010 STREG %rp, PT_GR2(%r1) << 2011 STREG %r3, PT_GR3(%r1) << 2012 STREG %r4, PT_GR4(%r1) << 2013 STREG %r5, PT_GR5(%r1) << 2014 STREG %r6, PT_GR6(%r1) << 2015 STREG %r7, PT_GR7(%r1) << 2016 STREG %r8, PT_GR8(%r1) << 2017 STREG %r9, PT_GR9(%r1) << 2018 STREG %r10, PT_GR10(%r1) << 2019 STREG %r11, PT_GR11(%r1) << 2020 STREG %r12, PT_GR12(%r1) << 2021 STREG %r13, PT_GR13(%r1) << 2022 STREG %r14, PT_GR14(%r1) << 2023 STREG %r15, PT_GR15(%r1) << 2024 STREG %r16, PT_GR16(%r1) << 2025 STREG %r17, PT_GR17(%r1) << 2026 STREG %r18, PT_GR18(%r1) << 2027 STREG %r19, PT_GR19(%r1) << 2028 STREG %r20, PT_GR20(%r1) << 2029 STREG %r21, PT_GR21(%r1) << 2030 STREG %r22, PT_GR22(%r1) << 2031 STREG %r23, PT_GR23(%r1) << 2032 STREG %r24, PT_GR24(%r1) << 2033 STREG %r25, PT_GR25(%r1) << 2034 STREG %r26, PT_GR26(%r1) << 2035 STREG %r27, PT_GR27(%r1) << 2036 STREG %r28, PT_GR28(%r1) << 2037 STREG %r29, PT_GR29(%r1) << 2038 STREG %r30, PT_GR30(%r1) << 2039 STREG %r31, PT_GR31(%r1) << 2040 mfctl %cr11, %r26 << 2041 STREG %r26, PT_SAR(%r1) << 2042 << 2043 copy %rp, %r26 << 2044 LDREG -FTRACE_FRAME_SIZE-PT_SZ_ALGN << 2045 ldo -8(%r25), %r25 << 2046 ldo -FTRACE_FRAME_SIZE(%r1), %arg << 2047 b,l ftrace_function_trampoline, % << 2048 copy %r1, %arg3 /* struct pt_regs << 2049 << 2050 ldo -PT_SZ_ALGN(%sp), %r1 << 2051 << 2052 LDREG PT_SAR(%r1), %rp << 2053 mtctl %rp, %cr11 << 2054 << 2055 LDREG PT_GR2(%r1), %rp << 2056 LDREG PT_GR3(%r1), %r3 << 2057 LDREG PT_GR4(%r1), %r4 << 2058 LDREG PT_GR5(%r1), %r5 << 2059 LDREG PT_GR6(%r1), %r6 << 2060 LDREG PT_GR7(%r1), %r7 << 2061 LDREG PT_GR8(%r1), %r8 << 2062 LDREG PT_GR9(%r1), %r9 << 2063 LDREG PT_GR10(%r1),%r10 << 2064 LDREG PT_GR11(%r1),%r11 << 2065 LDREG PT_GR12(%r1),%r12 << 2066 LDREG PT_GR13(%r1),%r13 << 2067 LDREG PT_GR14(%r1),%r14 << 2068 LDREG PT_GR15(%r1),%r15 << 2069 LDREG PT_GR16(%r1),%r16 << 2070 LDREG PT_GR17(%r1),%r17 << 2071 LDREG PT_GR18(%r1),%r18 << 2072 LDREG PT_GR19(%r1),%r19 << 2073 LDREG PT_GR20(%r1),%r20 << 2074 LDREG PT_GR21(%r1),%r21 << 2075 LDREG PT_GR22(%r1),%r22 << 2076 LDREG PT_GR23(%r1),%r23 << 2077 LDREG PT_GR24(%r1),%r24 << 2078 LDREG PT_GR25(%r1),%r25 << 2079 LDREG PT_GR26(%r1),%r26 << 2080 LDREG PT_GR27(%r1),%r27 << 2081 LDREG PT_GR28(%r1),%r28 << 2082 LDREG PT_GR29(%r1),%r29 << 2083 LDREG PT_GR30(%r1),%r30 << 2084 LDREG PT_GR31(%r1),%r31 << 2085 << 2086 ldo -PT_SZ_ALGN(%sp), %sp << 2087 LDREGM -FTRACE_FRAME_SIZE(%sp), %r1 << 2088 /* Adjust return point to jump back t << 2089 ldo -4(%r1), %r1 << 2090 bv,n (%r1) << 2091 << 2092 ENDPROC_CFI(ftrace_regs_caller) << 2093 << 2094 #endif << 2095 #endif << 2096 << 2097 #ifdef CONFIG_FUNCTION_GRAPH_TRACER << 2098 .align 8 << 2099 ENTRY_CFI(return_to_handler, caller,frame=FRA << 2100 .export parisc_return_to_handler,data << 2101 parisc_return_to_handler: << 2102 copy %r3,%r1 << 2103 STREG %r0,-RP_OFFSET(%sp) /* st << 2104 copy %sp,%r3 << 2105 STREGM %r1,FRAME_SIZE(%sp) << 2106 STREG %ret0,8(%r3) << 2107 STREG %ret1,16(%r3) << 2108 << 2109 #ifdef CONFIG_64BIT << 2110 loadgp << 2111 #endif << 2112 << 2113 /* call ftrace_return_to_handler(0) * << 2114 .import ftrace_return_to_handler,code << 2115 load32 ftrace_return_to_handler,%ret0 << 2116 load32 .Lftrace_ret,%r2 << 2117 #ifdef CONFIG_64BIT << 2118 ldo -16(%sp),%ret1 /* Re << 2119 bve (%ret0) << 2120 #else << 2121 bv %r0(%ret0) << 2122 #endif << 2123 ldi 0,%r26 << 2124 .Lftrace_ret: << 2125 copy %ret0,%rp << 2126 << 2127 /* restore original return values */ << 2128 LDREG 8(%r3),%ret0 << 2129 LDREG 16(%r3),%ret1 << 2130 << 2131 /* return from function */ << 2132 #ifdef CONFIG_64BIT << 2133 bve (%rp) << 2134 #else << 2135 bv %r0(%rp) << 2136 #endif 378 #endif 2137 LDREGM -FRAME_SIZE(%sp),%r3 !! 379 2: fmovemx %fp0-%fp7,%a0@(TASK_THREAD+THREAD_FPREG) 2138 ENDPROC_CFI(return_to_handler) !! 380 fmoveml %fpcr/%fpsr/%fpiar,%a0@(TASK_THREAD+THREAD_FPCNTL) 2139 !! 381 3: 2140 #endif /* CONFIG_FUNCTION_GRAPH_TRACER */ !! 382 #endif /* CONFIG_M68KFPU_EMU_ONLY */ 2141 !! 383 /* Return previous task in %d1 */ 2142 #endif /* CONFIG_FUNCTION_TRACER */ !! 384 movel %curptr,%d1 2143 !! 385 2144 #ifdef CONFIG_IRQSTACKS !! 386 /* switch to new task (a1 contains new task) */ 2145 /* void call_on_stack(unsigned long param1, v !! 387 movel %a1,%curptr 2146 unsigned long new_stack !! 388 2147 ENTRY_CFI(call_on_stack, FRAME=2*FRAME_SIZE,C !! 389 /* restore floating point context */ 2148 ENTRY(_call_on_stack) !! 390 #ifndef CONFIG_M68KFPU_EMU_ONLY 2149 copy %sp, %r1 !! 391 #ifdef CONFIG_M68KFPU_EMU 2150 !! 392 tstl m68k_fputype 2151 /* Regarding the HPPA calling convent !! 393 jeq 4f 2152 we assume the PIC register is not !! 394 #endif 2153 CONFIG_64BIT, the argument pointer !! 395 #if defined(CONFIG_M68060) 2154 argument region allocated for the !! 396 #if !defined(CPU_M68060_ONLY) 2155 !! 397 btst #3,m68k_cputype+3 2156 /* Switch to new stack. We allocate !! 398 beqs 1f 2157 ldo 2*FRAME_SIZE(%arg2), %sp !! 399 #endif 2158 # ifdef CONFIG_64BIT !! 400 /* The 060 FPU keeps status in bits 15-8 of the first longword */ 2159 /* Save previous stack pointer and re !! 401 tstb %a1@(TASK_THREAD+THREAD_FPSTATE+2) 2160 STREG %rp, -FRAME_SIZE-RP_OFFSET(%s !! 402 jeq 3f 2161 /* Calls always use function descript !! 403 #if !defined(CPU_M68060_ONLY) 2162 LDREG 16(%arg1), %arg1 !! 404 jra 2f 2163 bve,l (%arg1), %rp !! 405 #endif 2164 STREG %r1, -FRAME_SIZE-REG_SZ(%sp) !! 406 #endif /* CONFIG_M68060 */ 2165 LDREG -FRAME_SIZE-RP_OFFSET(%sp), % !! 407 #if !defined(CPU_M68060_ONLY) 2166 bve (%rp) !! 408 1: tstb %a1@(TASK_THREAD+THREAD_FPSTATE) 2167 LDREG -FRAME_SIZE-REG_SZ(%sp), %sp !! 409 jeq 3f 2168 # else !! 410 #endif 2169 /* Save previous stack pointer and re !! 411 2: fmovemx %a1@(TASK_THREAD+THREAD_FPREG),%fp0-%fp7 2170 STREG %r1, -FRAME_SIZE-REG_SZ(%sp) !! 412 fmoveml %a1@(TASK_THREAD+THREAD_FPCNTL),%fpcr/%fpsr/%fpiar 2171 STREG %rp, -FRAME_SIZE-RP_OFFSET(%s !! 413 3: frestore %a1@(TASK_THREAD+THREAD_FPSTATE) 2172 /* Calls use function descriptor if P !! 414 4: 2173 bb,>=,n %arg1, 30, 1f !! 415 #endif /* CONFIG_M68KFPU_EMU_ONLY */ 2174 depwi 0,31,2, %arg1 !! 416 2175 LDREG 0(%arg1), %arg1 !! 417 /* restore the kernel stack pointer */ 2176 1: !! 418 movel %a1@(TASK_THREAD+THREAD_KSP),%sp 2177 be,l 0(%sr4,%arg1), %sr0, %r31 !! 419 2178 copy %r31, %rp !! 420 /* restore non-scratch registers */ 2179 LDREG -FRAME_SIZE-RP_OFFSET(%sp), % !! 421 RESTORE_SWITCH_STACK 2180 bv (%rp) !! 422 2181 LDREG -FRAME_SIZE-REG_SZ(%sp), %sp !! 423 /* restore user stack pointer */ 2182 # endif /* CONFIG_64BIT */ !! 424 movel %a1@(TASK_THREAD+THREAD_USP),%a0 2183 ENDPROC_CFI(call_on_stack) !! 425 movel %a0,%usp 2184 #endif /* CONFIG_IRQSTACKS */ !! 426 2185 !! 427 /* restore fs (sfc,%dfc) */ 2186 ENTRY_CFI(get_register) !! 428 movew %a1@(TASK_THREAD+THREAD_FC),%a0 2187 /* !! 429 movec %a0,%sfc 2188 * get_register is used by the non ac !! 430 movec %a0,%dfc 2189 * copy the value of the general regi !! 431 2190 * r1. This routine can't be used for !! 432 /* restore status register */ 2191 * the rfir will restore the original !! 433 movew %a1@(TASK_THREAD+THREAD_SR),%d0 2192 * registers we put a -1 into r1 to i !! 434 oriw #0x0700,%d0 2193 * should not be used (the register b !! 435 movew %d0,%sr 2194 * a -1 in it, but that is OK, it jus << 2195 * to use the slow path instead). << 2196 */ << 2197 blr %r8,%r0 << 2198 nop << 2199 bv %r0(%r25) /* r0 */ << 2200 copy %r0,%r1 << 2201 bv %r0(%r25) /* r1 - shadowed << 2202 ldi -1,%r1 << 2203 bv %r0(%r25) /* r2 */ << 2204 copy %r2,%r1 << 2205 bv %r0(%r25) /* r3 */ << 2206 copy %r3,%r1 << 2207 bv %r0(%r25) /* r4 */ << 2208 copy %r4,%r1 << 2209 bv %r0(%r25) /* r5 */ << 2210 copy %r5,%r1 << 2211 bv %r0(%r25) /* r6 */ << 2212 copy %r6,%r1 << 2213 bv %r0(%r25) /* r7 */ << 2214 copy %r7,%r1 << 2215 bv %r0(%r25) /* r8 - shadowed << 2216 ldi -1,%r1 << 2217 bv %r0(%r25) /* r9 - shadowed << 2218 ldi -1,%r1 << 2219 bv %r0(%r25) /* r10 */ << 2220 copy %r10,%r1 << 2221 bv %r0(%r25) /* r11 */ << 2222 copy %r11,%r1 << 2223 bv %r0(%r25) /* r12 */ << 2224 copy %r12,%r1 << 2225 bv %r0(%r25) /* r13 */ << 2226 copy %r13,%r1 << 2227 bv %r0(%r25) /* r14 */ << 2228 copy %r14,%r1 << 2229 bv %r0(%r25) /* r15 */ << 2230 copy %r15,%r1 << 2231 bv %r0(%r25) /* r16 - shadowe << 2232 ldi -1,%r1 << 2233 bv %r0(%r25) /* r17 - shadowe << 2234 ldi -1,%r1 << 2235 bv %r0(%r25) /* r18 */ << 2236 copy %r18,%r1 << 2237 bv %r0(%r25) /* r19 */ << 2238 copy %r19,%r1 << 2239 bv %r0(%r25) /* r20 */ << 2240 copy %r20,%r1 << 2241 bv %r0(%r25) /* r21 */ << 2242 copy %r21,%r1 << 2243 bv %r0(%r25) /* r22 */ << 2244 copy %r22,%r1 << 2245 bv %r0(%r25) /* r23 */ << 2246 copy %r23,%r1 << 2247 bv %r0(%r25) /* r24 - shadowe << 2248 ldi -1,%r1 << 2249 bv %r0(%r25) /* r25 - shadowe << 2250 ldi -1,%r1 << 2251 bv %r0(%r25) /* r26 */ << 2252 copy %r26,%r1 << 2253 bv %r0(%r25) /* r27 */ << 2254 copy %r27,%r1 << 2255 bv %r0(%r25) /* r28 */ << 2256 copy %r28,%r1 << 2257 bv %r0(%r25) /* r29 */ << 2258 copy %r29,%r1 << 2259 bv %r0(%r25) /* r30 */ << 2260 copy %r30,%r1 << 2261 bv %r0(%r25) /* r31 */ << 2262 copy %r31,%r1 << 2263 ENDPROC_CFI(get_register) << 2264 << 2265 436 2266 ENTRY_CFI(set_register) !! 437 rts 2267 /* << 2268 * set_register is used by the non ac << 2269 * copy the value of r1 into the gene << 2270 * r8. << 2271 */ << 2272 blr %r8,%r0 << 2273 nop << 2274 bv %r0(%r25) /* r0 (silly, bu << 2275 copy %r1,%r0 << 2276 bv %r0(%r25) /* r1 */ << 2277 copy %r1,%r1 << 2278 bv %r0(%r25) /* r2 */ << 2279 copy %r1,%r2 << 2280 bv %r0(%r25) /* r3 */ << 2281 copy %r1,%r3 << 2282 bv %r0(%r25) /* r4 */ << 2283 copy %r1,%r4 << 2284 bv %r0(%r25) /* r5 */ << 2285 copy %r1,%r5 << 2286 bv %r0(%r25) /* r6 */ << 2287 copy %r1,%r6 << 2288 bv %r0(%r25) /* r7 */ << 2289 copy %r1,%r7 << 2290 bv %r0(%r25) /* r8 */ << 2291 copy %r1,%r8 << 2292 bv %r0(%r25) /* r9 */ << 2293 copy %r1,%r9 << 2294 bv %r0(%r25) /* r10 */ << 2295 copy %r1,%r10 << 2296 bv %r0(%r25) /* r11 */ << 2297 copy %r1,%r11 << 2298 bv %r0(%r25) /* r12 */ << 2299 copy %r1,%r12 << 2300 bv %r0(%r25) /* r13 */ << 2301 copy %r1,%r13 << 2302 bv %r0(%r25) /* r14 */ << 2303 copy %r1,%r14 << 2304 bv %r0(%r25) /* r15 */ << 2305 copy %r1,%r15 << 2306 bv %r0(%r25) /* r16 */ << 2307 copy %r1,%r16 << 2308 bv %r0(%r25) /* r17 */ << 2309 copy %r1,%r17 << 2310 bv %r0(%r25) /* r18 */ << 2311 copy %r1,%r18 << 2312 bv %r0(%r25) /* r19 */ << 2313 copy %r1,%r19 << 2314 bv %r0(%r25) /* r20 */ << 2315 copy %r1,%r20 << 2316 bv %r0(%r25) /* r21 */ << 2317 copy %r1,%r21 << 2318 bv %r0(%r25) /* r22 */ << 2319 copy %r1,%r22 << 2320 bv %r0(%r25) /* r23 */ << 2321 copy %r1,%r23 << 2322 bv %r0(%r25) /* r24 */ << 2323 copy %r1,%r24 << 2324 bv %r0(%r25) /* r25 */ << 2325 copy %r1,%r25 << 2326 bv %r0(%r25) /* r26 */ << 2327 copy %r1,%r26 << 2328 bv %r0(%r25) /* r27 */ << 2329 copy %r1,%r27 << 2330 bv %r0(%r25) /* r28 */ << 2331 copy %r1,%r28 << 2332 bv %r0(%r25) /* r29 */ << 2333 copy %r1,%r29 << 2334 bv %r0(%r25) /* r30 */ << 2335 copy %r1,%r30 << 2336 bv %r0(%r25) /* r31 */ << 2337 copy %r1,%r31 << 2338 ENDPROC_CFI(set_register) << 2339 438 >> 439 #endif /* CONFIG_MMU && !CONFIG_COLDFIRE */
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.