mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

While the GCC and Clang compilers already define __ASSEMBLER__ automatically when compiling assembly code, __ASSEMBLY__ is a macro that only gets defined by the Makefiles in the kernel. This can be very confusing when switching between userspace and kernelspace coding, or when dealing with uapi headers that rather should use __ASSEMBLER__ instead. So let's standardize on the __ASSEMBLER__ macro that is provided by the compilers now. This is mostly a completely mechanical patch (done with a simple "sed -i" statement), except for some manual tweaks in the files arch/parisc/include/asm/smp.h, arch/parisc/include/asm/signal.h, arch/parisc/include/asm/thread_info.h and arch/parisc/include/asm/vdso.h that had the macro spelled in a wrong way. Cc: "James E.J. Bottomley" <James.Bottomley@HansenPartnership.com> Cc: Helge Deller <deller@gmx.de> Cc: linux-parisc@vger.kernel.org Signed-off-by: Thomas Huth <thuth@redhat.com> Signed-off-by: Helge Deller <deller@gmx.de>
586 lines
14 KiB
C
586 lines
14 KiB
C
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
/*
|
|
* Copyright (C) 1999 Hewlett-Packard (Frank Rowand)
|
|
* Copyright (C) 1999 Philipp Rumpf <prumpf@tux.org>
|
|
* Copyright (C) 1999 SuSE GmbH
|
|
* Copyright (C) 2021 Helge Deller <deller@gmx.de>
|
|
*/
|
|
|
|
#ifndef _PARISC_ASSEMBLY_H
|
|
#define _PARISC_ASSEMBLY_H
|
|
|
|
#ifdef CONFIG_64BIT
|
|
#define RP_OFFSET 16
|
|
#define FRAME_SIZE 128
|
|
#define CALLEE_REG_FRAME_SIZE 144
|
|
#define REG_SZ 8
|
|
#define ASM_ULONG_INSN .dword
|
|
#else /* CONFIG_64BIT */
|
|
#define RP_OFFSET 20
|
|
#define FRAME_SIZE 64
|
|
#define CALLEE_REG_FRAME_SIZE 128
|
|
#define REG_SZ 4
|
|
#define ASM_ULONG_INSN .word
|
|
#endif
|
|
|
|
/* Frame alignment for 32- and 64-bit */
|
|
#define FRAME_ALIGN 64
|
|
|
|
#define CALLEE_FLOAT_FRAME_SIZE 80
|
|
#define CALLEE_SAVE_FRAME_SIZE (CALLEE_REG_FRAME_SIZE + CALLEE_FLOAT_FRAME_SIZE)
|
|
|
|
#ifdef CONFIG_PA20
|
|
#define LDCW ldcw,co
|
|
#define BL b,l
|
|
# ifdef CONFIG_64BIT
|
|
# define PA_ASM_LEVEL 2.0w
|
|
# else
|
|
# define PA_ASM_LEVEL 2.0
|
|
# endif
|
|
#else
|
|
#define LDCW ldcw
|
|
#define BL bl
|
|
#define PA_ASM_LEVEL 1.1
|
|
#endif
|
|
|
|
/* Privilege level field in the rightmost two bits of the IA queues */
|
|
#define PRIV_USER 3
|
|
#define PRIV_KERNEL 0
|
|
|
|
/* Space register used inside kernel */
|
|
#define SR_KERNEL 0
|
|
#define SR_TEMP1 1
|
|
#define SR_TEMP2 2
|
|
#define SR_USER 3
|
|
|
|
#ifdef __ASSEMBLER__
|
|
|
|
#ifdef CONFIG_64BIT
|
|
#define LDREG ldd
|
|
#define STREG std
|
|
#define LDREGX ldd,s
|
|
#define LDREGM ldd,mb
|
|
#define STREGM std,ma
|
|
#define SHRREG shrd
|
|
#define SHLREG shld
|
|
#define ANDCM andcm,*
|
|
#define COND(x) * ## x
|
|
#else /* CONFIG_64BIT */
|
|
#define LDREG ldw
|
|
#define STREG stw
|
|
#define LDREGX ldwx,s
|
|
#define LDREGM ldwm
|
|
#define STREGM stwm
|
|
#define SHRREG shr
|
|
#define SHLREG shlw
|
|
#define ANDCM andcm
|
|
#define COND(x) x
|
|
#endif
|
|
|
|
#ifdef CONFIG_64BIT
|
|
/* the 64-bit pa gnu assembler unfortunately defaults to .level 1.1 or 2.0 so
|
|
* work around that for now... */
|
|
.level 2.0w
|
|
#endif
|
|
|
|
#include <asm/asm-offsets.h>
|
|
#include <asm/page.h>
|
|
#include <asm/types.h>
|
|
|
|
#include <asm/asmregs.h>
|
|
#include <asm/psw.h>
|
|
|
|
/*
|
|
* We provide two versions of each macro to convert from physical
|
|
* to virtual and vice versa. The "_r1" versions take one argument
|
|
* register, but trashes r1 to do the conversion. The other
|
|
* version takes two arguments: a src and destination register.
|
|
* However, the source and destination registers can not be
|
|
* the same register.
|
|
*
|
|
* We use add,l to avoid clobbering the C/B bits in the PSW.
|
|
*/
|
|
|
|
.macro tophys grvirt, grphys
|
|
ldil L%(-__PAGE_OFFSET), \grphys
|
|
addl \grvirt, \grphys, \grphys
|
|
.endm
|
|
|
|
.macro tovirt grphys, grvirt
|
|
ldil L%(__PAGE_OFFSET), \grvirt
|
|
addl \grphys, \grvirt, \grvirt
|
|
.endm
|
|
|
|
.macro tophys_r1 gr
|
|
ldil L%(-__PAGE_OFFSET), %r1
|
|
addl \gr, %r1, \gr
|
|
.endm
|
|
|
|
.macro tovirt_r1 gr
|
|
ldil L%(__PAGE_OFFSET), %r1
|
|
addl \gr, %r1, \gr
|
|
.endm
|
|
|
|
.macro delay value
|
|
ldil L%\value, 1
|
|
ldo R%\value(1), 1
|
|
addib,UV,n -1,1,.
|
|
addib,NUV,n -1,1,.+8
|
|
nop
|
|
.endm
|
|
|
|
.macro debug value
|
|
.endm
|
|
|
|
.macro shlw r, sa, t
|
|
zdep \r, 31-(\sa), 32-(\sa), \t
|
|
.endm
|
|
|
|
/* And the PA 2.0W shift left */
|
|
.macro shld r, sa, t
|
|
depd,z \r, 63-(\sa), 64-(\sa), \t
|
|
.endm
|
|
|
|
/* Shift Right for 32-bit. Clobbers upper 32-bit on PA2.0. */
|
|
.macro shr r, sa, t
|
|
extru \r, 31-(\sa), 32-(\sa), \t
|
|
.endm
|
|
|
|
/* pa20w version of shift right */
|
|
.macro shrd r, sa, t
|
|
extrd,u \r, 63-(\sa), 64-(\sa), \t
|
|
.endm
|
|
|
|
/* Extract unsigned for 32- and 64-bit
|
|
* The extru instruction leaves the most significant 32 bits of the
|
|
* target register in an undefined state on PA 2.0 systems. */
|
|
.macro extru_safe r, p, len, t
|
|
#ifdef CONFIG_64BIT
|
|
extrd,u \r, 32+(\p), \len, \t
|
|
#else
|
|
extru \r, \p, \len, \t
|
|
#endif
|
|
.endm
|
|
|
|
/* The depi instruction leaves the most significant 32 bits of the
|
|
* target register in an undefined state on PA 2.0 systems. */
|
|
.macro depi_safe i, p, len, t
|
|
#ifdef CONFIG_64BIT
|
|
depdi \i, 32+(\p), \len, \t
|
|
#else
|
|
depi \i, \p, \len, \t
|
|
#endif
|
|
.endm
|
|
|
|
/* The depw instruction leaves the most significant 32 bits of the
|
|
* target register in an undefined state on PA 2.0 systems. */
|
|
.macro dep_safe i, p, len, t
|
|
#ifdef CONFIG_64BIT
|
|
depd \i, 32+(\p), \len, \t
|
|
#else
|
|
depw \i, \p, \len, \t
|
|
#endif
|
|
.endm
|
|
|
|
/* load 32-bit 'value' into 'reg' compensating for the ldil
|
|
* sign-extension when running in wide mode.
|
|
* WARNING!! neither 'value' nor 'reg' can be expressions
|
|
* containing '.'!!!! */
|
|
.macro load32 value, reg
|
|
ldil L%\value, \reg
|
|
ldo R%\value(\reg), \reg
|
|
.endm
|
|
|
|
.macro loadgp
|
|
#ifdef CONFIG_64BIT
|
|
ldil L%__gp, %r27
|
|
ldo R%__gp(%r27), %r27
|
|
#else
|
|
ldil L%$global$, %r27
|
|
ldo R%$global$(%r27), %r27
|
|
#endif
|
|
.endm
|
|
|
|
#define SAVE_SP(r, where) mfsp r, %r1 ! STREG %r1, where
|
|
#define REST_SP(r, where) LDREG where, %r1 ! mtsp %r1, r
|
|
#define SAVE_CR(r, where) mfctl r, %r1 ! STREG %r1, where
|
|
#define REST_CR(r, where) LDREG where, %r1 ! mtctl %r1, r
|
|
|
|
.macro save_general regs
|
|
STREG %r1, PT_GR1 (\regs)
|
|
STREG %r2, PT_GR2 (\regs)
|
|
STREG %r3, PT_GR3 (\regs)
|
|
STREG %r4, PT_GR4 (\regs)
|
|
STREG %r5, PT_GR5 (\regs)
|
|
STREG %r6, PT_GR6 (\regs)
|
|
STREG %r7, PT_GR7 (\regs)
|
|
STREG %r8, PT_GR8 (\regs)
|
|
STREG %r9, PT_GR9 (\regs)
|
|
STREG %r10, PT_GR10(\regs)
|
|
STREG %r11, PT_GR11(\regs)
|
|
STREG %r12, PT_GR12(\regs)
|
|
STREG %r13, PT_GR13(\regs)
|
|
STREG %r14, PT_GR14(\regs)
|
|
STREG %r15, PT_GR15(\regs)
|
|
STREG %r16, PT_GR16(\regs)
|
|
STREG %r17, PT_GR17(\regs)
|
|
STREG %r18, PT_GR18(\regs)
|
|
STREG %r19, PT_GR19(\regs)
|
|
STREG %r20, PT_GR20(\regs)
|
|
STREG %r21, PT_GR21(\regs)
|
|
STREG %r22, PT_GR22(\regs)
|
|
STREG %r23, PT_GR23(\regs)
|
|
STREG %r24, PT_GR24(\regs)
|
|
STREG %r25, PT_GR25(\regs)
|
|
/* r26 is saved in get_stack and used to preserve a value across virt_map */
|
|
STREG %r27, PT_GR27(\regs)
|
|
STREG %r28, PT_GR28(\regs)
|
|
/* r29 is saved in get_stack and used to point to saved registers */
|
|
/* r30 stack pointer saved in get_stack */
|
|
STREG %r31, PT_GR31(\regs)
|
|
.endm
|
|
|
|
.macro rest_general regs
|
|
/* r1 used as a temp in rest_stack and is restored there */
|
|
LDREG PT_GR2 (\regs), %r2
|
|
LDREG PT_GR3 (\regs), %r3
|
|
LDREG PT_GR4 (\regs), %r4
|
|
LDREG PT_GR5 (\regs), %r5
|
|
LDREG PT_GR6 (\regs), %r6
|
|
LDREG PT_GR7 (\regs), %r7
|
|
LDREG PT_GR8 (\regs), %r8
|
|
LDREG PT_GR9 (\regs), %r9
|
|
LDREG PT_GR10(\regs), %r10
|
|
LDREG PT_GR11(\regs), %r11
|
|
LDREG PT_GR12(\regs), %r12
|
|
LDREG PT_GR13(\regs), %r13
|
|
LDREG PT_GR14(\regs), %r14
|
|
LDREG PT_GR15(\regs), %r15
|
|
LDREG PT_GR16(\regs), %r16
|
|
LDREG PT_GR17(\regs), %r17
|
|
LDREG PT_GR18(\regs), %r18
|
|
LDREG PT_GR19(\regs), %r19
|
|
LDREG PT_GR20(\regs), %r20
|
|
LDREG PT_GR21(\regs), %r21
|
|
LDREG PT_GR22(\regs), %r22
|
|
LDREG PT_GR23(\regs), %r23
|
|
LDREG PT_GR24(\regs), %r24
|
|
LDREG PT_GR25(\regs), %r25
|
|
LDREG PT_GR26(\regs), %r26
|
|
LDREG PT_GR27(\regs), %r27
|
|
LDREG PT_GR28(\regs), %r28
|
|
/* r29 points to register save area, and is restored in rest_stack */
|
|
/* r30 stack pointer restored in rest_stack */
|
|
LDREG PT_GR31(\regs), %r31
|
|
.endm
|
|
|
|
.macro save_fp regs
|
|
fstd,ma %fr0, 8(\regs)
|
|
fstd,ma %fr1, 8(\regs)
|
|
fstd,ma %fr2, 8(\regs)
|
|
fstd,ma %fr3, 8(\regs)
|
|
fstd,ma %fr4, 8(\regs)
|
|
fstd,ma %fr5, 8(\regs)
|
|
fstd,ma %fr6, 8(\regs)
|
|
fstd,ma %fr7, 8(\regs)
|
|
fstd,ma %fr8, 8(\regs)
|
|
fstd,ma %fr9, 8(\regs)
|
|
fstd,ma %fr10, 8(\regs)
|
|
fstd,ma %fr11, 8(\regs)
|
|
fstd,ma %fr12, 8(\regs)
|
|
fstd,ma %fr13, 8(\regs)
|
|
fstd,ma %fr14, 8(\regs)
|
|
fstd,ma %fr15, 8(\regs)
|
|
fstd,ma %fr16, 8(\regs)
|
|
fstd,ma %fr17, 8(\regs)
|
|
fstd,ma %fr18, 8(\regs)
|
|
fstd,ma %fr19, 8(\regs)
|
|
fstd,ma %fr20, 8(\regs)
|
|
fstd,ma %fr21, 8(\regs)
|
|
fstd,ma %fr22, 8(\regs)
|
|
fstd,ma %fr23, 8(\regs)
|
|
fstd,ma %fr24, 8(\regs)
|
|
fstd,ma %fr25, 8(\regs)
|
|
fstd,ma %fr26, 8(\regs)
|
|
fstd,ma %fr27, 8(\regs)
|
|
fstd,ma %fr28, 8(\regs)
|
|
fstd,ma %fr29, 8(\regs)
|
|
fstd,ma %fr30, 8(\regs)
|
|
fstd %fr31, 0(\regs)
|
|
.endm
|
|
|
|
.macro rest_fp regs
|
|
fldd 0(\regs), %fr31
|
|
fldd,mb -8(\regs), %fr30
|
|
fldd,mb -8(\regs), %fr29
|
|
fldd,mb -8(\regs), %fr28
|
|
fldd,mb -8(\regs), %fr27
|
|
fldd,mb -8(\regs), %fr26
|
|
fldd,mb -8(\regs), %fr25
|
|
fldd,mb -8(\regs), %fr24
|
|
fldd,mb -8(\regs), %fr23
|
|
fldd,mb -8(\regs), %fr22
|
|
fldd,mb -8(\regs), %fr21
|
|
fldd,mb -8(\regs), %fr20
|
|
fldd,mb -8(\regs), %fr19
|
|
fldd,mb -8(\regs), %fr18
|
|
fldd,mb -8(\regs), %fr17
|
|
fldd,mb -8(\regs), %fr16
|
|
fldd,mb -8(\regs), %fr15
|
|
fldd,mb -8(\regs), %fr14
|
|
fldd,mb -8(\regs), %fr13
|
|
fldd,mb -8(\regs), %fr12
|
|
fldd,mb -8(\regs), %fr11
|
|
fldd,mb -8(\regs), %fr10
|
|
fldd,mb -8(\regs), %fr9
|
|
fldd,mb -8(\regs), %fr8
|
|
fldd,mb -8(\regs), %fr7
|
|
fldd,mb -8(\regs), %fr6
|
|
fldd,mb -8(\regs), %fr5
|
|
fldd,mb -8(\regs), %fr4
|
|
fldd,mb -8(\regs), %fr3
|
|
fldd,mb -8(\regs), %fr2
|
|
fldd,mb -8(\regs), %fr1
|
|
fldd,mb -8(\regs), %fr0
|
|
.endm
|
|
|
|
.macro callee_save_float
|
|
fstd,ma %fr12, 8(%r30)
|
|
fstd,ma %fr13, 8(%r30)
|
|
fstd,ma %fr14, 8(%r30)
|
|
fstd,ma %fr15, 8(%r30)
|
|
fstd,ma %fr16, 8(%r30)
|
|
fstd,ma %fr17, 8(%r30)
|
|
fstd,ma %fr18, 8(%r30)
|
|
fstd,ma %fr19, 8(%r30)
|
|
fstd,ma %fr20, 8(%r30)
|
|
fstd,ma %fr21, 8(%r30)
|
|
.endm
|
|
|
|
.macro callee_rest_float
|
|
fldd,mb -8(%r30), %fr21
|
|
fldd,mb -8(%r30), %fr20
|
|
fldd,mb -8(%r30), %fr19
|
|
fldd,mb -8(%r30), %fr18
|
|
fldd,mb -8(%r30), %fr17
|
|
fldd,mb -8(%r30), %fr16
|
|
fldd,mb -8(%r30), %fr15
|
|
fldd,mb -8(%r30), %fr14
|
|
fldd,mb -8(%r30), %fr13
|
|
fldd,mb -8(%r30), %fr12
|
|
.endm
|
|
|
|
#ifdef CONFIG_64BIT
|
|
.macro callee_save
|
|
std,ma %r3, CALLEE_REG_FRAME_SIZE(%r30)
|
|
mfctl %cr27, %r3
|
|
std %r4, -136(%r30)
|
|
std %r5, -128(%r30)
|
|
std %r6, -120(%r30)
|
|
std %r7, -112(%r30)
|
|
std %r8, -104(%r30)
|
|
std %r9, -96(%r30)
|
|
std %r10, -88(%r30)
|
|
std %r11, -80(%r30)
|
|
std %r12, -72(%r30)
|
|
std %r13, -64(%r30)
|
|
std %r14, -56(%r30)
|
|
std %r15, -48(%r30)
|
|
std %r16, -40(%r30)
|
|
std %r17, -32(%r30)
|
|
std %r18, -24(%r30)
|
|
std %r3, -16(%r30)
|
|
.endm
|
|
|
|
.macro callee_rest
|
|
ldd -16(%r30), %r3
|
|
ldd -24(%r30), %r18
|
|
ldd -32(%r30), %r17
|
|
ldd -40(%r30), %r16
|
|
ldd -48(%r30), %r15
|
|
ldd -56(%r30), %r14
|
|
ldd -64(%r30), %r13
|
|
ldd -72(%r30), %r12
|
|
ldd -80(%r30), %r11
|
|
ldd -88(%r30), %r10
|
|
ldd -96(%r30), %r9
|
|
ldd -104(%r30), %r8
|
|
ldd -112(%r30), %r7
|
|
ldd -120(%r30), %r6
|
|
ldd -128(%r30), %r5
|
|
ldd -136(%r30), %r4
|
|
mtctl %r3, %cr27
|
|
ldd,mb -CALLEE_REG_FRAME_SIZE(%r30), %r3
|
|
.endm
|
|
|
|
#else /* ! CONFIG_64BIT */
|
|
|
|
.macro callee_save
|
|
stw,ma %r3, CALLEE_REG_FRAME_SIZE(%r30)
|
|
mfctl %cr27, %r3
|
|
stw %r4, -124(%r30)
|
|
stw %r5, -120(%r30)
|
|
stw %r6, -116(%r30)
|
|
stw %r7, -112(%r30)
|
|
stw %r8, -108(%r30)
|
|
stw %r9, -104(%r30)
|
|
stw %r10, -100(%r30)
|
|
stw %r11, -96(%r30)
|
|
stw %r12, -92(%r30)
|
|
stw %r13, -88(%r30)
|
|
stw %r14, -84(%r30)
|
|
stw %r15, -80(%r30)
|
|
stw %r16, -76(%r30)
|
|
stw %r17, -72(%r30)
|
|
stw %r18, -68(%r30)
|
|
stw %r3, -64(%r30)
|
|
.endm
|
|
|
|
.macro callee_rest
|
|
ldw -64(%r30), %r3
|
|
ldw -68(%r30), %r18
|
|
ldw -72(%r30), %r17
|
|
ldw -76(%r30), %r16
|
|
ldw -80(%r30), %r15
|
|
ldw -84(%r30), %r14
|
|
ldw -88(%r30), %r13
|
|
ldw -92(%r30), %r12
|
|
ldw -96(%r30), %r11
|
|
ldw -100(%r30), %r10
|
|
ldw -104(%r30), %r9
|
|
ldw -108(%r30), %r8
|
|
ldw -112(%r30), %r7
|
|
ldw -116(%r30), %r6
|
|
ldw -120(%r30), %r5
|
|
ldw -124(%r30), %r4
|
|
mtctl %r3, %cr27
|
|
ldw,mb -CALLEE_REG_FRAME_SIZE(%r30), %r3
|
|
.endm
|
|
#endif /* ! CONFIG_64BIT */
|
|
|
|
.macro save_specials regs
|
|
|
|
SAVE_SP (%sr0, PT_SR0 (\regs))
|
|
SAVE_SP (%sr1, PT_SR1 (\regs))
|
|
SAVE_SP (%sr2, PT_SR2 (\regs))
|
|
SAVE_SP (%sr3, PT_SR3 (\regs))
|
|
SAVE_SP (%sr4, PT_SR4 (\regs))
|
|
SAVE_SP (%sr5, PT_SR5 (\regs))
|
|
SAVE_SP (%sr6, PT_SR6 (\regs))
|
|
|
|
SAVE_CR (%cr17, PT_IASQ0(\regs))
|
|
mtctl %r0, %cr17
|
|
SAVE_CR (%cr17, PT_IASQ1(\regs))
|
|
|
|
SAVE_CR (%cr18, PT_IAOQ0(\regs))
|
|
mtctl %r0, %cr18
|
|
SAVE_CR (%cr18, PT_IAOQ1(\regs))
|
|
|
|
#ifdef CONFIG_64BIT
|
|
/* cr11 (sar) is a funny one. 5 bits on PA1.1 and 6 bit on PA2.0
|
|
* For PA2.0 mtsar or mtctl always write 6 bits, but mfctl only
|
|
* reads 5 bits. Use mfctl,w to read all six bits. Otherwise
|
|
* we lose the 6th bit on a save/restore over interrupt.
|
|
*/
|
|
mfctl,w %cr11, %r1
|
|
STREG %r1, PT_SAR (\regs)
|
|
#else
|
|
SAVE_CR (%cr11, PT_SAR (\regs))
|
|
#endif
|
|
SAVE_CR (%cr19, PT_IIR (\regs))
|
|
|
|
/*
|
|
* Code immediately following this macro (in intr_save) relies
|
|
* on r8 containing ipsw.
|
|
*/
|
|
mfctl %cr22, %r8
|
|
STREG %r8, PT_PSW(\regs)
|
|
.endm
|
|
|
|
.macro rest_specials regs
|
|
|
|
REST_SP (%sr0, PT_SR0 (\regs))
|
|
REST_SP (%sr1, PT_SR1 (\regs))
|
|
REST_SP (%sr2, PT_SR2 (\regs))
|
|
REST_SP (%sr3, PT_SR3 (\regs))
|
|
REST_SP (%sr4, PT_SR4 (\regs))
|
|
REST_SP (%sr5, PT_SR5 (\regs))
|
|
REST_SP (%sr6, PT_SR6 (\regs))
|
|
REST_SP (%sr7, PT_SR7 (\regs))
|
|
|
|
REST_CR (%cr17, PT_IASQ0(\regs))
|
|
REST_CR (%cr17, PT_IASQ1(\regs))
|
|
|
|
REST_CR (%cr18, PT_IAOQ0(\regs))
|
|
REST_CR (%cr18, PT_IAOQ1(\regs))
|
|
|
|
REST_CR (%cr11, PT_SAR (\regs))
|
|
|
|
REST_CR (%cr22, PT_PSW (\regs))
|
|
.endm
|
|
|
|
|
|
/* First step to create a "relied upon translation"
|
|
* See PA 2.0 Arch. page F-4 and F-5.
|
|
*
|
|
* The ssm was originally necessary due to a "PCxT bug".
|
|
* But someone decided it needed to be added to the architecture
|
|
* and this "feature" went into rev3 of PA-RISC 1.1 Arch Manual.
|
|
* It's been carried forward into PA 2.0 Arch as well. :^(
|
|
*
|
|
* "ssm 0,%r0" is a NOP with side effects (prefetch barrier).
|
|
* rsm/ssm prevents the ifetch unit from speculatively fetching
|
|
* instructions past this line in the code stream.
|
|
* PA 2.0 processor will single step all insn in the same QUAD (4 insn).
|
|
*/
|
|
.macro pcxt_ssm_bug
|
|
rsm PSW_SM_I,%r0
|
|
nop /* 1 */
|
|
nop /* 2 */
|
|
nop /* 3 */
|
|
nop /* 4 */
|
|
nop /* 5 */
|
|
nop /* 6 */
|
|
nop /* 7 */
|
|
.endm
|
|
|
|
/* Switch to virtual mapping, trashing only %r1 */
|
|
.macro virt_map
|
|
/* pcxt_ssm_bug */
|
|
rsm PSW_SM_I, %r0 /* barrier for "Relied upon Translation */
|
|
mtsp %r0, %sr4
|
|
mtsp %r0, %sr5
|
|
mtsp %r0, %sr6
|
|
tovirt_r1 %r29
|
|
load32 KERNEL_PSW, %r1
|
|
|
|
rsm PSW_SM_QUIET,%r0 /* second "heavy weight" ctl op */
|
|
mtctl %r0, %cr17 /* Clear IIASQ tail */
|
|
mtctl %r0, %cr17 /* Clear IIASQ head */
|
|
mtctl %r1, %ipsw
|
|
load32 4f, %r1
|
|
mtctl %r1, %cr18 /* Set IIAOQ tail */
|
|
ldo 4(%r1), %r1
|
|
mtctl %r1, %cr18 /* Set IIAOQ head */
|
|
rfir
|
|
nop
|
|
4:
|
|
.endm
|
|
|
|
|
|
/*
|
|
* ASM_EXCEPTIONTABLE_ENTRY
|
|
*
|
|
* Creates an exception table entry.
|
|
* Do not convert to a assembler macro. This won't work.
|
|
*/
|
|
#define ASM_EXCEPTIONTABLE_ENTRY(fault_addr, except_addr) \
|
|
.section __ex_table,"aw" ! \
|
|
.align 4 ! \
|
|
.word (fault_addr - .), (except_addr - .) ! \
|
|
or %r0,%r0,%r0 ! \
|
|
.previous
|
|
|
|
|
|
#endif /* __ASSEMBLER__ */
|
|
#endif
|