mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-04 08:17:46 +00:00

When saving or restoring scalar FP context we want to access the least significant 64 bits of each FP register. When the FP registers are 64 bits wide that is trivially the start of the registers value in memory. However when the FP registers are wider this equivalence will no longer be true for big endian systems. Define a new set of offset macros for the least significant 64 bits of each saved FP register within thread context, and make use of them when saving and restoring scalar FP context. Signed-off-by: Paul Burton <paul.burton@imgtec.com> Cc: linux-mips@linux-mips.org Patchwork: https://patchwork.linux-mips.org/patch/6428/ Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
210 lines
5.2 KiB
C
210 lines
5.2 KiB
C
/*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* Copyright (C) 2003 Ralf Baechle
|
|
*/
|
|
#ifndef _ASM_ASMMACRO_H
|
|
#define _ASM_ASMMACRO_H
|
|
|
|
#include <asm/hazards.h>
|
|
|
|
#ifdef CONFIG_32BIT
|
|
#include <asm/asmmacro-32.h>
|
|
#endif
|
|
#ifdef CONFIG_64BIT
|
|
#include <asm/asmmacro-64.h>
|
|
#endif
|
|
#ifdef CONFIG_MIPS_MT_SMTC
|
|
#include <asm/mipsmtregs.h>
|
|
#endif
|
|
|
|
#ifdef CONFIG_MIPS_MT_SMTC
|
|
.macro local_irq_enable reg=t0
|
|
mfc0 \reg, CP0_TCSTATUS
|
|
ori \reg, \reg, TCSTATUS_IXMT
|
|
xori \reg, \reg, TCSTATUS_IXMT
|
|
mtc0 \reg, CP0_TCSTATUS
|
|
_ehb
|
|
.endm
|
|
|
|
.macro local_irq_disable reg=t0
|
|
mfc0 \reg, CP0_TCSTATUS
|
|
ori \reg, \reg, TCSTATUS_IXMT
|
|
mtc0 \reg, CP0_TCSTATUS
|
|
_ehb
|
|
.endm
|
|
#elif defined(CONFIG_CPU_MIPSR2)
|
|
.macro local_irq_enable reg=t0
|
|
ei
|
|
irq_enable_hazard
|
|
.endm
|
|
|
|
.macro local_irq_disable reg=t0
|
|
di
|
|
irq_disable_hazard
|
|
.endm
|
|
#else
|
|
.macro local_irq_enable reg=t0
|
|
mfc0 \reg, CP0_STATUS
|
|
ori \reg, \reg, 1
|
|
mtc0 \reg, CP0_STATUS
|
|
irq_enable_hazard
|
|
.endm
|
|
|
|
.macro local_irq_disable reg=t0
|
|
mfc0 \reg, CP0_STATUS
|
|
ori \reg, \reg, 1
|
|
xori \reg, \reg, 1
|
|
mtc0 \reg, CP0_STATUS
|
|
irq_disable_hazard
|
|
.endm
|
|
#endif /* CONFIG_MIPS_MT_SMTC */
|
|
|
|
.macro fpu_save_16even thread tmp=t0
|
|
cfc1 \tmp, fcr31
|
|
sdc1 $f0, THREAD_FPR0_LS64(\thread)
|
|
sdc1 $f2, THREAD_FPR2_LS64(\thread)
|
|
sdc1 $f4, THREAD_FPR4_LS64(\thread)
|
|
sdc1 $f6, THREAD_FPR6_LS64(\thread)
|
|
sdc1 $f8, THREAD_FPR8_LS64(\thread)
|
|
sdc1 $f10, THREAD_FPR10_LS64(\thread)
|
|
sdc1 $f12, THREAD_FPR12_LS64(\thread)
|
|
sdc1 $f14, THREAD_FPR14_LS64(\thread)
|
|
sdc1 $f16, THREAD_FPR16_LS64(\thread)
|
|
sdc1 $f18, THREAD_FPR18_LS64(\thread)
|
|
sdc1 $f20, THREAD_FPR20_LS64(\thread)
|
|
sdc1 $f22, THREAD_FPR22_LS64(\thread)
|
|
sdc1 $f24, THREAD_FPR24_LS64(\thread)
|
|
sdc1 $f26, THREAD_FPR26_LS64(\thread)
|
|
sdc1 $f28, THREAD_FPR28_LS64(\thread)
|
|
sdc1 $f30, THREAD_FPR30_LS64(\thread)
|
|
sw \tmp, THREAD_FCR31(\thread)
|
|
.endm
|
|
|
|
.macro fpu_save_16odd thread
|
|
.set push
|
|
.set mips64r2
|
|
sdc1 $f1, THREAD_FPR1_LS64(\thread)
|
|
sdc1 $f3, THREAD_FPR3_LS64(\thread)
|
|
sdc1 $f5, THREAD_FPR5_LS64(\thread)
|
|
sdc1 $f7, THREAD_FPR7_LS64(\thread)
|
|
sdc1 $f9, THREAD_FPR9_LS64(\thread)
|
|
sdc1 $f11, THREAD_FPR11_LS64(\thread)
|
|
sdc1 $f13, THREAD_FPR13_LS64(\thread)
|
|
sdc1 $f15, THREAD_FPR15_LS64(\thread)
|
|
sdc1 $f17, THREAD_FPR17_LS64(\thread)
|
|
sdc1 $f19, THREAD_FPR19_LS64(\thread)
|
|
sdc1 $f21, THREAD_FPR21_LS64(\thread)
|
|
sdc1 $f23, THREAD_FPR23_LS64(\thread)
|
|
sdc1 $f25, THREAD_FPR25_LS64(\thread)
|
|
sdc1 $f27, THREAD_FPR27_LS64(\thread)
|
|
sdc1 $f29, THREAD_FPR29_LS64(\thread)
|
|
sdc1 $f31, THREAD_FPR31_LS64(\thread)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro fpu_save_double thread status tmp
|
|
#if defined(CONFIG_MIPS64) || defined(CONFIG_CPU_MIPS32_R2)
|
|
sll \tmp, \status, 5
|
|
bgez \tmp, 10f
|
|
fpu_save_16odd \thread
|
|
10:
|
|
#endif
|
|
fpu_save_16even \thread \tmp
|
|
.endm
|
|
|
|
.macro fpu_restore_16even thread tmp=t0
|
|
lw \tmp, THREAD_FCR31(\thread)
|
|
ldc1 $f0, THREAD_FPR0_LS64(\thread)
|
|
ldc1 $f2, THREAD_FPR2_LS64(\thread)
|
|
ldc1 $f4, THREAD_FPR4_LS64(\thread)
|
|
ldc1 $f6, THREAD_FPR6_LS64(\thread)
|
|
ldc1 $f8, THREAD_FPR8_LS64(\thread)
|
|
ldc1 $f10, THREAD_FPR10_LS64(\thread)
|
|
ldc1 $f12, THREAD_FPR12_LS64(\thread)
|
|
ldc1 $f14, THREAD_FPR14_LS64(\thread)
|
|
ldc1 $f16, THREAD_FPR16_LS64(\thread)
|
|
ldc1 $f18, THREAD_FPR18_LS64(\thread)
|
|
ldc1 $f20, THREAD_FPR20_LS64(\thread)
|
|
ldc1 $f22, THREAD_FPR22_LS64(\thread)
|
|
ldc1 $f24, THREAD_FPR24_LS64(\thread)
|
|
ldc1 $f26, THREAD_FPR26_LS64(\thread)
|
|
ldc1 $f28, THREAD_FPR28_LS64(\thread)
|
|
ldc1 $f30, THREAD_FPR30_LS64(\thread)
|
|
ctc1 \tmp, fcr31
|
|
.endm
|
|
|
|
.macro fpu_restore_16odd thread
|
|
.set push
|
|
.set mips64r2
|
|
ldc1 $f1, THREAD_FPR1_LS64(\thread)
|
|
ldc1 $f3, THREAD_FPR3_LS64(\thread)
|
|
ldc1 $f5, THREAD_FPR5_LS64(\thread)
|
|
ldc1 $f7, THREAD_FPR7_LS64(\thread)
|
|
ldc1 $f9, THREAD_FPR9_LS64(\thread)
|
|
ldc1 $f11, THREAD_FPR11_LS64(\thread)
|
|
ldc1 $f13, THREAD_FPR13_LS64(\thread)
|
|
ldc1 $f15, THREAD_FPR15_LS64(\thread)
|
|
ldc1 $f17, THREAD_FPR17_LS64(\thread)
|
|
ldc1 $f19, THREAD_FPR19_LS64(\thread)
|
|
ldc1 $f21, THREAD_FPR21_LS64(\thread)
|
|
ldc1 $f23, THREAD_FPR23_LS64(\thread)
|
|
ldc1 $f25, THREAD_FPR25_LS64(\thread)
|
|
ldc1 $f27, THREAD_FPR27_LS64(\thread)
|
|
ldc1 $f29, THREAD_FPR29_LS64(\thread)
|
|
ldc1 $f31, THREAD_FPR31_LS64(\thread)
|
|
.set pop
|
|
.endm
|
|
|
|
.macro fpu_restore_double thread status tmp
|
|
#if defined(CONFIG_MIPS64) || defined(CONFIG_CPU_MIPS32_R2)
|
|
sll \tmp, \status, 5
|
|
bgez \tmp, 10f # 16 register mode?
|
|
|
|
fpu_restore_16odd \thread
|
|
10:
|
|
#endif
|
|
fpu_restore_16even \thread \tmp
|
|
.endm
|
|
|
|
#ifdef CONFIG_CPU_MIPSR2
|
|
.macro _EXT rd, rs, p, s
|
|
ext \rd, \rs, \p, \s
|
|
.endm
|
|
#else /* !CONFIG_CPU_MIPSR2 */
|
|
.macro _EXT rd, rs, p, s
|
|
srl \rd, \rs, \p
|
|
andi \rd, \rd, (1 << \s) - 1
|
|
.endm
|
|
#endif /* !CONFIG_CPU_MIPSR2 */
|
|
|
|
/*
|
|
* Temporary until all gas have MT ASE support
|
|
*/
|
|
.macro DMT reg=0
|
|
.word 0x41600bc1 | (\reg << 16)
|
|
.endm
|
|
|
|
.macro EMT reg=0
|
|
.word 0x41600be1 | (\reg << 16)
|
|
.endm
|
|
|
|
.macro DVPE reg=0
|
|
.word 0x41600001 | (\reg << 16)
|
|
.endm
|
|
|
|
.macro EVPE reg=0
|
|
.word 0x41600021 | (\reg << 16)
|
|
.endm
|
|
|
|
.macro MFTR rt=0, rd=0, u=0, sel=0
|
|
.word 0x41000000 | (\rt << 16) | (\rd << 11) | (\u << 5) | (\sel)
|
|
.endm
|
|
|
|
.macro MTTR rt=0, rd=0, u=0, sel=0
|
|
.word 0x41800000 | (\rt << 16) | (\rd << 11) | (\u << 5) | (\sel)
|
|
.endm
|
|
|
|
#endif /* _ASM_ASMMACRO_H */
|