245 lines
3.9 KiB
ArmAsm
245 lines
3.9 KiB
ArmAsm
|
/*
|
||
|
* Floating-point, VMX/Altivec and VSX loads and stores
|
||
|
* for use in instruction emulation.
|
||
|
*
|
||
|
* Copyright 2010 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com>
|
||
|
*
|
||
|
* This program is free software; you can redistribute it and/or
|
||
|
* modify it under the terms of the GNU General Public License
|
||
|
* as published by the Free Software Foundation; either version
|
||
|
* 2 of the License, or (at your option) any later version.
|
||
|
*/
|
||
|
|
||
|
#include <asm/processor.h>
|
||
|
#include <asm/ppc_asm.h>
|
||
|
#include <asm/ppc-opcode.h>
|
||
|
#include <asm/reg.h>
|
||
|
#include <asm/asm-offsets.h>
|
||
|
#include <linux/errno.h>
|
||
|
|
||
|
#ifdef CONFIG_PPC_FPU
|
||
|
|
||
|
#define STKFRM (PPC_MIN_STKFRM + 16)
|
||
|
|
||
|
/* Get the contents of frN into *p; N is in r3 and p is in r4. */
|
||
|
_GLOBAL(get_fpr)
|
||
|
mflr r0
|
||
|
mfmsr r6
|
||
|
ori r7, r6, MSR_FP
|
||
|
MTMSRD(r7)
|
||
|
isync
|
||
|
rlwinm r3,r3,3,0xf8
|
||
|
bcl 20,31,1f
|
||
|
reg = 0
|
||
|
.rept 32
|
||
|
stfd reg, 0(r4)
|
||
|
b 2f
|
||
|
reg = reg + 1
|
||
|
.endr
|
||
|
1: mflr r5
|
||
|
add r5,r3,r5
|
||
|
mtctr r5
|
||
|
mtlr r0
|
||
|
bctr
|
||
|
2: MTMSRD(r6)
|
||
|
isync
|
||
|
blr
|
||
|
|
||
|
/* Put the contents of *p into frN; N is in r3 and p is in r4. */
|
||
|
_GLOBAL(put_fpr)
|
||
|
mflr r0
|
||
|
mfmsr r6
|
||
|
ori r7, r6, MSR_FP
|
||
|
MTMSRD(r7)
|
||
|
isync
|
||
|
rlwinm r3,r3,3,0xf8
|
||
|
bcl 20,31,1f
|
||
|
reg = 0
|
||
|
.rept 32
|
||
|
lfd reg, 0(r4)
|
||
|
b 2f
|
||
|
reg = reg + 1
|
||
|
.endr
|
||
|
1: mflr r5
|
||
|
add r5,r3,r5
|
||
|
mtctr r5
|
||
|
mtlr r0
|
||
|
bctr
|
||
|
2: MTMSRD(r6)
|
||
|
isync
|
||
|
blr
|
||
|
|
||
|
#ifdef CONFIG_ALTIVEC
|
||
|
/* Get the contents of vrN into *p; N is in r3 and p is in r4. */
|
||
|
_GLOBAL(get_vr)
|
||
|
mflr r0
|
||
|
mfmsr r6
|
||
|
oris r7, r6, MSR_VEC@h
|
||
|
MTMSRD(r7)
|
||
|
isync
|
||
|
rlwinm r3,r3,3,0xf8
|
||
|
bcl 20,31,1f
|
||
|
reg = 0
|
||
|
.rept 32
|
||
|
stvx reg, 0, r4
|
||
|
b 2f
|
||
|
reg = reg + 1
|
||
|
.endr
|
||
|
1: mflr r5
|
||
|
add r5,r3,r5
|
||
|
mtctr r5
|
||
|
mtlr r0
|
||
|
bctr
|
||
|
2: MTMSRD(r6)
|
||
|
isync
|
||
|
blr
|
||
|
|
||
|
/* Put the contents of *p into vrN; N is in r3 and p is in r4. */
|
||
|
_GLOBAL(put_vr)
|
||
|
mflr r0
|
||
|
mfmsr r6
|
||
|
oris r7, r6, MSR_VEC@h
|
||
|
MTMSRD(r7)
|
||
|
isync
|
||
|
rlwinm r3,r3,3,0xf8
|
||
|
bcl 20,31,1f
|
||
|
reg = 0
|
||
|
.rept 32
|
||
|
lvx reg, 0, r4
|
||
|
b 2f
|
||
|
reg = reg + 1
|
||
|
.endr
|
||
|
1: mflr r5
|
||
|
add r5,r3,r5
|
||
|
mtctr r5
|
||
|
mtlr r0
|
||
|
bctr
|
||
|
2: MTMSRD(r6)
|
||
|
isync
|
||
|
blr
|
||
|
#endif /* CONFIG_ALTIVEC */
|
||
|
|
||
|
#ifdef CONFIG_VSX
|
||
|
/* Get the contents of vsN into vs0; N is in r3. */
|
||
|
_GLOBAL(get_vsr)
|
||
|
mflr r0
|
||
|
rlwinm r3,r3,3,0x1f8
|
||
|
bcl 20,31,1f
|
||
|
blr /* vs0 is already in vs0 */
|
||
|
nop
|
||
|
reg = 1
|
||
|
.rept 63
|
||
|
XXLOR(0,reg,reg)
|
||
|
blr
|
||
|
reg = reg + 1
|
||
|
.endr
|
||
|
1: mflr r5
|
||
|
add r5,r3,r5
|
||
|
mtctr r5
|
||
|
mtlr r0
|
||
|
bctr
|
||
|
|
||
|
/* Put the contents of vs0 into vsN; N is in r3. */
|
||
|
_GLOBAL(put_vsr)
|
||
|
mflr r0
|
||
|
rlwinm r3,r3,3,0x1f8
|
||
|
bcl 20,31,1f
|
||
|
blr /* v0 is already in v0 */
|
||
|
nop
|
||
|
reg = 1
|
||
|
.rept 63
|
||
|
XXLOR(reg,0,0)
|
||
|
blr
|
||
|
reg = reg + 1
|
||
|
.endr
|
||
|
1: mflr r5
|
||
|
add r5,r3,r5
|
||
|
mtctr r5
|
||
|
mtlr r0
|
||
|
bctr
|
||
|
|
||
|
/* Load VSX reg N from vector doubleword *p. N is in r3, p in r4. */
|
||
|
_GLOBAL(load_vsrn)
|
||
|
PPC_STLU r1,-STKFRM(r1)
|
||
|
mflr r0
|
||
|
PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
|
||
|
mfmsr r6
|
||
|
oris r7,r6,MSR_VSX@h
|
||
|
cmpwi cr7,r3,0
|
||
|
li r8,STKFRM-16
|
||
|
MTMSRD(r7)
|
||
|
isync
|
||
|
beq cr7,1f
|
||
|
STXVD2X(0,R1,R8)
|
||
|
1: LXVD2X(0,R0,R4)
|
||
|
#ifdef __LITTLE_ENDIAN__
|
||
|
XXSWAPD(0,0)
|
||
|
#endif
|
||
|
beq cr7,4f
|
||
|
bl put_vsr
|
||
|
LXVD2X(0,R1,R8)
|
||
|
4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
|
||
|
mtlr r0
|
||
|
MTMSRD(r6)
|
||
|
isync
|
||
|
addi r1,r1,STKFRM
|
||
|
blr
|
||
|
|
||
|
/* Store VSX reg N to vector doubleword *p. N is in r3, p in r4. */
|
||
|
_GLOBAL(store_vsrn)
|
||
|
PPC_STLU r1,-STKFRM(r1)
|
||
|
mflr r0
|
||
|
PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
|
||
|
mfmsr r6
|
||
|
oris r7,r6,MSR_VSX@h
|
||
|
li r8,STKFRM-16
|
||
|
MTMSRD(r7)
|
||
|
isync
|
||
|
STXVD2X(0,R1,R8)
|
||
|
bl get_vsr
|
||
|
#ifdef __LITTLE_ENDIAN__
|
||
|
XXSWAPD(0,0)
|
||
|
#endif
|
||
|
STXVD2X(0,R0,R4)
|
||
|
LXVD2X(0,R1,R8)
|
||
|
PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
|
||
|
mtlr r0
|
||
|
MTMSRD(r6)
|
||
|
isync
|
||
|
mr r3,r9
|
||
|
addi r1,r1,STKFRM
|
||
|
blr
|
||
|
#endif /* CONFIG_VSX */
|
||
|
|
||
|
/* Convert single-precision to double, without disturbing FPRs. */
|
||
|
/* conv_sp_to_dp(float *sp, double *dp) */
|
||
|
_GLOBAL(conv_sp_to_dp)
|
||
|
mfmsr r6
|
||
|
ori r7, r6, MSR_FP
|
||
|
MTMSRD(r7)
|
||
|
isync
|
||
|
stfd fr0, -16(r1)
|
||
|
lfs fr0, 0(r3)
|
||
|
stfd fr0, 0(r4)
|
||
|
lfd fr0, -16(r1)
|
||
|
MTMSRD(r6)
|
||
|
isync
|
||
|
blr
|
||
|
|
||
|
/* Convert single-precision to double, without disturbing FPRs. */
|
||
|
/* conv_sp_to_dp(double *dp, float *sp) */
|
||
|
_GLOBAL(conv_dp_to_sp)
|
||
|
mfmsr r6
|
||
|
ori r7, r6, MSR_FP
|
||
|
MTMSRD(r7)
|
||
|
isync
|
||
|
stfd fr0, -16(r1)
|
||
|
lfd fr0, 0(r3)
|
||
|
stfs fr0, 0(r4)
|
||
|
lfd fr0, -16(r1)
|
||
|
MTMSRD(r6)
|
||
|
isync
|
||
|
blr
|
||
|
|
||
|
#endif /* CONFIG_PPC_FPU */
|