[PATCH 07/19] powerpc: Create mtmsrd_isync()
Benjamin Herrenschmidt
benh at kernel.crashing.org
Sat Sep 2 13:44:42 AEST 2017
On Thu, 2015-10-29 at 11:43 +1100, Anton Blanchard wrote:
> mtmsrd_isync() will do an mtmsrd followed by an isync on older
> processors. On newer processors we avoid the isync via a feature fixup.
The isync is needed specifically when enabling/disable FP etc... right
?
I'd like to make the name a bit clearer. Maybe something like
set_msr_fpvec() or maybe you can come up with something even better, ie
use a name that represents what it's for rather than what it does.
> Signed-off-by: Anton Blanchard <anton at samba.org>
> ---
> arch/powerpc/include/asm/reg.h | 8 ++++++++
> arch/powerpc/kernel/process.c | 30 ++++++++++++++++++++++--------
> 2 files changed, 30 insertions(+), 8 deletions(-)
>
> diff --git a/arch/powerpc/include/asm/reg.h b/arch/powerpc/include/asm/reg.h
> index a908ada..987dac0 100644
> --- a/arch/powerpc/include/asm/reg.h
> +++ b/arch/powerpc/include/asm/reg.h
> @@ -1193,12 +1193,20 @@
> #define __mtmsrd(v, l) asm volatile("mtmsrd %0," __stringify(l) \
> : : "r" (v) : "memory")
> #define mtmsr(v) __mtmsrd((v), 0)
> +#define __MTMSR "mtmsrd"
> #else
> #define mtmsr(v) asm volatile("mtmsr %0" : \
> : "r" ((unsigned long)(v)) \
> : "memory")
> +#define __MTMSR "mtmsr"
> #endif
>
> +static inline void mtmsr_isync(unsigned long val)
> +{
> + asm volatile(__MTMSR " %0; " ASM_FTR_IFCLR("isync", "nop", %1) : :
> + "r" (val), "i" (CPU_FTR_ARCH_206) : "memory");
> +}
> +
> #define mfspr(rn) ({unsigned long rval; \
> asm volatile("mfspr %0," __stringify(rn) \
> : "=r" (rval)); rval;})
> diff --git a/arch/powerpc/kernel/process.c b/arch/powerpc/kernel/process.c
> index ef64219..5bf8ec2 100644
> --- a/arch/powerpc/kernel/process.c
> +++ b/arch/powerpc/kernel/process.c
> @@ -130,7 +130,10 @@ void enable_kernel_fp(void)
> check_if_tm_restore_required(current);
> giveup_fpu(current);
> } else {
> - giveup_fpu(NULL); /* just enables FP for kernel */
> + u64 oldmsr = mfmsr();
> +
> + if (!(oldmsr & MSR_FP))
> + mtmsr_isync(oldmsr | MSR_FP);
> }
> }
> EXPORT_SYMBOL(enable_kernel_fp);
> @@ -144,7 +147,10 @@ void enable_kernel_altivec(void)
> check_if_tm_restore_required(current);
> giveup_altivec(current);
> } else {
> - giveup_altivec_notask();
> + u64 oldmsr = mfmsr();
> +
> + if (!(oldmsr & MSR_VEC))
> + mtmsr_isync(oldmsr | MSR_VEC);
> }
> }
> EXPORT_SYMBOL(enable_kernel_altivec);
> @@ -173,10 +179,14 @@ void enable_kernel_vsx(void)
> {
> WARN_ON(preemptible());
>
> - if (current->thread.regs && (current->thread.regs->msr & MSR_VSX))
> + if (current->thread.regs && (current->thread.regs->msr & MSR_VSX)) {
> giveup_vsx(current);
> - else
> - giveup_vsx(NULL); /* just enable vsx for kernel - force */
> + } else {
> + u64 oldmsr = mfmsr();
> +
> + if (!(oldmsr & MSR_VSX))
> + mtmsr_isync(oldmsr | MSR_VSX);
> + }
> }
> EXPORT_SYMBOL(enable_kernel_vsx);
>
> @@ -209,10 +219,14 @@ void enable_kernel_spe(void)
> {
> WARN_ON(preemptible());
>
> - if (current->thread.regs && (current->thread.regs->msr & MSR_SPE))
> + if (current->thread.regs && (current->thread.regs->msr & MSR_SPE)) {
> giveup_spe(current);
> - else
> - giveup_spe(NULL); /* just enable SPE for kernel - force */
> + } else {
> + u64 oldmsr = mfmsr();
> +
> + if (!(oldmsr & MSR_SPE))
> + mtmsr_isync(oldmsr | MSR_SPE);
> + }
> }
> EXPORT_SYMBOL(enable_kernel_spe);
>
More information about the Linuxppc-dev
mailing list