[RFC PATCH 7/9] powerpc: Add support to mask perf interrupts

Nicholas Piggin npiggin at gmail.com
Tue Jul 26 15:46:05 AEST 2016


On Mon, 25 Jul 2016 20:22:20 +0530
Madhavan Srinivasan <maddy at linux.vnet.ibm.com> wrote:

> To support masking of the PMI interrupts, couple of new interrupt
> handler macros are added MASKABLE_EXCEPTION_PSERIES_OOL and
> MASKABLE_RELON_EXCEPTION_PSERIES_OOL. These are needed to include the
> SOFTEN_TEST and implement the support at both host and guest kernel.
> 
> Couple of new irq #defs "PACA_IRQ_PMI" and "SOFTEN_VALUE_0xf0*" added
> to use in the exception code to check for PMI interrupts.
> 
> __SOFTEN_TEST macro is modified to support the PMI interrupt.
> Present __SOFTEN_TEST code loads the soft_enabled from paca and check
> to call masked_interrupt handler code. To support both current
> behaviour and PMI masking, these changes are added,
> 
> 1) Current LR register content are saved in R11
> 2) "bge" branch operation is changed to "bgel".
> 3) restore R11 to LR
> 
> Reason:
> 
> To retain PMI as NMI behaviour for flag state of 1, we save the LR
> regsiter value in R11 and branch to "masked_interrupt" handler with
> LR update. And in "masked_interrupt" handler, we check for the
> "SOFTEN_VALUE_*" value in R10 for PMI and branch back with "blr" if
> PMI.
> 
> To mask PMI for a flag >1 value, masked_interrupt vaoid's the above
> check and continue to execute the masked_interrupt code and disabled
> MSR[EE] and updated the irq_happend with PMI info.
> 
> Finally, saving of R11 is moved before calling SOFTEN_TEST in the
> __EXCEPTION_PROLOG_1 macro to support saving of LR values in
> SOFTEN_TEST.
> 
> Signed-off-by: Madhavan Srinivasan <maddy at linux.vnet.ibm.com>
> ---
>  arch/powerpc/include/asm/exception-64s.h | 22 ++++++++++++++++++++--
>  arch/powerpc/include/asm/hw_irq.h        |  1 +
>  arch/powerpc/kernel/exceptions-64s.S     | 27
> ++++++++++++++++++++++++--- 3 files changed, 45 insertions(+), 5
> deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/exception-64s.h
> b/arch/powerpc/include/asm/exception-64s.h index
> 44d3f539d8a5..c951b7ab5108 100644 ---
> a/arch/powerpc/include/asm/exception-64s.h +++
> b/arch/powerpc/include/asm/exception-64s.h @@ -166,8 +166,8 @@
> END_FTR_SECTION_NESTED(ftr,ftr,943)
> OPT_SAVE_REG_TO_PACA(area+EX_CFAR, r10, CPU_FTR_CFAR);
> \ SAVE_CTR(r10, area);
> \ mfcr
> r9;							\
> -
> extra(vec);							\
> std
> r11,area+EX_R11(r13);					\
> +
> extra(vec);							\
> std
> r12,area+EX_R12(r13);					\
> GET_SCRATCH0(r10);						\
> std	r10,area+EX_R13(r13) @@ -403,12 +403,17 @@
> label##_relon_hv:						\
> #define SOFTEN_VALUE_0xe82	PACA_IRQ_DBELL #define
> SOFTEN_VALUE_0xe60	PACA_IRQ_HMI #define
> SOFTEN_VALUE_0xe62	PACA_IRQ_HMI +#define
> SOFTEN_VALUE_0xf01	PACA_IRQ_PMI +#define
> SOFTEN_VALUE_0xf00	PACA_IRQ_PMI

#define __SOFTEN_TEST(h,
> vec)						\ lbz
> r10,PACASOFTIRQEN(r13);					\
> cmpwi
> r10,LAZY_INTERRUPT_DISABLED;				\
> li
> r10,SOFTEN_VALUE_##vec;					\
> -	bge	masked_##h##interrupt

At which point, can't we pass in the interrupt level we want to mask
for to SOFTEN_TEST, and avoid all this extra code changes?


PMU masked interrupt will compare with SOFTEN_LEVEL_PMU, existing
interrupts will compare with SOFTEN_LEVEL_EE (or whatever suitable
names there are).


> +	mflr
> r11;							\
> +	bgel
> masked_##h##interrupt;					\
> +	mtlr	r11;

This might corrupt return prediction when masked_interrupt does not
return. I guess that's uncommon case though. But I think we can avoid
this if we do the above, no?

Thanks,
Nick


More information about the Linuxppc-dev mailing list