[PATCH v5 2/4] powerpc/64: enhance memcmp() with VMX instruction for long bytes comparision
Simon Guo
wei.guo.simon at gmail.com
Thu May 24 01:37:34 AEST 2018
Hi Michael,
On Thu, May 24, 2018 at 05:44:33PM +1000, Michael Ellerman wrote:
> Hi Simon,
>
> wei.guo.simon at gmail.com writes:
> > From: Simon Guo <wei.guo.simon at gmail.com>
> >
> > This patch add VMX primitives to do memcmp() in case the compare size
> > exceeds 4K bytes. KSM feature can benefit from this.
>
> You say "exceeds 4K" here.
>
it should be >= 4k. I will correct the message.
> > diff --git a/arch/powerpc/lib/memcmp_64.S b/arch/powerpc/lib/memcmp_64.S
> > index f20e883..6303bbf 100644
> > --- a/arch/powerpc/lib/memcmp_64.S
> > +++ b/arch/powerpc/lib/memcmp_64.S
> > @@ -27,12 +27,73 @@
> > #define LH lhbrx
> > #define LW lwbrx
> > #define LD ldbrx
> > +#define LVS lvsr
> > +#define VPERM(_VRT,_VRA,_VRB,_VRC) \
> > + vperm _VRT,_VRB,_VRA,_VRC
> > #else
> > #define LH lhzx
> > #define LW lwzx
> > #define LD ldx
> > +#define LVS lvsl
> > +#define VPERM(_VRT,_VRA,_VRB,_VRC) \
> > + vperm _VRT,_VRA,_VRB,_VRC
> > #endif
> >
> > +#define VMX_OPS_THRES 4096
>
> THRES == 4096
>
> BTW, can we call it VMX_THRESH ?
>
Sure. I will update it.
> > +#define ENTER_VMX_OPS \
> > + mflr r0; \
> > + std r3,-STACKFRAMESIZE+STK_REG(R31)(r1); \
> > + std r4,-STACKFRAMESIZE+STK_REG(R30)(r1); \
> > + std r5,-STACKFRAMESIZE+STK_REG(R29)(r1); \
> > + std r0,16(r1); \
> > + stdu r1,-STACKFRAMESIZE(r1); \
> > + bl enter_vmx_ops; \
> > + cmpwi cr1,r3,0; \
> > + ld r0,STACKFRAMESIZE+16(r1); \
> > + ld r3,STK_REG(R31)(r1); \
> > + ld r4,STK_REG(R30)(r1); \
> > + ld r5,STK_REG(R29)(r1); \
> > + addi r1,r1,STACKFRAMESIZE; \
> > + mtlr r0
> > +
> > +#define EXIT_VMX_OPS \
> > + mflr r0; \
> > + std r3,-STACKFRAMESIZE+STK_REG(R31)(r1); \
> > + std r4,-STACKFRAMESIZE+STK_REG(R30)(r1); \
> > + std r5,-STACKFRAMESIZE+STK_REG(R29)(r1); \
> > + std r0,16(r1); \
> > + stdu r1,-STACKFRAMESIZE(r1); \
> > + bl exit_vmx_ops; \
> > + ld r0,STACKFRAMESIZE+16(r1); \
> > + ld r3,STK_REG(R31)(r1); \
> > + ld r4,STK_REG(R30)(r1); \
> > + ld r5,STK_REG(R29)(r1); \
> > + addi r1,r1,STACKFRAMESIZE; \
> > + mtlr r0
> > +
> > +/*
> > + * LD_VSR_CROSS16B load the 2nd 16 bytes for _vaddr which is unaligned with
> > + * 16 bytes boundary and permute the result with the 1st 16 bytes.
> > +
> > + * | y y y y y y y y y y y y y 0 1 2 | 3 4 5 6 7 8 9 a b c d e f z z z |
> > + * ^ ^ ^
> > + * 0xbbbb10 0xbbbb20 0xbbb30
> > + * ^
> > + * _vaddr
> > + *
> > + *
> > + * _vmask is the mask generated by LVS
> > + * _v1st_qw is the 1st aligned QW of current addr which is already loaded.
> > + * for example: 0xyyyyyyyyyyyyy012 for big endian
> > + * _v2nd_qw is the 2nd aligned QW of cur _vaddr to be loaded.
> > + * for example: 0x3456789abcdefzzz for big endian
> > + * The permute result is saved in _v_res.
> > + * for example: 0x0123456789abcdef for big endian.
> > + */
> > +#define LD_VSR_CROSS16B(_vaddr,_vmask,_v1st_qw,_v2nd_qw,_v_res) \
> > + lvx _v2nd_qw,_vaddr,off16; \
> > + VPERM(_v_res,_v1st_qw,_v2nd_qw,_vmask)
> > +
> > /*
> > * There are 2 categories for memcmp:
> > * 1) src/dst has the same offset to the 8 bytes boundary. The handlers
> > @@ -174,6 +235,13 @@ _GLOBAL(memcmp)
> > blr
> >
> > .Llong:
> > +#ifdef CONFIG_ALTIVEC
> > + /* Try to use vmx loop if length is larger than 4K */
> > + cmpldi cr6,r5,VMX_OPS_THRES
> > + bge cr6,.Lsameoffset_vmx_cmp
>
> Here we compare the length to 4K and if it's greater *or equal* then we
> go to the VMX case. Or am I reading it backward?
>
> So we should say "if the size is 4K or more we do VMX" shouldn't we?
Yes. Again I need reword the comment to "equal or greater than 4K"
here.
Thanks,
- Simon
More information about the Linuxppc-dev
mailing list