Hi Simon,

wei.guo.si...@gmail.com writes:
> From: Simon Guo <wei.guo.si...@gmail.com>
>
> This patch add VMX primitives to do memcmp() in case the compare size
> exceeds 4K bytes. KSM feature can benefit from this.

You say "exceeds 4K" here.

> diff --git a/arch/powerpc/lib/memcmp_64.S b/arch/powerpc/lib/memcmp_64.S
> index f20e883..6303bbf 100644
> --- a/arch/powerpc/lib/memcmp_64.S
> +++ b/arch/powerpc/lib/memcmp_64.S
> @@ -27,12 +27,73 @@
>  #define LH   lhbrx
>  #define LW   lwbrx
>  #define LD   ldbrx
> +#define LVS  lvsr
> +#define VPERM(_VRT,_VRA,_VRB,_VRC) \
> +     vperm _VRT,_VRB,_VRA,_VRC
>  #else
>  #define LH   lhzx
>  #define LW   lwzx
>  #define LD   ldx
> +#define LVS  lvsl
> +#define VPERM(_VRT,_VRA,_VRB,_VRC) \
> +     vperm _VRT,_VRA,_VRB,_VRC
>  #endif
>  
> +#define VMX_OPS_THRES 4096

THRES == 4096

BTW, can we call it VMX_THRESH ?

> +#define ENTER_VMX_OPS        \
> +     mflr    r0;     \
> +     std     r3,-STACKFRAMESIZE+STK_REG(R31)(r1); \
> +     std     r4,-STACKFRAMESIZE+STK_REG(R30)(r1); \
> +     std     r5,-STACKFRAMESIZE+STK_REG(R29)(r1); \
> +     std     r0,16(r1); \
> +     stdu    r1,-STACKFRAMESIZE(r1); \
> +     bl      enter_vmx_ops; \
> +     cmpwi   cr1,r3,0; \
> +     ld      r0,STACKFRAMESIZE+16(r1); \
> +     ld      r3,STK_REG(R31)(r1); \
> +     ld      r4,STK_REG(R30)(r1); \
> +     ld      r5,STK_REG(R29)(r1); \
> +     addi    r1,r1,STACKFRAMESIZE; \
> +     mtlr    r0
> +
> +#define EXIT_VMX_OPS \
> +     mflr    r0; \
> +     std     r3,-STACKFRAMESIZE+STK_REG(R31)(r1); \
> +     std     r4,-STACKFRAMESIZE+STK_REG(R30)(r1); \
> +     std     r5,-STACKFRAMESIZE+STK_REG(R29)(r1); \
> +     std     r0,16(r1); \
> +     stdu    r1,-STACKFRAMESIZE(r1); \
> +     bl      exit_vmx_ops; \
> +     ld      r0,STACKFRAMESIZE+16(r1); \
> +     ld      r3,STK_REG(R31)(r1); \
> +     ld      r4,STK_REG(R30)(r1); \
> +     ld      r5,STK_REG(R29)(r1); \
> +     addi    r1,r1,STACKFRAMESIZE; \
> +     mtlr    r0
> +
> +/*
> + * LD_VSR_CROSS16B load the 2nd 16 bytes for _vaddr which is unaligned with
> + * 16 bytes boundary and permute the result with the 1st 16 bytes.
> +
> + *    |  y y y y y y y y y y y y y 0 1 2 | 3 4 5 6 7 8 9 a b c d e f z z z |
> + *    ^                                  ^                                 ^
> + * 0xbbbb10                          0xbbbb20                          
> 0xbbb30
> + *                                 ^
> + *                                _vaddr
> + *
> + *
> + * _vmask is the mask generated by LVS
> + * _v1st_qw is the 1st aligned QW of current addr which is already loaded.
> + *   for example: 0xyyyyyyyyyyyyy012 for big endian
> + * _v2nd_qw is the 2nd aligned QW of cur _vaddr to be loaded.
> + *   for example: 0x3456789abcdefzzz for big endian
> + * The permute result is saved in _v_res.
> + *   for example: 0x0123456789abcdef for big endian.
> + */
> +#define LD_VSR_CROSS16B(_vaddr,_vmask,_v1st_qw,_v2nd_qw,_v_res) \
> +        lvx     _v2nd_qw,_vaddr,off16; \
> +        VPERM(_v_res,_v1st_qw,_v2nd_qw,_vmask)
> +
>  /*
>   * There are 2 categories for memcmp:
>   * 1) src/dst has the same offset to the 8 bytes boundary. The handlers
> @@ -174,6 +235,13 @@ _GLOBAL(memcmp)
>       blr
>  
>  .Llong:
> +#ifdef CONFIG_ALTIVEC
> +     /* Try to use vmx loop if length is larger than 4K */
> +     cmpldi  cr6,r5,VMX_OPS_THRES
> +     bge     cr6,.Lsameoffset_vmx_cmp

Here we compare the length to 4K and if it's greater *or equal* then we
go to the VMX case. Or am I reading it backward?

So we should say "if the size is 4K or more we do VMX" shouldn't we?

cheers

Reply via email to