]> granicus.if.org Git - libx264/commitdiff
ppc: Use xxpermdi in sad_x3/x4 and use macros to avoid redundant code
authorAlexandra Hájková <alexandra.khirnova@gmail.com>
Sat, 8 Sep 2018 07:15:53 +0000 (07:15 +0000)
committerAnton Mitrofanov <BugMaster@narod.ru>
Wed, 6 Mar 2019 19:45:51 +0000 (22:45 +0300)
common/ppc/pixel.c

index 35dd934e10a54705f55f6dd05b336560fc6cdcac..0ccf78861dc2ca056171b6a487e24a0eddef4884 100644 (file)
@@ -820,6 +820,82 @@ static void pixel_sad_x4_16x8_altivec( uint8_t *fenc, uint8_t *pix0, uint8_t *pi
     scores[3] = sum3;
 }
 
+#define PROCESS_PIXS                                                                  \
+        vec_u8_t pix0vH = vec_vsx_ld( 0, pix0 );                                      \
+        pix0 += i_stride;                                                             \
+                                                                                      \
+        vec_u8_t pix1vH = vec_vsx_ld( 0, pix1 );                                      \
+        pix1 += i_stride;                                                             \
+                                                                                      \
+        vec_u8_t fencvH = vec_vsx_ld( 0, fenc );                                      \
+        fenc += FENC_STRIDE;                                                          \
+                                                                                      \
+        vec_u8_t pix2vH = vec_vsx_ld( 0, pix2 );                                      \
+        pix2 += i_stride;                                                             \
+                                                                                      \
+        vec_u8_t pix0vL = vec_vsx_ld( 0, pix0 );                                      \
+        pix0 += i_stride;                                                             \
+                                                                                      \
+        vec_u8_t pix1vL = vec_vsx_ld( 0, pix1 );                                      \
+        pix1 += i_stride;                                                             \
+                                                                                      \
+        vec_u8_t fencvL = vec_vsx_ld( 0, fenc );                                      \
+        fenc += FENC_STRIDE;                                                          \
+                                                                                      \
+        vec_u8_t pix2vL = vec_vsx_ld( 0, pix2 );                                      \
+        pix2 += i_stride;                                                             \
+                                                                                      \
+        fencv = xxpermdi( fencvH, fencvL, 0 );                                        \
+        pix0v = xxpermdi( pix0vH, pix0vL, 0 );                                        \
+        pix1v = xxpermdi( pix1vH, pix1vL, 0 );                                        \
+        pix2v = xxpermdi( pix2vH, pix2vL, 0 );                                        \
+                                                                                      \
+        sum0v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix0v ), (vec_u32_t) sum0v ); \
+        sum1v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix1v ), (vec_u32_t) sum1v ); \
+        sum2v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix2v ), (vec_u32_t) sum2v );
+
+#define PIXEL_SAD_X3_ALTIVEC( name, ly )            \
+static int name( uint8_t *fenc, uint8_t *pix0,      \
+                 uint8_t *pix1, uint8_t *pix2,      \
+                 intptr_t i_stride, int scores[3] ) \
+{                                                   \
+    ALIGNED_16( int sum0 );                         \
+    ALIGNED_16( int sum1 );                         \
+    ALIGNED_16( int sum2 );                         \
+                                                    \
+    LOAD_ZERO;                                      \
+    vec_u8_t fencv, pix0v, pix1v, pix2v;            \
+    vec_s32_t sum0v, sum1v, sum2v;                  \
+                                                    \
+    sum0v = vec_splat_s32( 0 );                     \
+    sum1v = vec_splat_s32( 0 );                     \
+    sum2v = vec_splat_s32( 0 );                     \
+                                                    \
+    for( int y = 0; y < ly; y++ )                   \
+    {                                               \
+        PROCESS_PIXS                                \
+    }                                               \
+                                                    \
+    sum0v = vec_sums( sum0v, zero_s32v );           \
+    sum1v = vec_sums( sum1v, zero_s32v );           \
+    sum2v = vec_sums( sum2v, zero_s32v );           \
+                                                    \
+    sum0v = vec_splat( sum0v, 3 );                  \
+    sum1v = vec_splat( sum1v, 3 );                  \
+    sum2v = vec_splat( sum2v, 3 );                  \
+                                                    \
+    vec_ste( sum0v, 0, &sum0 );                     \
+    vec_ste( sum1v, 0, &sum1 );                     \
+    vec_ste( sum2v, 0, &sum2 );                     \
+                                                    \
+    scores[0] = sum0;                               \
+    scores[1] = sum1;                               \
+    scores[2] = sum2;                               \
+}
+
+PIXEL_SAD_X3_ALTIVEC( pixel_sad_x3_8x8_altivec, 4 )
+PIXEL_SAD_X3_ALTIVEC( pixel_sad_x3_8x16_altivec, 8 )
+
 static void pixel_sad_x3_16x8_altivec( uint8_t *fenc, uint8_t *pix0,
                                        uint8_t *pix1, uint8_t *pix2,
                                        intptr_t i_stride, int scores[3] )
@@ -888,300 +964,49 @@ static void pixel_sad_x3_16x8_altivec( uint8_t *fenc, uint8_t *pix0,
     scores[2] = sum2;
 }
 
-
-static void pixel_sad_x4_8x16_altivec( uint8_t *fenc,
-                                       uint8_t *pix0, uint8_t *pix1,
-                                       uint8_t *pix2, uint8_t *pix3,
-                                       intptr_t i_stride, int scores[4] )
-{
-    ALIGNED_16( int sum0 );
-    ALIGNED_16( int sum1 );
-    ALIGNED_16( int sum2 );
-    ALIGNED_16( int sum3 );
-
-    LOAD_ZERO;
-    vec_u8_t fencv, pix0v, pix1v, pix2v, pix3v;
-    vec_s32_t sum0v, sum1v, sum2v, sum3v;
-
-    sum0v = vec_splat_s32(0);
-    sum1v = vec_splat_s32(0);
-    sum2v = vec_splat_s32(0);
-    sum3v = vec_splat_s32(0);
-
-    for( int y = 0; y < 8; y++ )
-    {
-        pix0v = vec_vsx_ld(0, pix0);
-        pix0 += i_stride;
-
-        pix1v = vec_vsx_ld(0, pix1);
-        pix1 += i_stride;
-
-        fencv = vec_vsx_ld(0, fenc);
-        fenc += FENC_STRIDE;
-
-        pix2v = vec_vsx_ld(0, pix2);
-        pix2 += i_stride;
-
-        pix3v = vec_vsx_ld(0, pix3);
-        pix3 += i_stride;
-
-        sum0v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix0v ), (vec_u32_t) sum0v );
-        sum1v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix1v ), (vec_u32_t) sum1v );
-        sum2v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix2v ), (vec_u32_t) sum2v );
-        sum3v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix3v ), (vec_u32_t) sum3v );
-
-        pix0v = vec_vsx_ld(0, pix0);
-        pix0 += i_stride;
-
-        pix1v = vec_vsx_ld(0, pix1);
-        pix1 += i_stride;
-
-        fencv = vec_vsx_ld(0, fenc);
-        fenc += FENC_STRIDE;
-
-        pix2v = vec_vsx_ld(0, pix2);
-        pix2 += i_stride;
-
-        pix3v = vec_vsx_ld(0, pix3);
-        pix3 += i_stride;
-
-        sum0v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix0v ), (vec_u32_t) sum0v );
-        sum1v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix1v ), (vec_u32_t) sum1v );
-        sum2v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix2v ), (vec_u32_t) sum2v );
-        sum3v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix3v ), (vec_u32_t) sum3v );
-    }
-
-    sum0v = vec_sum2s( sum0v, zero_s32v );
-    sum1v = vec_sum2s( sum1v, zero_s32v );
-    sum2v = vec_sum2s( sum2v, zero_s32v );
-    sum3v = vec_sum2s( sum3v, zero_s32v );
-
-    sum0v = vec_splat( sum0v, 1 );
-    sum1v = vec_splat( sum1v, 1 );
-    sum2v = vec_splat( sum2v, 1 );
-    sum3v = vec_splat( sum3v, 1 );
-
-    vec_ste( sum0v, 0, &sum0);
-    vec_ste( sum1v, 0, &sum1);
-    vec_ste( sum2v, 0, &sum2);
-    vec_ste( sum3v, 0, &sum3);
-
-    scores[0] = sum0;
-    scores[1] = sum1;
-    scores[2] = sum2;
-    scores[3] = sum3;
-}
-
-static void pixel_sad_x3_8x16_altivec( uint8_t *fenc, uint8_t *pix0,
-                                       uint8_t *pix1, uint8_t *pix2,
-                                       intptr_t i_stride, int scores[3] )
-{
-    ALIGNED_16( int sum0 );
-    ALIGNED_16( int sum1 );
-    ALIGNED_16( int sum2 );
-
-    LOAD_ZERO;
-    vec_u8_t fencv, pix0v, pix1v, pix2v;
-    vec_s32_t sum0v, sum1v, sum2v;
-
-    sum0v = vec_splat_s32(0);
-    sum1v = vec_splat_s32(0);
-    sum2v = vec_splat_s32(0);
-
-    for( int y = 0; y < 8; y++ )
-    {
-        pix0v = vec_vsx_ld(0, pix0);
-        pix0 += i_stride;
-
-        pix1v = vec_vsx_ld(0, pix1);
-        pix1 += i_stride;
-
-        fencv = vec_vsx_ld(0, fenc);
-        fenc += FENC_STRIDE;
-
-        pix2v = vec_vsx_ld(0, pix2);
-        pix2 += i_stride;
-
-        sum0v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix0v ), (vec_u32_t) sum0v );
-        sum1v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix1v ), (vec_u32_t) sum1v );
-        sum2v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix2v ), (vec_u32_t) sum2v );
-
-        pix0v = vec_vsx_ld(0, pix0);
-        pix0 += i_stride;
-
-        pix1v = vec_vsx_ld(0, pix1);
-        pix1 += i_stride;
-
-        fencv = vec_vsx_ld(0, fenc);
-        fenc += FENC_STRIDE;
-
-        pix2v = vec_vsx_ld(0, pix2);
-        pix2 += i_stride;
-
-        sum0v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix0v ), (vec_u32_t) sum0v );
-        sum1v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix1v ), (vec_u32_t) sum1v );
-        sum2v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix2v ), (vec_u32_t) sum2v );
-    }
-
-    sum0v = vec_sum2s( sum0v, zero_s32v );
-    sum1v = vec_sum2s( sum1v, zero_s32v );
-    sum2v = vec_sum2s( sum2v, zero_s32v );
-
-    sum0v = vec_splat( sum0v, 1 );
-    sum1v = vec_splat( sum1v, 1 );
-    sum2v = vec_splat( sum2v, 1 );
-
-    vec_ste( sum0v, 0, &sum0);
-    vec_ste( sum1v, 0, &sum1);
-    vec_ste( sum2v, 0, &sum2);
-
-    scores[0] = sum0;
-    scores[1] = sum1;
-    scores[2] = sum2;
-}
-
-static void pixel_sad_x4_8x8_altivec( uint8_t *fenc,
-                                      uint8_t *pix0, uint8_t *pix1,
-                                      uint8_t *pix2, uint8_t *pix3,
-                                      intptr_t i_stride, int scores[4] )
-{
-    ALIGNED_16( int sum0 );
-    ALIGNED_16( int sum1 );
-    ALIGNED_16( int sum2 );
-    ALIGNED_16( int sum3 );
-
-    LOAD_ZERO;
-    vec_u8_t fencv, pix0v, pix1v, pix2v, pix3v;
-    vec_s32_t sum0v, sum1v, sum2v, sum3v;
-
-    sum0v = vec_splat_s32(0);
-    sum1v = vec_splat_s32(0);
-    sum2v = vec_splat_s32(0);
-    sum3v = vec_splat_s32(0);
-
-    for( int y = 0; y < 4; y++ )
-    {
-        vec_u8_t pix0vH = vec_vsx_ld( 0, pix0 );
-        pix0 += i_stride;
-
-        vec_u8_t pix1vH = vec_vsx_ld( 0, pix1 );
-        pix1 += i_stride;
-
-        vec_u8_t fencvH = vec_vsx_ld( 0, fenc );
-        fenc += FENC_STRIDE;
-
-        vec_u8_t pix2vH = vec_vsx_ld( 0, pix2 );
-        pix2 += i_stride;
-
-        vec_u8_t pix3vH = vec_vsx_ld( 0, pix3 );
-        pix3 += i_stride;
-
-        vec_u8_t pix0vL = vec_vsx_ld( 0, pix0 );
-        pix0 += i_stride;
-
-        vec_u8_t pix1vL = vec_vsx_ld( 0, pix1 );
-        pix1 += i_stride;
-
-        vec_u8_t fencvL = vec_vsx_ld( 0, fenc );
-        fenc += FENC_STRIDE;
-
-        vec_u8_t pix2vL = vec_vsx_ld( 0, pix2 );
-        pix2 += i_stride;
-
-        vec_u8_t pix3vL = vec_vsx_ld( 0, pix3 );
-        pix3 += i_stride;
-
-        fencv = xxpermdi( fencvH, fencvL, 0 );
-        pix0v = xxpermdi( pix0vH, pix0vL, 0 );
-        pix1v = xxpermdi( pix1vH, pix1vL, 0 );
-        pix2v = xxpermdi( pix2vH, pix2vL, 0 );
-        pix3v = xxpermdi( pix3vH, pix3vL, 0 );
-
-        sum0v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix0v ), (vec_u32_t) sum0v );
-        sum1v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix1v ), (vec_u32_t) sum1v );
-        sum2v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix2v ), (vec_u32_t) sum2v );
-        sum3v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix3v ), (vec_u32_t) sum3v );
-    }
-
-    sum0v = vec_sums( sum0v, zero_s32v );
-    sum1v = vec_sums( sum1v, zero_s32v );
-    sum2v = vec_sums( sum2v, zero_s32v );
-    sum3v = vec_sums( sum3v, zero_s32v );
-
-    vec_s32_t s01 = vec_mergel( sum0v, sum1v );
-    vec_s32_t s23 = vec_mergel( sum2v, sum3v );
-    vec_s32_t s = xxpermdi( s01, s23, 3 );
-
-    vec_vsx_st( s, 0, scores );
+#define PIXEL_SAD_X4_ALTIVEC( name, ly )                                              \
+static int name( uint8_t *fenc,                                                       \
+                 uint8_t *pix0, uint8_t *pix1,                                        \
+                 uint8_t *pix2, uint8_t *pix3,                                        \
+                 intptr_t i_stride, int scores[4] )                                   \
+{                                                                                     \
+    ALIGNED_16( int sum0 );                                                           \
+    ALIGNED_16( int sum1 );                                                           \
+    ALIGNED_16( int sum2 );                                                           \
+                                                                                      \
+    LOAD_ZERO;                                                                        \
+    vec_u8_t fencv, pix0v, pix1v, pix2v, pix3v;                                       \
+    vec_s32_t sum0v, sum1v, sum2v, sum3v;                                             \
+                                                                                      \
+    sum0v = vec_splat_s32( 0 );                                                       \
+    sum1v = vec_splat_s32( 0 );                                                       \
+    sum2v = vec_splat_s32( 0 );                                                       \
+                                                                                      \
+    for( int y = 0; y < ly; y++ )                                                     \
+    {                                                                                 \
+        PROCESS_PIXS                                                                  \
+        vec_u8_t pix3vH = vec_vsx_ld( 0, pix3 );                                      \
+        pix3 += i_stride;                                                             \
+        vec_u8_t pix3vL = vec_vsx_ld( 0, pix3 );                                      \
+        pix3 += i_stride;                                                             \
+        pix3v = xxpermdi( pix3vH, pix3vL, 0 );                                        \
+        sum3v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix3v ), (vec_u32_t) sum3v ); \
+    }                                                                                 \
+                                                                                      \
+    sum0v = vec_sums( sum0v, zero_s32v );                                             \
+    sum1v = vec_sums( sum1v, zero_s32v );                                             \
+    sum2v = vec_sums( sum2v, zero_s32v );                                             \
+    sum3v = vec_sums( sum3v, zero_s32v );                                             \
+                                                                                      \
+    vec_s32_t s01 = vec_mergel( sum0v, sum1v );                                       \
+    vec_s32_t s23 = vec_mergel( sum2v, sum3v );                                       \
+    vec_s32_t s = xxpermdi( s01, s23, 3 );                                            \
+                                                                                      \
+    vec_vsx_st( s, 0, scores );                                                       \
 }
 
-static void pixel_sad_x3_8x8_altivec( uint8_t *fenc, uint8_t *pix0,
-                                      uint8_t *pix1, uint8_t *pix2,
-                                      intptr_t i_stride, int scores[3] )
-{
-    ALIGNED_16( int sum0 );
-    ALIGNED_16( int sum1 );
-    ALIGNED_16( int sum2 );
-
-    LOAD_ZERO;
-    vec_u8_t fencv, pix0v, pix1v, pix2v;
-    vec_s32_t sum0v, sum1v, sum2v;
-
-    sum0v = vec_splat_s32(0);
-    sum1v = vec_splat_s32(0);
-    sum2v = vec_splat_s32(0);
-
-    for( int y = 0; y < 4; y++ )
-    {
-        pix0v = vec_vsx_ld(0, pix0);
-        pix0 += i_stride;
-
-        pix1v = vec_vsx_ld(0, pix1);
-        pix1 += i_stride;
-
-        fencv = vec_vsx_ld(0, fenc);
-        fenc += FENC_STRIDE;
-
-        pix2v = vec_vsx_ld(0, pix2);
-        pix2 += i_stride;
-
-        sum0v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix0v ), (vec_u32_t) sum0v );
-        sum1v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix1v ), (vec_u32_t) sum1v );
-        sum2v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix2v ), (vec_u32_t) sum2v );
-
-        pix0v = vec_vsx_ld(0, pix0);
-        pix0 += i_stride;
-
-        pix1v = vec_vsx_ld(0, pix1);
-        pix1 += i_stride;
-
-        fencv = vec_vsx_ld(0, fenc);
-        fenc += FENC_STRIDE;
-
-        pix2v = vec_vsx_ld(0, pix2);
-        pix2 += i_stride;
-
-        sum0v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix0v ), (vec_u32_t) sum0v );
-        sum1v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix1v ), (vec_u32_t) sum1v );
-        sum2v = (vec_s32_t) vec_sum4s( vec_absd( fencv, pix2v ), (vec_u32_t) sum2v );
-    }
-
-    sum0v = vec_sum2s( sum0v, zero_s32v );
-    sum1v = vec_sum2s( sum1v, zero_s32v );
-    sum2v = vec_sum2s( sum2v, zero_s32v );
-
-    sum0v = vec_splat( sum0v, 1 );
-    sum1v = vec_splat( sum1v, 1 );
-    sum2v = vec_splat( sum2v, 1 );
-
-    vec_ste( sum0v, 0, &sum0);
-    vec_ste( sum1v, 0, &sum1);
-    vec_ste( sum2v, 0, &sum2);
-
-    scores[0] = sum0;
-    scores[1] = sum1;
-    scores[2] = sum2;
-}
+PIXEL_SAD_X4_ALTIVEC( pixel_sad_x4_8x8_altivec, 4 )
+PIXEL_SAD_X4_ALTIVEC( pixel_sad_x4_8x16_altivec, 8 )
 
 /***********************************************************************
 * SSD routines