---
 libavcodec/x86/h264dsp_mmx.c |  730 ++++++++++++++++++++++--------------------
 1 files changed, 379 insertions(+), 351 deletions(-)

diff --git a/libavcodec/x86/h264dsp_mmx.c b/libavcodec/x86/h264dsp_mmx.c
index f5c7b93..458f348 100644
--- a/libavcodec/x86/h264dsp_mmx.c
+++ b/libavcodec/x86/h264dsp_mmx.c
@@ -23,12 +23,14 @@
 #include "libavcodec/h264dsp.h"
 #include "dsputil_mmx.h"
 
-DECLARE_ALIGNED(8, static const uint64_t, ff_pb_3_1  ) = 0x0103010301030103ULL;
+DECLARE_ALIGNED(8, static const uint64_t, ff_pb_3_1) = 0x0103010301030103ULL;
 
 /***********************************/
 /* IDCT */
-#define IDCT_ADD_FUNC(NUM, DEPTH, OPT) \
-void ff_h264_idct ## NUM ## _add_ ## DEPTH ## _ ## OPT (uint8_t *dst, int16_t 
*block, int stride);
+#define IDCT_ADD_FUNC(NUM, DEPTH, OPT)                                  \
+void ff_h264_idct ## NUM ## _add_ ## DEPTH ## _ ## OPT(uint8_t *dst,    \
+                                                       int16_t *block,  \
+                                                       int stride);
 
 IDCT_ADD_FUNC(, 8, mmx)
 IDCT_ADD_FUNC(, 10, sse2)
@@ -46,10 +48,10 @@ IDCT_ADD_FUNC(8, 10, avx)
 #endif
 
 
-#define IDCT_ADD_REP_FUNC(NUM, REP, DEPTH, OPT) \
-void ff_h264_idct ## NUM ## _add ## REP ## _ ## DEPTH ## _ ## OPT \
-                              (uint8_t *dst, const int *block_offset, \
-                              DCTELEM *block, int stride, const uint8_t 
nnzc[6*8]);
+#define IDCT_ADD_REP_FUNC(NUM, REP, DEPTH, OPT)                         \
+void ff_h264_idct ## NUM ## _add ## REP ## _ ## DEPTH ## _ ## OPT       \
+    (uint8_t *dst, const int *block_offset,                             \
+     DCTELEM *block, int stride, const uint8_t nnzc[6 * 8]);
 
 IDCT_ADD_REP_FUNC(8, 4, 8, mmx)
 IDCT_ADD_REP_FUNC(8, 4, 8, mmx2)
@@ -70,10 +72,11 @@ IDCT_ADD_REP_FUNC(, 16intra, 10, avx)
 #endif
 
 
-#define IDCT_ADD_REP_FUNC2(NUM, REP, DEPTH, OPT) \
-void ff_h264_idct ## NUM ## _add ## REP ## _ ## DEPTH ## _ ## OPT \
-                              (uint8_t **dst, const int *block_offset, \
-                              DCTELEM *block, int stride, const uint8_t 
nnzc[6*8]);
+#define IDCT_ADD_REP_FUNC2(NUM, REP, DEPTH, OPT)                      \
+void ff_h264_idct ## NUM ## _add ## REP ## _ ## DEPTH ## _ ## OPT     \
+    (uint8_t **dst, const int *block_offset,                          \
+     DCTELEM *block, int stride, const uint8_t nnzc[6 * 8]);
+
 IDCT_ADD_REP_FUNC2(, 8, 8, mmx)
 IDCT_ADD_REP_FUNC2(, 8, 8, mmx2)
 IDCT_ADD_REP_FUNC2(, 8, 8, sse2)
@@ -82,151 +85,157 @@ IDCT_ADD_REP_FUNC2(, 8, 10, sse2)
 IDCT_ADD_REP_FUNC2(, 8, 10, avx)
 #endif
 
-void ff_h264_luma_dc_dequant_idct_mmx (DCTELEM *output, DCTELEM *input, int 
qmul);
+void ff_h264_luma_dc_dequant_idct_mmx(DCTELEM *output, DCTELEM *input, int 
qmul);
 void ff_h264_luma_dc_dequant_idct_sse2(DCTELEM *output, DCTELEM *input, int 
qmul);
 
 /***********************************/
 /* deblocking */
 
-#define h264_loop_filter_strength_iteration_mmx2(bS, nz, ref, mv, bidir, 
edges, step, mask_mv, dir, d_idx, mask_dir) \
-    do { \
-        x86_reg b_idx; \
-        mask_mv <<= 3; \
-        for( b_idx=0; b_idx<edges; b_idx+=step ) { \
-            if (!mask_dir) \
-            __asm__ volatile( \
-                    "pxor %%mm0, %%mm0 \n\t" \
-                    :: \
-            ); \
-            if(!(mask_mv & b_idx)) { \
-                if(bidir) { \
-                    __asm__ volatile( \
-                        "movd         %a3(%0,%2), %%mm2 \n" \
+#define h264_loop_filter_strength_iteration_mmx2(bS, nz, ref, mv,       \
+                                                 bidir, edges, step,    \
+                                                 mask_mv, dir, d_idx,   \
+                                                 mask_dir)              \
+    do {                                                                \
+        x86_reg b_idx;                                                  \
+        mask_mv <<= 3;                                                  \
+        for (b_idx = 0; b_idx < edges; b_idx += step) {                 \
+            if (!mask_dir)                                              \
+                __asm__ volatile (                                      \
+                    "pxor %%mm0, %%mm0 \n\t"                            \
+                    ::                                                  \
+                    );                                                  \
+            if (!(mask_mv & b_idx)) {                                   \
+                if (bidir) {                                            \
+                    __asm__ volatile (                                  \
+                        "movd         %a3(%0,%2), %%mm2 \n"             \
                         "punpckldq    %a4(%0,%2), %%mm2 \n" /* { ref0[bn], 
ref1[bn] } */ \
-                        "pshufw $0x44, 12(%0,%2), %%mm0 \n" /* { ref0[b], 
ref0[b] } */ \
-                        "pshufw $0x44, 52(%0,%2), %%mm1 \n" /* { ref1[b], 
ref1[b] } */ \
-                        "pshufw $0x4E, %%mm2, %%mm3 \n" \
-                        "psubb         %%mm2, %%mm0 \n" /* { 
ref0[b]!=ref0[bn], ref0[b]!=ref1[bn] } */ \
-                        "psubb         %%mm3, %%mm1 \n" /* { 
ref1[b]!=ref1[bn], ref1[b]!=ref0[bn] } */ \
- \
-                        "por           %%mm1, %%mm0 \n" \
-                        "movq   %a5(%1,%2,4), %%mm1 \n" \
-                        "movq   %a6(%1,%2,4), %%mm2 \n" \
-                        "movq          %%mm1, %%mm3 \n" \
-                        "movq          %%mm2, %%mm4 \n" \
-                        "psubw   48(%1,%2,4), %%mm1 \n" \
-                        "psubw   56(%1,%2,4), %%mm2 \n" \
-                        "psubw  208(%1,%2,4), %%mm3 \n" \
-                        "psubw  216(%1,%2,4), %%mm4 \n" \
-                        "packsswb      %%mm2, %%mm1 \n" \
-                        "packsswb      %%mm4, %%mm3 \n" \
-                        "paddb         %%mm6, %%mm1 \n" \
-                        "paddb         %%mm6, %%mm3 \n" \
+                        "pshufw $0x44, 12(%0,%2), %%mm0 \n" /* { ref0[b],  
ref0[b] } */  \
+                        "pshufw $0x44, 52(%0,%2), %%mm1 \n" /* { ref1[b],  
ref1[b] } */  \
+                        "pshufw $0x4E, %%mm2, %%mm3 \n"                 \
+                        "psubb         %%mm2, %%mm0 \n" /* { ref0[b] != 
ref0[bn], ref0[b] != ref1[bn] } */ \
+                        "psubb         %%mm3, %%mm1 \n" /* { ref1[b] != 
ref1[bn], ref1[b] != ref0[bn] } */ \
+                                                                        \
+                        "por           %%mm1, %%mm0 \n"                 \
+                        "movq   %a5(%1,%2,4), %%mm1 \n"                 \
+                        "movq   %a6(%1,%2,4), %%mm2 \n"                 \
+                        "movq          %%mm1, %%mm3 \n"                 \
+                        "movq          %%mm2, %%mm4 \n"                 \
+                        "psubw   48(%1,%2,4), %%mm1 \n"                 \
+                        "psubw   56(%1,%2,4), %%mm2 \n"                 \
+                        "psubw  208(%1,%2,4), %%mm3 \n"                 \
+                        "psubw  216(%1,%2,4), %%mm4 \n"                 \
+                        "packsswb      %%mm2, %%mm1 \n"                 \
+                        "packsswb      %%mm4, %%mm3 \n"                 \
+                        "paddb         %%mm6, %%mm1 \n"                 \
+                        "paddb         %%mm6, %%mm3 \n"                 \
                         "psubusb       %%mm5, %%mm1 \n" /* abs(mv[b] - mv[bn]) 
>= limit */ \
-                        "psubusb       %%mm5, %%mm3 \n" \
-                        "packsswb      %%mm3, %%mm1 \n" \
- \
-                        "por           %%mm1, %%mm0 \n" \
-                        "movq   %a7(%1,%2,4), %%mm1 \n" \
-                        "movq   %a8(%1,%2,4), %%mm2 \n" \
-                        "movq          %%mm1, %%mm3 \n" \
-                        "movq          %%mm2, %%mm4 \n" \
-                        "psubw   48(%1,%2,4), %%mm1 \n" \
-                        "psubw   56(%1,%2,4), %%mm2 \n" \
-                        "psubw  208(%1,%2,4), %%mm3 \n" \
-                        "psubw  216(%1,%2,4), %%mm4 \n" \
-                        "packsswb      %%mm2, %%mm1 \n" \
-                        "packsswb      %%mm4, %%mm3 \n" \
-                        "paddb         %%mm6, %%mm1 \n" \
-                        "paddb         %%mm6, %%mm3 \n" \
+                        "psubusb       %%mm5, %%mm3 \n"                 \
+                        "packsswb      %%mm3, %%mm1 \n"                 \
+                                                                        \
+                        "por           %%mm1, %%mm0 \n"                 \
+                        "movq   %a7(%1,%2,4), %%mm1 \n"                 \
+                        "movq   %a8(%1,%2,4), %%mm2 \n"                 \
+                        "movq          %%mm1, %%mm3 \n"                 \
+                        "movq          %%mm2, %%mm4 \n"                 \
+                        "psubw   48(%1,%2,4), %%mm1 \n"                 \
+                        "psubw   56(%1,%2,4), %%mm2 \n"                 \
+                        "psubw  208(%1,%2,4), %%mm3 \n"                 \
+                        "psubw  216(%1,%2,4), %%mm4 \n"                 \
+                        "packsswb      %%mm2, %%mm1 \n"                 \
+                        "packsswb      %%mm4, %%mm3 \n"                 \
+                        "paddb         %%mm6, %%mm1 \n"                 \
+                        "paddb         %%mm6, %%mm3 \n"                 \
                         "psubusb       %%mm5, %%mm1 \n" /* abs(mv[b] - mv[bn]) 
>= limit */ \
-                        "psubusb       %%mm5, %%mm3 \n" \
-                        "packsswb      %%mm3, %%mm1 \n" \
- \
-                        "pshufw $0x4E, %%mm1, %%mm1 \n" \
-                        "por           %%mm1, %%mm0 \n" \
-                        "pshufw $0x4E, %%mm0, %%mm1 \n" \
-                        "pminub        %%mm1, %%mm0 \n" \
-                        ::"r"(ref), \
-                          "r"(mv), \
-                          "r"(b_idx), \
-                          "i"(d_idx+12), \
-                          "i"(d_idx+52), \
-                          "i"(d_idx*4+48), \
-                          "i"(d_idx*4+56), \
-                          "i"(d_idx*4+208), \
-                          "i"(d_idx*4+216) \
-                    ); \
-                } else { \
-                    __asm__ volatile( \
-                        "movd   12(%0,%2), %%mm0 \n" \
+                        "psubusb       %%mm5, %%mm3 \n"                 \
+                        "packsswb      %%mm3, %%mm1 \n"                 \
+                                                                        \
+                        "pshufw $0x4E, %%mm1, %%mm1 \n"                 \
+                        "por           %%mm1, %%mm0 \n"                 \
+                        "pshufw $0x4E, %%mm0, %%mm1 \n"                 \
+                        "pminub        %%mm1, %%mm0 \n"                 \
+                        :: "r" (ref),                                   \
+                         "r" (mv),                                      \
+                         "r" (b_idx),                                   \
+                         "i" (d_idx + 12),                              \
+                         "i" (d_idx + 52),                              \
+                         "i" (d_idx * 4 + 48),                          \
+                         "i" (d_idx * 4 + 56),                          \
+                         "i" (d_idx * 4 + 208),                         \
+                         "i" (d_idx * 4 + 216)                          \
+                        );                                              \
+                } else {                                                \
+                    __asm__ volatile (                                  \
+                        "movd   12(%0,%2), %%mm0 \n"                    \
                         "psubb %a3(%0,%2), %%mm0 \n" /* ref[b] != ref[bn] */ \
-                        "movq   48(%1,%2,4), %%mm1 \n" \
-                        "movq   56(%1,%2,4), %%mm2 \n" \
-                        "psubw %a4(%1,%2,4), %%mm1 \n" \
-                        "psubw %a5(%1,%2,4), %%mm2 \n" \
-                        "packsswb   %%mm2, %%mm1 \n" \
-                        "paddb      %%mm6, %%mm1 \n" \
+                        "movq   48(%1,%2,4), %%mm1 \n"                  \
+                        "movq   56(%1,%2,4), %%mm2 \n"                  \
+                        "psubw %a4(%1,%2,4), %%mm1 \n"                  \
+                        "psubw %a5(%1,%2,4), %%mm2 \n"                  \
+                        "packsswb   %%mm2, %%mm1 \n"                    \
+                        "paddb      %%mm6, %%mm1 \n"                    \
                         "psubusb    %%mm5, %%mm1 \n" /* abs(mv[b] - mv[bn]) >= 
limit */ \
-                        "packsswb   %%mm1, %%mm1 \n" \
-                        "por        %%mm1, %%mm0 \n" \
-                        ::"r"(ref), \
-                          "r"(mv), \
-                          "r"(b_idx), \
-                          "i"(d_idx+12), \
-                          "i"(d_idx*4+48), \
-                          "i"(d_idx*4+56) \
-                    ); \
-                } \
-            } \
-            __asm__ volatile( \
-                "movd 12(%0,%1), %%mm1 \n" \
-                "por %a2(%0,%1), %%mm1 \n" /* nnz[b] || nnz[bn] */ \
-                ::"r"(nnz), \
-                  "r"(b_idx), \
-                  "i"(d_idx+12) \
-            ); \
-            __asm__ volatile( \
-                "pminub    %%mm7, %%mm1 \n" \
-                "pminub    %%mm7, %%mm0 \n" \
-                "psllw        $1, %%mm1 \n" \
-                "pxor      %%mm2, %%mm2 \n" \
-                "pmaxub    %%mm0, %%mm1 \n" \
-                "punpcklbw %%mm2, %%mm1 \n" \
-                "movq      %%mm1, %a1(%0,%2) \n" \
-                ::"r"(bS), \
-                  "i"(32*dir), \
-                  "r"(b_idx) \
-                :"memory" \
-            ); \
-        } \
+                        "packsswb   %%mm1, %%mm1 \n"                    \
+                        "por        %%mm1, %%mm0 \n"                    \
+                        :: "r" (ref),                                   \
+                         "r" (mv),                                      \
+                         "r" (b_idx),                                   \
+                         "i" (d_idx + 12),                              \
+                         "i" (d_idx * 4 + 48),                          \
+                         "i" (d_idx * 4 + 56)                           \
+                        );                                              \
+                }                                                       \
+            }                                                           \
+            __asm__ volatile (                                          \
+                "movd 12(%0,%1), %%mm1 \n"                              \
+                "por %a2(%0,%1), %%mm1 \n" /* nnz[b] || nnz[bn] */      \
+                :: "r" (nnz),                                           \
+                 "r" (b_idx),                                           \
+                 "i" (d_idx + 12)                                       \
+                );                                                      \
+            __asm__ volatile (                                          \
+                "pminub    %%mm7, %%mm1 \n"                             \
+                "pminub    %%mm7, %%mm0 \n"                             \
+                "psllw        $1, %%mm1 \n"                             \
+                "pxor      %%mm2, %%mm2 \n"                             \
+                "pmaxub    %%mm0, %%mm1 \n"                             \
+                "punpcklbw %%mm2, %%mm1 \n"                             \
+                "movq      %%mm1, %a1(%0,%2) \n"                        \
+                :: "r" (bS),                                            \
+                 "i" (32 * dir),                                        \
+                 "r" (b_idx)                                            \
+                : "memory"                                              \
+                );                                                      \
+        }                                                               \
     } while (0)
 
-static void h264_loop_filter_strength_mmx2( int16_t bS[2][4][4], uint8_t 
nnz[40], int8_t ref[2][40], int16_t mv[2][40][2],
-                                            int bidir, int edges, int step, 
int mask_mv0, int mask_mv1, int field ) {
-    __asm__ volatile(
+static void h264_loop_filter_strength_mmx2(int16_t bS[2][4][4], uint8_t 
nnz[40],
+                                           int8_t ref[2][40], int16_t 
mv[2][40][2],
+                                           int bidir, int edges, int step,
+                                           int mask_mv0, int mask_mv1, int 
field)
+{
+    __asm__ volatile (
         "movq %0, %%mm7 \n"
         "movq %1, %%mm6 \n"
-        ::"m"(ff_pb_1), "m"(ff_pb_3)
-    );
-    if(field)
-        __asm__ volatile(
-            "movq %0, %%mm6 \n"
-            ::"m"(ff_pb_3_1)
+        :: "m" (ff_pb_1), "m" (ff_pb_3)
         );
-    __asm__ volatile(
+    if (field)
+        __asm__ volatile (
+            "movq %0, %%mm6 \n"
+            :: "m" (ff_pb_3_1)
+            );
+    __asm__ volatile (
         "movq  %%mm6, %%mm5 \n"
         "paddb %%mm5, %%mm5 \n"
-    :);
+        :);
 
-    // could do a special case for dir==0 && edges==1, but it only reduces the
-    // average filter time by 1.2%
+    /* could do a special case for dir==0 && edges==1, but it only reduces the
+     * average filter time by 1.2% */
     step  <<= 3;
     edges <<= 3;
     h264_loop_filter_strength_iteration_mmx2(bS, nnz, ref, mv, bidir, edges, 
step, mask_mv1, 1, -8,  0);
     h264_loop_filter_strength_iteration_mmx2(bS, nnz, ref, mv, bidir,    32,   
 8, mask_mv0, 0, -1, -1);
 
-    __asm__ volatile(
+    __asm__ volatile (
         "movq   (%0), %%mm0 \n\t"
         "movq  8(%0), %%mm1 \n\t"
         "movq 16(%0), %%mm2 \n\t"
@@ -236,278 +245,297 @@ static void h264_loop_filter_strength_mmx2( int16_t 
bS[2][4][4], uint8_t nnz[40]
         "movq %%mm3,  8(%0) \n\t"
         "movq %%mm4, 16(%0) \n\t"
         "movq %%mm2, 24(%0) \n\t"
-        ::"r"(bS[0])
-        :"memory"
-    );
+        :: "r" (bS[0])
+        : "memory"
+        );
 }
 
-#define LF_FUNC(DIR, TYPE, DEPTH, OPT) \
-void ff_deblock_ ## DIR ## _ ## TYPE ## _ ## DEPTH ## _ ## OPT (uint8_t *pix, 
int stride, \
-                                                                int alpha, int 
beta, int8_t *tc0);
+#define LF_FUNC(DIR, TYPE, DEPTH, OPT)                                        \
+void ff_deblock_ ## DIR ## _ ## TYPE ## _ ## DEPTH ## _ ## OPT(uint8_t *pix,  \
+                                                               int stride,    \
+                                                               int alpha,     \
+                                                               int beta,      \
+                                                               int8_t *tc0);
 #define LF_IFUNC(DIR, TYPE, DEPTH, OPT) \
-void ff_deblock_ ## DIR ## _ ## TYPE ## _ ## DEPTH ## _ ## OPT (uint8_t *pix, 
int stride, \
-                                                                int alpha, int 
beta);
-
-#define LF_FUNCS(type, depth)\
-LF_FUNC (h,  chroma,       depth, mmx2)\
-LF_IFUNC(h,  chroma_intra, depth, mmx2)\
-LF_FUNC (v,  chroma,       depth, mmx2)\
-LF_IFUNC(v,  chroma_intra, depth, mmx2)\
-LF_FUNC (h,  luma,         depth, mmx2)\
-LF_IFUNC(h,  luma_intra,   depth, mmx2)\
-LF_FUNC (h,  luma,         depth, sse2)\
-LF_IFUNC(h,  luma_intra,   depth, sse2)\
-LF_FUNC (v,  luma,         depth, sse2)\
-LF_IFUNC(v,  luma_intra,   depth, sse2)\
-LF_FUNC (h,  chroma,       depth, sse2)\
-LF_IFUNC(h,  chroma_intra, depth, sse2)\
-LF_FUNC (v,  chroma,       depth, sse2)\
-LF_IFUNC(v,  chroma_intra, depth, sse2)\
-LF_FUNC (h,  luma,         depth,  avx)\
-LF_IFUNC(h,  luma_intra,   depth,  avx)\
-LF_FUNC (v,  luma,         depth,  avx)\
-LF_IFUNC(v,  luma_intra,   depth,  avx)\
-LF_FUNC (h,  chroma,       depth,  avx)\
-LF_IFUNC(h,  chroma_intra, depth,  avx)\
-LF_FUNC (v,  chroma,       depth,  avx)\
-LF_IFUNC(v,  chroma_intra, depth,  avx)
-
-LF_FUNCS( uint8_t,  8)
+void ff_deblock_ ## DIR ## _ ## TYPE ## _ ## DEPTH ## _ ## OPT(uint8_t *pix,  \
+                                                               int stride,    \
+                                                               int alpha,     \
+                                                               int beta);
+
+#define LF_FUNCS(type, depth)                   \
+LF_FUNC(h,  chroma,       depth, mmx2)          \
+LF_IFUNC(h, chroma_intra, depth, mmx2)          \
+LF_FUNC(v,  chroma,       depth, mmx2)          \
+LF_IFUNC(v, chroma_intra, depth, mmx2)          \
+LF_FUNC(h,  luma,         depth, mmx2)          \
+LF_IFUNC(h, luma_intra,   depth, mmx2)          \
+LF_FUNC(h,  luma,         depth, sse2)          \
+LF_IFUNC(h, luma_intra,   depth, sse2)          \
+LF_FUNC(v,  luma,         depth, sse2)          \
+LF_IFUNC(v, luma_intra,   depth, sse2)          \
+LF_FUNC(h,  chroma,       depth, sse2)          \
+LF_IFUNC(h, chroma_intra, depth, sse2)          \
+LF_FUNC(v,  chroma,       depth, sse2)          \
+LF_IFUNC(v, chroma_intra, depth, sse2)          \
+LF_FUNC(h,  luma,         depth, avx)           \
+LF_IFUNC(h, luma_intra,   depth, avx)           \
+LF_FUNC(v,  luma,         depth, avx)           \
+LF_IFUNC(v, luma_intra,   depth, avx)           \
+LF_FUNC(h,  chroma,       depth, avx)           \
+LF_IFUNC(h, chroma_intra, depth, avx)           \
+LF_FUNC(v,  chroma,       depth, avx)           \
+LF_IFUNC(v, chroma_intra, depth, avx)
+
+LF_FUNCS(uint8_t,   8)
 LF_FUNCS(uint16_t, 10)
 
 #if ARCH_X86_32
-LF_FUNC (v8, luma,             8, mmx2)
-static void ff_deblock_v_luma_8_mmx2(uint8_t *pix, int stride, int alpha, int 
beta, int8_t *tc0)
+LF_FUNC(v8, luma, 8, mmx2)
+static void ff_deblock_v_luma_8_mmx2(uint8_t *pix, int stride, int alpha,
+                                     int beta, int8_t *tc0)
 {
-    if((tc0[0] & tc0[1]) >= 0)
-        ff_deblock_v8_luma_8_mmx2(pix+0, stride, alpha, beta, tc0);
-    if((tc0[2] & tc0[3]) >= 0)
-        ff_deblock_v8_luma_8_mmx2(pix+8, stride, alpha, beta, tc0+2);
+    if ((tc0[0] & tc0[1]) >= 0)
+        ff_deblock_v8_luma_8_mmx2(pix + 0, stride, alpha, beta, tc0);
+    if ((tc0[2] & tc0[3]) >= 0)
+        ff_deblock_v8_luma_8_mmx2(pix + 8, stride, alpha, beta, tc0 + 2);
 }
-LF_IFUNC(v8, luma_intra,        8, mmx2)
-static void ff_deblock_v_luma_intra_8_mmx2(uint8_t *pix, int stride, int 
alpha, int beta)
+
+LF_IFUNC(v8, luma_intra, 8, mmx2)
+static void ff_deblock_v_luma_intra_8_mmx2(uint8_t *pix, int stride,
+                                           int alpha, int beta)
 {
-    ff_deblock_v8_luma_intra_8_mmx2(pix+0, stride, alpha, beta);
-    ff_deblock_v8_luma_intra_8_mmx2(pix+8, stride, alpha, beta);
+    ff_deblock_v8_luma_intra_8_mmx2(pix + 0, stride, alpha, beta);
+    ff_deblock_v8_luma_intra_8_mmx2(pix + 8, stride, alpha, beta);
 }
 #endif /* ARCH_X86_32 */
 
-LF_FUNC (v,  luma,            10, mmx2)
-LF_IFUNC(v,  luma_intra,      10, mmx2)
+LF_FUNC(v,  luma,       10, mmx2)
+LF_IFUNC(v, luma_intra, 10, mmx2)
 
 /***********************************/
 /* weighted prediction */
 
-#define H264_WEIGHT(W, OPT) \
-void ff_h264_weight_ ## W ## _ ## OPT(uint8_t *dst, \
-    int stride, int height, int log2_denom, int weight, int offset);
+#define H264_WEIGHT(W, OPT)                                             \
+void ff_h264_weight_ ## W ## _ ## OPT(uint8_t *dst, int stride,         \
+                                      int height, int log2_denom,       \
+                                      int weight, int offset);
 
-#define H264_BIWEIGHT(W, OPT) \
-void ff_h264_biweight_ ## W ## _ ## OPT(uint8_t *dst, \
-    uint8_t *src, int stride, int height, int log2_denom, int weightd, \
-    int weights, int offset);
+#define H264_BIWEIGHT(W, OPT)                                           \
+void ff_h264_biweight_ ## W ## _ ## OPT(uint8_t *dst, uint8_t *src,     \
+                                        int stride, int height,         \
+                                        int log2_denom, int weightd,    \
+                                        int weights, int offset);
 
-#define H264_BIWEIGHT_MMX(W) \
-H264_WEIGHT  (W, mmx2) \
-H264_BIWEIGHT(W, mmx2)
+#define H264_BIWEIGHT_MMX(W)                    \
+    H264_WEIGHT(W, mmx2)                        \
+    H264_BIWEIGHT(W, mmx2)
 
-#define H264_BIWEIGHT_MMX_SSE(W) \
-H264_BIWEIGHT_MMX(W) \
-H264_WEIGHT      (W, sse2) \
-H264_BIWEIGHT    (W, sse2) \
-H264_BIWEIGHT    (W, ssse3)
+#define H264_BIWEIGHT_MMX_SSE(W)                \
+    H264_BIWEIGHT_MMX(W)                        \
+    H264_WEIGHT(W, sse2)                        \
+    H264_BIWEIGHT(W, sse2)                      \
+    H264_BIWEIGHT(W, ssse3)
 
 H264_BIWEIGHT_MMX_SSE(16)
-H264_BIWEIGHT_MMX_SSE( 8)
-H264_BIWEIGHT_MMX    ( 4)
-
-#define H264_WEIGHT_10(W, DEPTH, OPT) \
-void ff_h264_weight_ ## W ## _ ## DEPTH ## _ ## OPT(uint8_t *dst, \
-    int stride, int height, int log2_denom, int weight, int offset);
-
-#define H264_BIWEIGHT_10(W, DEPTH, OPT) \
-void ff_h264_biweight_ ## W ## _ ## DEPTH ## _ ## OPT \
-    (uint8_t *dst, uint8_t *src, int stride, int height, int log2_denom, \
-     int weightd, int weights, int offset);
-
-#define H264_BIWEIGHT_10_SSE(W, DEPTH) \
-H264_WEIGHT_10  (W, DEPTH, sse2) \
-H264_WEIGHT_10  (W, DEPTH, sse4) \
-H264_BIWEIGHT_10(W, DEPTH, sse2) \
-H264_BIWEIGHT_10(W, DEPTH, sse4)
+H264_BIWEIGHT_MMX_SSE(8)
+H264_BIWEIGHT_MMX(4)
+
+#define H264_WEIGHT_10(W, DEPTH, OPT)                                   \
+void ff_h264_weight_ ## W ## _ ## DEPTH ## _ ## OPT(uint8_t *dst,       \
+                                                    int stride,         \
+                                                    int height,         \
+                                                    int log2_denom,     \
+                                                    int weight,         \
+                                                    int offset);
+
+#define H264_BIWEIGHT_10(W, DEPTH, OPT)                                 \
+void ff_h264_biweight_ ## W ## _ ## DEPTH ## _ ## OPT(uint8_t *dst,     \
+                                                      uint8_t *src,     \
+                                                      int stride,       \
+                                                      int height,       \
+                                                      int log2_denom,   \
+                                                      int weightd,      \
+                                                      int weights,      \
+                                                      int offset);
+
+#define H264_BIWEIGHT_10_SSE(W, DEPTH)          \
+    H264_WEIGHT_10(W, DEPTH, sse2)              \
+    H264_WEIGHT_10(W, DEPTH, sse4)              \
+    H264_BIWEIGHT_10(W, DEPTH, sse2)            \
+    H264_BIWEIGHT_10(W, DEPTH, sse4)
 
 H264_BIWEIGHT_10_SSE(16, 10)
-H264_BIWEIGHT_10_SSE( 8, 10)
-H264_BIWEIGHT_10_SSE( 4, 10)
+H264_BIWEIGHT_10_SSE(8,  10)
+H264_BIWEIGHT_10_SSE(4,  10)
 
-void ff_h264dsp_init_x86(H264DSPContext *c, const int bit_depth, const int 
chroma_format_idc)
+void ff_h264dsp_init_x86(H264DSPContext *c, const int bit_depth,
+                         const int chroma_format_idc)
 {
     int mm_flags = av_get_cpu_flags();
 
-    if (chroma_format_idc == 1 && mm_flags & AV_CPU_FLAG_MMX2) {
-        c->h264_loop_filter_strength= h264_loop_filter_strength_mmx2;
-    }
+    if (chroma_format_idc == 1 && mm_flags & AV_CPU_FLAG_MMX2)
+        c->h264_loop_filter_strength = h264_loop_filter_strength_mmx2;
 
 #if HAVE_YASM
     if (bit_depth == 8) {
-    if (mm_flags & AV_CPU_FLAG_MMX) {
-        c->h264_idct_dc_add         =
-        c->h264_idct_add            = ff_h264_idct_add_8_mmx;
-        c->h264_idct8_dc_add        =
-        c->h264_idct8_add           = ff_h264_idct8_add_8_mmx;
-
-        c->h264_idct_add16          = ff_h264_idct_add16_8_mmx;
-        c->h264_idct8_add4          = ff_h264_idct8_add4_8_mmx;
-        if (chroma_format_idc == 1)
-            c->h264_idct_add8       = ff_h264_idct_add8_8_mmx;
-        c->h264_idct_add16intra     = ff_h264_idct_add16intra_8_mmx;
-        if (mm_flags & AV_CPU_FLAG_CMOV)
-            c->h264_luma_dc_dequant_idct = ff_h264_luma_dc_dequant_idct_mmx;
-
-        if (mm_flags & AV_CPU_FLAG_MMX2) {
-            c->h264_idct_dc_add    = ff_h264_idct_dc_add_8_mmx2;
-            c->h264_idct8_dc_add   = ff_h264_idct8_dc_add_8_mmx2;
-            c->h264_idct_add16     = ff_h264_idct_add16_8_mmx2;
-            c->h264_idct8_add4     = ff_h264_idct8_add4_8_mmx2;
+        if (mm_flags & AV_CPU_FLAG_MMX) {
+            c->h264_idct_dc_add   =
+            c->h264_idct_add      = ff_h264_idct_add_8_mmx;
+            c->h264_idct8_dc_add  =
+            c->h264_idct8_add     = ff_h264_idct8_add_8_mmx;
+
+            c->h264_idct_add16 = ff_h264_idct_add16_8_mmx;
+            c->h264_idct8_add4 = ff_h264_idct8_add4_8_mmx;
             if (chroma_format_idc == 1)
-                c->h264_idct_add8  = ff_h264_idct_add8_8_mmx2;
-            c->h264_idct_add16intra= ff_h264_idct_add16intra_8_mmx2;
-
-            c->h264_v_loop_filter_chroma= ff_deblock_v_chroma_8_mmx2;
-            c->h264_v_loop_filter_chroma_intra= 
ff_deblock_v_chroma_intra_8_mmx2;
-            if (chroma_format_idc == 1) {
-                c->h264_h_loop_filter_chroma= ff_deblock_h_chroma_8_mmx2;
-                c->h264_h_loop_filter_chroma_intra= 
ff_deblock_h_chroma_intra_8_mmx2;
-            }
+                c->h264_idct_add8 = ff_h264_idct_add8_8_mmx;
+            c->h264_idct_add16intra = ff_h264_idct_add16intra_8_mmx;
+            if (mm_flags & AV_CPU_FLAG_CMOV)
+                c->h264_luma_dc_dequant_idct = 
ff_h264_luma_dc_dequant_idct_mmx;
+
+            if (mm_flags & AV_CPU_FLAG_MMX2) {
+                c->h264_idct_dc_add  = ff_h264_idct_dc_add_8_mmx2;
+                c->h264_idct8_dc_add = ff_h264_idct8_dc_add_8_mmx2;
+                c->h264_idct_add16   = ff_h264_idct_add16_8_mmx2;
+                c->h264_idct8_add4   = ff_h264_idct8_add4_8_mmx2;
+                if (chroma_format_idc == 1)
+                    c->h264_idct_add8 = ff_h264_idct_add8_8_mmx2;
+                c->h264_idct_add16intra = ff_h264_idct_add16intra_8_mmx2;
+
+                c->h264_v_loop_filter_chroma       = 
ff_deblock_v_chroma_8_mmx2;
+                c->h264_v_loop_filter_chroma_intra = 
ff_deblock_v_chroma_intra_8_mmx2;
+                if (chroma_format_idc == 1) {
+                    c->h264_h_loop_filter_chroma       = 
ff_deblock_h_chroma_8_mmx2;
+                    c->h264_h_loop_filter_chroma_intra = 
ff_deblock_h_chroma_intra_8_mmx2;
+                }
 #if ARCH_X86_32
-            c->h264_v_loop_filter_luma= ff_deblock_v_luma_8_mmx2;
-            c->h264_h_loop_filter_luma= ff_deblock_h_luma_8_mmx2;
-            c->h264_v_loop_filter_luma_intra = ff_deblock_v_luma_intra_8_mmx2;
-            c->h264_h_loop_filter_luma_intra = ff_deblock_h_luma_intra_8_mmx2;
-#endif
-            c->weight_h264_pixels_tab[0]= ff_h264_weight_16_mmx2;
-            c->weight_h264_pixels_tab[1]= ff_h264_weight_8_mmx2;
-            c->weight_h264_pixels_tab[2]= ff_h264_weight_4_mmx2;
+                c->h264_v_loop_filter_luma       = ff_deblock_v_luma_8_mmx2;
+                c->h264_h_loop_filter_luma       = ff_deblock_h_luma_8_mmx2;
+                c->h264_v_loop_filter_luma_intra = 
ff_deblock_v_luma_intra_8_mmx2;
+                c->h264_h_loop_filter_luma_intra = 
ff_deblock_h_luma_intra_8_mmx2;
+#endif /* ARCH_X86_32 */
+                c->weight_h264_pixels_tab[0] = ff_h264_weight_16_mmx2;
+                c->weight_h264_pixels_tab[1] = ff_h264_weight_8_mmx2;
+                c->weight_h264_pixels_tab[2] = ff_h264_weight_4_mmx2;
 
-            c->biweight_h264_pixels_tab[0]= ff_h264_biweight_16_mmx2;
-            c->biweight_h264_pixels_tab[1]= ff_h264_biweight_8_mmx2;
-            c->biweight_h264_pixels_tab[2]= ff_h264_biweight_4_mmx2;
+                c->biweight_h264_pixels_tab[0] = ff_h264_biweight_16_mmx2;
+                c->biweight_h264_pixels_tab[1] = ff_h264_biweight_8_mmx2;
+                c->biweight_h264_pixels_tab[2] = ff_h264_biweight_4_mmx2;
 
-            if (mm_flags&AV_CPU_FLAG_SSE2) {
-                c->h264_idct8_add           = ff_h264_idct8_add_8_sse2;
+                if (mm_flags & AV_CPU_FLAG_SSE2) {
+                    c->h264_idct8_add  = ff_h264_idct8_add_8_sse2;
 
-                c->h264_idct_add16          = ff_h264_idct_add16_8_sse2;
-                c->h264_idct8_add4          = ff_h264_idct8_add4_8_sse2;
-                if (chroma_format_idc == 1)
-                    c->h264_idct_add8       = ff_h264_idct_add8_8_sse2;
-                c->h264_idct_add16intra     = ff_h264_idct_add16intra_8_sse2;
-                c->h264_luma_dc_dequant_idct= 
ff_h264_luma_dc_dequant_idct_sse2;
+                    c->h264_idct_add16 = ff_h264_idct_add16_8_sse2;
+                    c->h264_idct8_add4 = ff_h264_idct8_add4_8_sse2;
+                    if (chroma_format_idc == 1)
+                        c->h264_idct_add8 = ff_h264_idct_add8_8_sse2;
+                    c->h264_idct_add16intra      = 
ff_h264_idct_add16intra_8_sse2;
+                    c->h264_luma_dc_dequant_idct = 
ff_h264_luma_dc_dequant_idct_sse2;
 
-                c->weight_h264_pixels_tab[0]= ff_h264_weight_16_sse2;
-                c->weight_h264_pixels_tab[1]= ff_h264_weight_8_sse2;
+                    c->weight_h264_pixels_tab[0] = ff_h264_weight_16_sse2;
+                    c->weight_h264_pixels_tab[1] = ff_h264_weight_8_sse2;
 
-                c->biweight_h264_pixels_tab[0]= ff_h264_biweight_16_sse2;
-                c->biweight_h264_pixels_tab[1]= ff_h264_biweight_8_sse2;
+                    c->biweight_h264_pixels_tab[0] = ff_h264_biweight_16_sse2;
+                    c->biweight_h264_pixels_tab[1] = ff_h264_biweight_8_sse2;
 
 #if HAVE_ALIGNED_STACK
-                c->h264_v_loop_filter_luma = ff_deblock_v_luma_8_sse2;
-                c->h264_h_loop_filter_luma = ff_deblock_h_luma_8_sse2;
-                c->h264_v_loop_filter_luma_intra = 
ff_deblock_v_luma_intra_8_sse2;
-                c->h264_h_loop_filter_luma_intra = 
ff_deblock_h_luma_intra_8_sse2;
-#endif
-            }
-            if (mm_flags&AV_CPU_FLAG_SSSE3) {
-                c->biweight_h264_pixels_tab[0]= ff_h264_biweight_16_ssse3;
-                c->biweight_h264_pixels_tab[1]= ff_h264_biweight_8_ssse3;
-            }
-            if (mm_flags&AV_CPU_FLAG_AVX) {
+                    c->h264_v_loop_filter_luma       = 
ff_deblock_v_luma_8_sse2;
+                    c->h264_h_loop_filter_luma       = 
ff_deblock_h_luma_8_sse2;
+                    c->h264_v_loop_filter_luma_intra = 
ff_deblock_v_luma_intra_8_sse2;
+                    c->h264_h_loop_filter_luma_intra = 
ff_deblock_h_luma_intra_8_sse2;
+#endif /* HAVE_ALIGNED_STACK */
+                }
+                if (mm_flags & AV_CPU_FLAG_SSSE3) {
+                    c->biweight_h264_pixels_tab[0] = ff_h264_biweight_16_ssse3;
+                    c->biweight_h264_pixels_tab[1] = ff_h264_biweight_8_ssse3;
+                }
+                if (mm_flags & AV_CPU_FLAG_AVX) {
 #if HAVE_ALIGNED_STACK
-                c->h264_v_loop_filter_luma = ff_deblock_v_luma_8_avx;
-                c->h264_h_loop_filter_luma = ff_deblock_h_luma_8_avx;
-                c->h264_v_loop_filter_luma_intra = 
ff_deblock_v_luma_intra_8_avx;
-                c->h264_h_loop_filter_luma_intra = 
ff_deblock_h_luma_intra_8_avx;
-#endif
+                    c->h264_v_loop_filter_luma       = ff_deblock_v_luma_8_avx;
+                    c->h264_h_loop_filter_luma       = ff_deblock_h_luma_8_avx;
+                    c->h264_v_loop_filter_luma_intra = 
ff_deblock_v_luma_intra_8_avx;
+                    c->h264_h_loop_filter_luma_intra = 
ff_deblock_h_luma_intra_8_avx;
+#endif /* HAVE_ALIGNED_STACK */
+                }
             }
         }
-    }
     } else if (bit_depth == 10) {
-    if (mm_flags & AV_CPU_FLAG_MMX) {
-        if (mm_flags & AV_CPU_FLAG_MMX2) {
+        if (mm_flags & AV_CPU_FLAG_MMX) {
+            if (mm_flags & AV_CPU_FLAG_MMX2) {
 #if ARCH_X86_32
-            c->h264_v_loop_filter_chroma= ff_deblock_v_chroma_10_mmx2;
-            c->h264_v_loop_filter_chroma_intra= 
ff_deblock_v_chroma_intra_10_mmx2;
-            c->h264_v_loop_filter_luma= ff_deblock_v_luma_10_mmx2;
-            c->h264_h_loop_filter_luma= ff_deblock_h_luma_10_mmx2;
-            c->h264_v_loop_filter_luma_intra = ff_deblock_v_luma_intra_10_mmx2;
-            c->h264_h_loop_filter_luma_intra = ff_deblock_h_luma_intra_10_mmx2;
-#endif
-            c->h264_idct_dc_add= ff_h264_idct_dc_add_10_mmx2;
-            if (mm_flags&AV_CPU_FLAG_SSE2) {
-                c->h264_idct_add       = ff_h264_idct_add_10_sse2;
-                c->h264_idct8_dc_add   = ff_h264_idct8_dc_add_10_sse2;
-
-                c->h264_idct_add16     = ff_h264_idct_add16_10_sse2;
-                if (chroma_format_idc == 1)
-                    c->h264_idct_add8  = ff_h264_idct_add8_10_sse2;
-                c->h264_idct_add16intra= ff_h264_idct_add16intra_10_sse2;
+                c->h264_v_loop_filter_chroma       = 
ff_deblock_v_chroma_10_mmx2;
+                c->h264_v_loop_filter_chroma_intra = 
ff_deblock_v_chroma_intra_10_mmx2;
+                c->h264_v_loop_filter_luma         = ff_deblock_v_luma_10_mmx2;
+                c->h264_h_loop_filter_luma         = ff_deblock_h_luma_10_mmx2;
+                c->h264_v_loop_filter_luma_intra   = 
ff_deblock_v_luma_intra_10_mmx2;
+                c->h264_h_loop_filter_luma_intra   = 
ff_deblock_h_luma_intra_10_mmx2;
+#endif /* ARCH_X86_32 */
+                c->h264_idct_dc_add = ff_h264_idct_dc_add_10_mmx2;
+                if (mm_flags & AV_CPU_FLAG_SSE2) {
+                    c->h264_idct_add     = ff_h264_idct_add_10_sse2;
+                    c->h264_idct8_dc_add = ff_h264_idct8_dc_add_10_sse2;
+
+                    c->h264_idct_add16 = ff_h264_idct_add16_10_sse2;
+                    if (chroma_format_idc == 1)
+                        c->h264_idct_add8 = ff_h264_idct_add8_10_sse2;
+                    c->h264_idct_add16intra = ff_h264_idct_add16intra_10_sse2;
 #if HAVE_ALIGNED_STACK
-                c->h264_idct8_add      = ff_h264_idct8_add_10_sse2;
-                c->h264_idct8_add4     = ff_h264_idct8_add4_10_sse2;
-#endif
+                    c->h264_idct8_add  = ff_h264_idct8_add_10_sse2;
+                    c->h264_idct8_add4 = ff_h264_idct8_add4_10_sse2;
+#endif /* HAVE_ALIGNED_STACK */
 
-                c->weight_h264_pixels_tab[0] = ff_h264_weight_16_10_sse2;
-                c->weight_h264_pixels_tab[1] = ff_h264_weight_8_10_sse2;
-                c->weight_h264_pixels_tab[2] = ff_h264_weight_4_10_sse2;
+                    c->weight_h264_pixels_tab[0] = ff_h264_weight_16_10_sse2;
+                    c->weight_h264_pixels_tab[1] = ff_h264_weight_8_10_sse2;
+                    c->weight_h264_pixels_tab[2] = ff_h264_weight_4_10_sse2;
 
-                c->biweight_h264_pixels_tab[0] = ff_h264_biweight_16_10_sse2;
-                c->biweight_h264_pixels_tab[1] = ff_h264_biweight_8_10_sse2;
-                c->biweight_h264_pixels_tab[2] = ff_h264_biweight_4_10_sse2;
+                    c->biweight_h264_pixels_tab[0] = 
ff_h264_biweight_16_10_sse2;
+                    c->biweight_h264_pixels_tab[1] = 
ff_h264_biweight_8_10_sse2;
+                    c->biweight_h264_pixels_tab[2] = 
ff_h264_biweight_4_10_sse2;
 
-                c->h264_v_loop_filter_chroma= ff_deblock_v_chroma_10_sse2;
-                c->h264_v_loop_filter_chroma_intra= 
ff_deblock_v_chroma_intra_10_sse2;
+                    c->h264_v_loop_filter_chroma       = 
ff_deblock_v_chroma_10_sse2;
+                    c->h264_v_loop_filter_chroma_intra = 
ff_deblock_v_chroma_intra_10_sse2;
 #if HAVE_ALIGNED_STACK
-                c->h264_v_loop_filter_luma = ff_deblock_v_luma_10_sse2;
-                c->h264_h_loop_filter_luma = ff_deblock_h_luma_10_sse2;
-                c->h264_v_loop_filter_luma_intra = 
ff_deblock_v_luma_intra_10_sse2;
-                c->h264_h_loop_filter_luma_intra = 
ff_deblock_h_luma_intra_10_sse2;
-#endif
-            }
-            if (mm_flags&AV_CPU_FLAG_SSE4) {
-                c->weight_h264_pixels_tab[0] = ff_h264_weight_16_10_sse4;
-                c->weight_h264_pixels_tab[1] = ff_h264_weight_8_10_sse4;
-                c->weight_h264_pixels_tab[2] = ff_h264_weight_4_10_sse4;
-
-                c->biweight_h264_pixels_tab[0] = ff_h264_biweight_16_10_sse4;
-                c->biweight_h264_pixels_tab[1] = ff_h264_biweight_8_10_sse4;
-                c->biweight_h264_pixels_tab[2] = ff_h264_biweight_4_10_sse4;
-            }
+                    c->h264_v_loop_filter_luma       = 
ff_deblock_v_luma_10_sse2;
+                    c->h264_h_loop_filter_luma       = 
ff_deblock_h_luma_10_sse2;
+                    c->h264_v_loop_filter_luma_intra = 
ff_deblock_v_luma_intra_10_sse2;
+                    c->h264_h_loop_filter_luma_intra = 
ff_deblock_h_luma_intra_10_sse2;
+#endif /* HAVE_ALIGNED_STACK */
+                }
+                if (mm_flags & AV_CPU_FLAG_SSE4) {
+                    c->weight_h264_pixels_tab[0] = ff_h264_weight_16_10_sse4;
+                    c->weight_h264_pixels_tab[1] = ff_h264_weight_8_10_sse4;
+                    c->weight_h264_pixels_tab[2] = ff_h264_weight_4_10_sse4;
+
+                    c->biweight_h264_pixels_tab[0] = 
ff_h264_biweight_16_10_sse4;
+                    c->biweight_h264_pixels_tab[1] = 
ff_h264_biweight_8_10_sse4;
+                    c->biweight_h264_pixels_tab[2] = 
ff_h264_biweight_4_10_sse4;
+                }
 #if HAVE_AVX
-            if (mm_flags&AV_CPU_FLAG_AVX) {
-                c->h264_idct_dc_add    =
-                c->h264_idct_add       = ff_h264_idct_add_10_avx;
-                c->h264_idct8_dc_add   = ff_h264_idct8_dc_add_10_avx;
-
-                c->h264_idct_add16     = ff_h264_idct_add16_10_avx;
-                if (chroma_format_idc == 1)
-                    c->h264_idct_add8  = ff_h264_idct_add8_10_avx;
-                c->h264_idct_add16intra= ff_h264_idct_add16intra_10_avx;
+                if (mm_flags & AV_CPU_FLAG_AVX) {
+                    c->h264_idct_dc_add  =
+                    c->h264_idct_add     = ff_h264_idct_add_10_avx;
+                    c->h264_idct8_dc_add = ff_h264_idct8_dc_add_10_avx;
+
+                    c->h264_idct_add16 = ff_h264_idct_add16_10_avx;
+                    if (chroma_format_idc == 1)
+                        c->h264_idct_add8 = ff_h264_idct_add8_10_avx;
+                    c->h264_idct_add16intra = ff_h264_idct_add16intra_10_avx;
 #if HAVE_ALIGNED_STACK
-                c->h264_idct8_add      = ff_h264_idct8_add_10_avx;
-                c->h264_idct8_add4     = ff_h264_idct8_add4_10_avx;
-#endif
+                    c->h264_idct8_add  = ff_h264_idct8_add_10_avx;
+                    c->h264_idct8_add4 = ff_h264_idct8_add4_10_avx;
+#endif /* HAVE_ALIGNED_STACK */
 
-                c->h264_v_loop_filter_chroma= ff_deblock_v_chroma_10_avx;
-                c->h264_v_loop_filter_chroma_intra= 
ff_deblock_v_chroma_intra_10_avx;
+                    c->h264_v_loop_filter_chroma       = 
ff_deblock_v_chroma_10_avx;
+                    c->h264_v_loop_filter_chroma_intra = 
ff_deblock_v_chroma_intra_10_avx;
 #if HAVE_ALIGNED_STACK
-                c->h264_v_loop_filter_luma = ff_deblock_v_luma_10_avx;
-                c->h264_h_loop_filter_luma = ff_deblock_h_luma_10_avx;
-                c->h264_v_loop_filter_luma_intra = 
ff_deblock_v_luma_intra_10_avx;
-                c->h264_h_loop_filter_luma_intra = 
ff_deblock_h_luma_intra_10_avx;
-#endif
-            }
+                    c->h264_v_loop_filter_luma         = 
ff_deblock_v_luma_10_avx;
+                    c->h264_h_loop_filter_luma         = 
ff_deblock_h_luma_10_avx;
+                    c->h264_v_loop_filter_luma_intra   = 
ff_deblock_v_luma_intra_10_avx;
+                    c->h264_h_loop_filter_luma_intra   = 
ff_deblock_h_luma_intra_10_avx;
+#endif /* HAVE_ALIGNED_STACK */
+                }
 #endif /* HAVE_AVX */
+            }
         }
     }
-    }
-#endif
+#endif /* HAVE_YASM */
 }
-- 
1.7.1

_______________________________________________
libav-devel mailing list
[email protected]
https://lists.libav.org/mailman/listinfo/libav-devel

Reply via email to