struct
{
+ /* space for p_fenc and p_fdec */
+#define FENC_STRIDE 16
+#define FDEC_STRIDE 32
+ DECLARE_ALIGNED( uint8_t, fenc_buf[24*FENC_STRIDE], 16 );
+ DECLARE_ALIGNED( uint8_t, fdec_buf[27*FDEC_STRIDE], 16 );
+
/* pointer over mb of the frame to be compressed */
uint8_t *p_fenc[3];
uint8_t *p_fref[2][16][4+2]; /* last: lN, lH, lV, lHV, cU, cV */
uint16_t *p_integral[2][16];
- /* common stride */
+ /* fref stride */
int i_stride[3];
} pic;
const int mvy = x264_clip3( h->mb.cache.mv[0][i8][1], h->mb.mv_min[1], h->mb.mv_max[1] );
h->mc.mc_luma( h->mb.pic.p_fref[0][i_ref], h->mb.pic.i_stride[0],
- &h->mb.pic.p_fdec[0][4*y * h->mb.pic.i_stride[0]+4*x], h->mb.pic.i_stride[0],
+ &h->mb.pic.p_fdec[0][4*y*FDEC_STRIDE+4*x], FDEC_STRIDE,
mvx + 4*4*x, mvy + 4*4*y, 4*width, 4*height );
h->mc.mc_chroma( &h->mb.pic.p_fref[0][i_ref][4][2*y*h->mb.pic.i_stride[1]+2*x], h->mb.pic.i_stride[1],
- &h->mb.pic.p_fdec[1][2*y*h->mb.pic.i_stride[1]+2*x], h->mb.pic.i_stride[1],
+ &h->mb.pic.p_fdec[1][2*y*FDEC_STRIDE+2*x], FDEC_STRIDE,
mvx, mvy, 2*width, 2*height );
h->mc.mc_chroma( &h->mb.pic.p_fref[0][i_ref][5][2*y*h->mb.pic.i_stride[2]+2*x], h->mb.pic.i_stride[2],
- &h->mb.pic.p_fdec[2][2*y*h->mb.pic.i_stride[2]+2*x], h->mb.pic.i_stride[2],
+ &h->mb.pic.p_fdec[2][2*y*FDEC_STRIDE+2*x], FDEC_STRIDE,
mvx, mvy, 2*width, 2*height );
}
static inline void x264_mb_mc_1xywh( x264_t *h, int x, int y, int width, int height )
const int mvy = x264_clip3( h->mb.cache.mv[1][i8][1], h->mb.mv_min[1], h->mb.mv_max[1] );
h->mc.mc_luma( h->mb.pic.p_fref[1][i_ref], h->mb.pic.i_stride[0],
- &h->mb.pic.p_fdec[0][4*y *h->mb.pic.i_stride[0]+4*x], h->mb.pic.i_stride[0],
+ &h->mb.pic.p_fdec[0][4*y*FDEC_STRIDE+4*x], FDEC_STRIDE,
mvx + 4*4*x, mvy + 4*4*y, 4*width, 4*height );
h->mc.mc_chroma( &h->mb.pic.p_fref[1][i_ref][4][2*y*h->mb.pic.i_stride[1]+2*x], h->mb.pic.i_stride[1],
- &h->mb.pic.p_fdec[1][2*y*h->mb.pic.i_stride[1]+2*x], h->mb.pic.i_stride[1],
+ &h->mb.pic.p_fdec[1][2*y*FDEC_STRIDE+2*x], FDEC_STRIDE,
mvx, mvy, 2*width, 2*height );
h->mc.mc_chroma( &h->mb.pic.p_fref[1][i_ref][5][2*y*h->mb.pic.i_stride[2]+2*x], h->mb.pic.i_stride[2],
- &h->mb.pic.p_fdec[2][2*y*h->mb.pic.i_stride[2]+2*x], h->mb.pic.i_stride[2],
+ &h->mb.pic.p_fdec[2][2*y*FDEC_STRIDE+2*x], FDEC_STRIDE,
mvx, mvy, 2*width, 2*height );
}
const int i_ref0 = h->mb.cache.ref[0][i8];
const int weight = h->mb.bipred_weight[i_ref0][i_ref1];
- h->mc.avg_weight[i_mode]( &h->mb.pic.p_fdec[0][4*y *h->mb.pic.i_stride[0]+4*x], h->mb.pic.i_stride[0], tmp, 16, weight );
+ h->mc.avg_weight[i_mode]( &h->mb.pic.p_fdec[0][4*y*FDEC_STRIDE+4*x], FDEC_STRIDE, tmp, 16, weight );
h->mc.mc_chroma( &h->mb.pic.p_fref[1][i_ref1][4][2*y*h->mb.pic.i_stride[1]+2*x], h->mb.pic.i_stride[1],
tmp, 16, mvx1, mvy1, 2*width, 2*height );
- h->mc.avg_weight[i_mode+3]( &h->mb.pic.p_fdec[1][2*y*h->mb.pic.i_stride[1]+2*x], h->mb.pic.i_stride[1], tmp, 16, weight );
+ h->mc.avg_weight[i_mode+3]( &h->mb.pic.p_fdec[1][2*y*FDEC_STRIDE+2*x], FDEC_STRIDE, tmp, 16, weight );
h->mc.mc_chroma( &h->mb.pic.p_fref[1][i_ref1][5][2*y*h->mb.pic.i_stride[2]+2*x], h->mb.pic.i_stride[2],
tmp, 16, mvx1, mvy1, 2*width, 2*height );
- h->mc.avg_weight[i_mode+3]( &h->mb.pic.p_fdec[2][2*y*h->mb.pic.i_stride[2]+2*x], h->mb.pic.i_stride[2], tmp, 16, weight );
+ h->mc.avg_weight[i_mode+3]( &h->mb.pic.p_fdec[2][2*y*FDEC_STRIDE+2*x], FDEC_STRIDE, tmp, 16, weight );
}
else
{
- h->mc.avg[i_mode]( &h->mb.pic.p_fdec[0][4*y *h->mb.pic.i_stride[0]+4*x], h->mb.pic.i_stride[0], tmp, 16 );
+ h->mc.avg[i_mode]( &h->mb.pic.p_fdec[0][4*y*FDEC_STRIDE+4*x], FDEC_STRIDE, tmp, 16 );
h->mc.mc_chroma( &h->mb.pic.p_fref[1][i_ref1][4][2*y*h->mb.pic.i_stride[1]+2*x], h->mb.pic.i_stride[1],
tmp, 16, mvx1, mvy1, 2*width, 2*height );
- h->mc.avg[i_mode+3]( &h->mb.pic.p_fdec[1][2*y*h->mb.pic.i_stride[1]+2*x], h->mb.pic.i_stride[1], tmp, 16 );
+ h->mc.avg[i_mode+3]( &h->mb.pic.p_fdec[1][2*y*FDEC_STRIDE+2*x], FDEC_STRIDE, tmp, 16 );
h->mc.mc_chroma( &h->mb.pic.p_fref[1][i_ref1][5][2*y*h->mb.pic.i_stride[2]+2*x], h->mb.pic.i_stride[2],
tmp, 16, mvx1, mvy1, 2*width, 2*height );
- h->mc.avg[i_mode+3]( &h->mb.pic.p_fdec[2][2*y*h->mb.pic.i_stride[2]+2*x], h->mb.pic.i_stride[2], tmp, 16 );
+ h->mc.avg[i_mode+3]( &h->mb.pic.p_fdec[2][2*y*FDEC_STRIDE+2*x], FDEC_STRIDE, tmp, 16 );
}
}
h->mb.i_b4_xy = i_mb_4x4;
h->mb.i_neighbour = 0;
+ /* fdec: fenc:
+ * yyyyyyy
+ * yYYYY YYYY
+ * yYYYY YYYY
+ * yYYYY YYYY
+ * yYYYY YYYY
+ * uuu vvv UUVV
+ * uUU vVV UUVV
+ * uUU vVV
+ */
+ h->mb.pic.p_fenc[0] = h->mb.pic.fenc_buf;
+ h->mb.pic.p_fenc[1] = h->mb.pic.fenc_buf + 16*FENC_STRIDE;
+ h->mb.pic.p_fenc[2] = h->mb.pic.fenc_buf + 16*FENC_STRIDE + 8;
+ h->mb.pic.p_fdec[0] = h->mb.pic.fdec_buf + 2*FDEC_STRIDE;
+ h->mb.pic.p_fdec[1] = h->mb.pic.fdec_buf + 19*FDEC_STRIDE;
+ h->mb.pic.p_fdec[2] = h->mb.pic.fdec_buf + 19*FDEC_STRIDE + 16;
+
/* load picture pointers */
for( i = 0; i < 3; i++ )
{
const int w = (i == 0 ? 16 : 8);
const int i_stride = h->fdec->i_stride[i];
+ const uint8_t *plane_fdec = &h->fdec->plane[i][ w * (i_mb_x + i_mb_y * i_stride) ];
int j;
h->mb.pic.i_stride[i] = i_stride;
- h->mb.pic.p_fenc[i] = &h->fenc->plane[i][ w * ( i_mb_x + i_mb_y * i_stride )];
- h->mb.pic.p_fdec[i] = &h->fdec->plane[i][ w * ( i_mb_x + i_mb_y * i_stride )];
+ h->mc.copy[i?PIXEL_8x8:PIXEL_16x16]( h->mb.pic.p_fenc[i], FENC_STRIDE,
+ &h->fenc->plane[i][ w * (i_mb_x + i_mb_y * i_stride) ], i_stride, w );
+ memcpy( &h->mb.pic.p_fdec[i][-1-FDEC_STRIDE], &plane_fdec[-1-i_stride], w*3/2+1 );
+ for( j = 0; j < w; j++ )
+ h->mb.pic.p_fdec[i][-1+j*FDEC_STRIDE] = plane_fdec[-1+j*i_stride];
for( j = 0; j < h->i_ref0; j++ )
{
int i;
+ for( i = 0; i < 3; i++ )
+ {
+ int w = i ? 8 : 16;
+ h->mc.copy[i?PIXEL_8x8:PIXEL_16x16](
+ &h->fdec->plane[i][ w * (h->mb.i_mb_x + h->mb.i_mb_y * h->fdec->i_stride[i]) ],
+ h->fdec->i_stride[i],
+ h->mb.pic.p_fdec[i], FDEC_STRIDE, w );
+ }
+
h->mb.type[i_mb_xy] = i_mb_type;
if( h->mb.i_type != I_16x16 && h->mb.i_cbp_luma == 0 && h->mb.i_cbp_chroma == 0 )
int predict_mode[9];
uint8_t *p_dstc[2], *p_srcc[2];
- int i_stride[2];
if( a->i_sad_i8x8chroma < COST_MAX )
return;
p_srcc[0] = h->mb.pic.p_fenc[1];
p_srcc[1] = h->mb.pic.p_fenc[2];
- i_stride[0] = h->mb.pic.i_stride[1];
- i_stride[1] = h->mb.pic.i_stride[2];
-
predict_8x8chroma_mode_available( h->mb.i_neighbour, predict_mode, &i_max );
a->i_sad_i8x8chroma = COST_MAX;
for( i = 0; i < i_max; i++ )
i_mode = predict_mode[i];
/* we do the prediction */
- h->predict_8x8c[i_mode]( p_dstc[0], i_stride[0] );
- h->predict_8x8c[i_mode]( p_dstc[1], i_stride[1] );
+ h->predict_8x8c[i_mode]( p_dstc[0], FDEC_STRIDE );
+ h->predict_8x8c[i_mode]( p_dstc[1], FDEC_STRIDE );
/* we calculate the cost */
- i_sad = h->pixf.mbcmp[PIXEL_8x8]( p_dstc[0], i_stride[0],
- p_srcc[0], i_stride[0] ) +
- h->pixf.mbcmp[PIXEL_8x8]( p_dstc[1], i_stride[1],
- p_srcc[1], i_stride[1] ) +
+ i_sad = h->pixf.mbcmp[PIXEL_8x8]( p_dstc[0], FDEC_STRIDE,
+ p_srcc[0], FENC_STRIDE ) +
+ h->pixf.mbcmp[PIXEL_8x8]( p_dstc[1], FDEC_STRIDE,
+ p_srcc[1], FENC_STRIDE ) +
a->i_lambda * bs_size_ue( x264_mb_pred_mode8x8c_fix[i_mode] );
/* if i_score is lower it is better */
static void x264_mb_analyse_intra( x264_t *h, x264_mb_analysis_t *a, int i_cost_inter )
{
const unsigned int flags = h->sh.i_type == SLICE_TYPE_I ? h->param.analyse.intra : h->param.analyse.inter;
- const int i_stride = h->mb.pic.i_stride[0];
uint8_t *p_src = h->mb.pic.p_fenc[0];
uint8_t *p_dst = h->mb.pic.p_fdec[0];
int f8_satd_rd_ratio = 0;
int i_mode;
i_mode = predict_mode[i];
- h->predict_16x16[i_mode]( p_dst, i_stride );
+ h->predict_16x16[i_mode]( p_dst, FDEC_STRIDE );
- i_sad = h->pixf.mbcmp[PIXEL_16x16]( p_dst, i_stride, p_src, i_stride ) +
+ i_sad = h->pixf.mbcmp[PIXEL_16x16]( p_dst, FDEC_STRIDE, p_src, FENC_STRIDE ) +
a->i_lambda * bs_size_ue( x264_mb_pred_mode16x16_fix[i_mode] );
if( a->i_sad_i16x16 > i_sad )
{
x = block_idx_x[idx];
y = block_idx_y[idx];
- p_src_by = p_src + 4 * x + 4 * y * i_stride;
- p_dst_by = p_dst + 4 * x + 4 * y * i_stride;
+ p_src_by = p_src + 4 * x + 4 * y * FENC_STRIDE;
+ p_dst_by = p_dst + 4 * x + 4 * y * FDEC_STRIDE;
i_best = COST_MAX;
predict_4x4_mode_available( h->mb.i_neighbour4[idx], predict_mode, &i_max );
if( (h->mb.i_neighbour4[idx] & (MB_TOPRIGHT|MB_TOP)) == MB_TOP )
/* emulate missing topright samples */
- *(uint32_t*) &p_dst_by[4 - i_stride] = p_dst_by[3 - i_stride] * 0x01010101U;
+ *(uint32_t*) &p_dst_by[4 - FDEC_STRIDE] = p_dst_by[3 - FDEC_STRIDE] * 0x01010101U;
for( i = 0; i < i_max; i++ )
{
int i_mode;
i_mode = predict_mode[i];
- h->predict_4x4[i_mode]( p_dst_by, i_stride );
+ h->predict_4x4[i_mode]( p_dst_by, FDEC_STRIDE );
- i_sad = h->pixf.mbcmp[PIXEL_4x4]( p_dst_by, i_stride,
- p_src_by, i_stride )
+ i_sad = h->pixf.mbcmp[PIXEL_4x4]( p_dst_by, FDEC_STRIDE,
+ p_src_by, FENC_STRIDE )
+ a->i_lambda * (i_pred_mode == x264_mb_pred_mode4x4_fix(i_mode) ? 1 : 4);
if( i_best > i_sad )
a->i_sad_i4x4 += i_best;
/* we need to encode this block now (for next ones) */
- h->predict_4x4[a->i_predict4x4[x][y]]( p_dst_by, i_stride );
+ h->predict_4x4[a->i_predict4x4[x][y]]( p_dst_by, FDEC_STRIDE );
x264_mb_encode_i4x4( h, idx, a->i_qp );
h->mb.cache.intra4x4_pred_mode[x264_scan8[idx]] = a->i_predict4x4[x][y];
x = idx&1;
y = idx>>1;
- p_src_by = p_src + 8 * x + 8 * y * i_stride;
- p_dst_by = p_dst + 8 * x + 8 * y * i_stride;
+ p_src_by = p_src + 8 * x + 8 * y * FENC_STRIDE;
+ p_dst_by = p_dst + 8 * x + 8 * y * FDEC_STRIDE;
i_best = COST_MAX;
predict_4x4_mode_available( h->mb.i_neighbour8[idx], predict_mode, &i_max );
int i_mode;
i_mode = predict_mode[i];
- h->predict_8x8[i_mode]( p_dst_by, i_stride, h->mb.i_neighbour8[idx] );
+ h->predict_8x8[i_mode]( p_dst_by, FDEC_STRIDE, h->mb.i_neighbour8[idx] );
/* could use sa8d, but it doesn't seem worth the speed cost (without mmx at least) */
- i_sad = h->pixf.mbcmp[PIXEL_8x8]( p_dst_by, i_stride,
- p_src_by, i_stride )
+ i_sad = h->pixf.mbcmp[PIXEL_8x8]( p_dst_by, FDEC_STRIDE,
+ p_src_by, FENC_STRIDE )
+ a->i_lambda * (i_pred_mode == x264_mb_pred_mode4x4_fix(i_mode) ? 1 : 4);
if( i_best > i_sad )
a->i_sad_i8x8 += i_best;
/* we need to encode this block now (for next ones) */
- h->predict_8x8[a->i_predict8x8[x][y]]( p_dst_by, i_stride, h->mb.i_neighbour );
+ h->predict_8x8[a->i_predict8x8[x][y]]( p_dst_by, FDEC_STRIDE, h->mb.i_neighbour );
x264_mb_encode_i8x8( h, idx, a->i_qp );
x264_macroblock_cache_intra8x8_pred( h, 2*x, 2*y, a->i_predict8x8[x][y] );
#define LOAD_FENC( m, src, xoff, yoff) \
(m)->i_stride[0] = h->mb.pic.i_stride[0]; \
(m)->i_stride[1] = h->mb.pic.i_stride[1]; \
- (m)->p_fenc[0] = &(src)[0][(xoff)+(yoff)*(m)->i_stride[0]]; \
- (m)->p_fenc[1] = &(src)[1][((xoff)>>1)+((yoff)>>1)*(m)->i_stride[1]]; \
- (m)->p_fenc[2] = &(src)[2][((xoff)>>1)+((yoff)>>1)*(m)->i_stride[1]];
+ (m)->p_fenc[0] = &(src)[0][(xoff)+(yoff)*FENC_STRIDE]; \
+ (m)->p_fenc[1] = &(src)[1][((xoff)>>1)+((yoff)>>1)*FENC_STRIDE]; \
+ (m)->p_fenc[2] = &(src)[2][((xoff)>>1)+((yoff)>>1)*FENC_STRIDE];
#define LOAD_HPELS(m, src, list, ref, xoff, yoff) \
(m)->p_fref[0] = &(src)[0][(xoff)+(yoff)*(m)->i_stride[0]]; \
static int x264_mb_analyse_inter_p4x4_chroma( x264_t *h, x264_mb_analysis_t *a, uint8_t **p_fref, int i8x8, int pixel )
{
- uint8_t pix1[8*8], pix2[8*8];
+ DECLARE_ALIGNED( uint8_t, pix1[8*8], 8 );
+ DECLARE_ALIGNED( uint8_t, pix2[8*8], 8 );
const int i_stride = h->mb.pic.i_stride[1];
const int off = 4*(i8x8&1) + 2*(i8x8&2)*i_stride;
CHROMA4x4MC( 2,4, a->l0.me4x8[i8x8][1], 2,0 );
}
- return h->pixf.mbcmp[PIXEL_4x4]( &h->mb.pic.p_fenc[1][off], i_stride, pix1, 8 )
- + h->pixf.mbcmp[PIXEL_4x4]( &h->mb.pic.p_fenc[2][off], i_stride, pix2, 8 );
+ return h->pixf.mbcmp[PIXEL_4x4]( &h->mb.pic.p_fenc[1][off], FENC_STRIDE, pix1, 8 )
+ + h->pixf.mbcmp[PIXEL_4x4]( &h->mb.pic.p_fenc[2][off], FENC_STRIDE, pix2, 8 );
}
static void x264_mb_analyse_inter_p4x4( x264_t *h, x264_mb_analysis_t *a, int i8x8 )
uint8_t **p_fenc = h->mb.pic.p_fenc;
uint8_t **p_fdec = h->mb.pic.p_fdec;
- int i_stride= h->mb.pic.i_stride[0];
int i;
a->i_cost16x16direct = 0;
for( i = 0; i < 4; i++ )
{
- const int x8 = i%2;
- const int y8 = i/2;
- const int off = 8 * x8 + 8 * i_stride * y8;
+ const int x = (i&1)*8;
+ const int y = (i>>1)*8;
a->i_cost16x16direct +=
a->i_cost8x8direct[i] =
- h->pixf.mbcmp[PIXEL_8x8]( &p_fenc[0][off], i_stride, &p_fdec[0][off], i_stride );
+ h->pixf.mbcmp[PIXEL_8x8]( &p_fenc[0][x+y*FENC_STRIDE], FENC_STRIDE, &p_fdec[0][x+y*FDEC_STRIDE], FDEC_STRIDE );
/* mb type cost */
a->i_cost8x8direct[i] += a->i_lambda * i_sub_mb_b_cost_table[D_DIRECT_8x8];
else
h->mc.avg[PIXEL_16x16]( pix1, 16, src2, stride2 );
- a->i_cost16x16bi = h->pixf.mbcmp[PIXEL_16x16]( h->mb.pic.p_fenc[0], h->mb.pic.i_stride[0], pix1, 16 )
+ a->i_cost16x16bi = h->pixf.mbcmp[PIXEL_16x16]( h->mb.pic.p_fenc[0], FENC_STRIDE, pix1, 16 )
+ REF_COST( 0, a->l0.i_ref )
+ REF_COST( 1, a->l1.i_ref )
+ a->l0.me16x16.cost_mv
}
WEIGHTED_AVG( PIXEL_8x8, pix[0], 8, pix[1], 8 );
- i_part_cost_bi += h->pixf.mbcmp[PIXEL_8x8]( a->l0.me8x8[i].p_fenc[0], h->mb.pic.i_stride[0], pix[0], 8 )
+ i_part_cost_bi += h->pixf.mbcmp[PIXEL_8x8]( a->l0.me8x8[i].p_fenc[0], FENC_STRIDE, pix[0], 8 )
+ a->i_lambda * i_sub_mb_b_cost_table[D_BI_8x8];
a->l0.me8x8[i].cost += a->i_lambda * i_sub_mb_b_cost_table[D_L0_8x8];
a->l1.me8x8[i].cost += a->i_lambda * i_sub_mb_b_cost_table[D_L1_8x8];
uint8_t **p_fref[2] =
{ h->mb.pic.p_fref[0][a->l0.i_ref],
h->mb.pic.p_fref[1][a->l1.i_ref] };
- uint8_t pix[2][16*8];
+ DECLARE_ALIGNED( uint8_t, pix[2][16*8], 16 );
int mvc[2][2];
int i, l;
}
WEIGHTED_AVG( PIXEL_16x8, pix[0], 16, pix[1], 16 );
- i_part_cost_bi += h->pixf.mbcmp[PIXEL_16x8]( a->l0.me16x8[i].p_fenc[0], h->mb.pic.i_stride[0], pix[0], 16 );
+ i_part_cost_bi += h->pixf.mbcmp[PIXEL_16x8]( a->l0.me16x8[i].p_fenc[0], FENC_STRIDE, pix[0], 16 );
i_part_cost = a->l0.me16x8[i].cost;
a->i_mb_partition16x8[i] = D_L0_8x8; /* not actually 8x8, only the L0 matters */
}
WEIGHTED_AVG( PIXEL_8x16, pix[0], 8, pix[1], 8 );
- i_part_cost_bi += h->pixf.mbcmp[PIXEL_8x16]( a->l0.me8x16[i].p_fenc[0], h->mb.pic.i_stride[0], pix[0], 8 );
+ i_part_cost_bi += h->pixf.mbcmp[PIXEL_8x16]( a->l0.me8x16[i].p_fenc[0], FENC_STRIDE, pix[0], 8 );
i_part_cost = a->l0.me8x16[i].cost;
a->i_mb_partition8x16[i] = D_L0_8x8;
/* FIXME only luma mc is needed */
x264_mb_mc( h );
- i_cost8 = h->pixf.sa8d[PIXEL_16x16]( h->mb.pic.p_fenc[0], h->mb.pic.i_stride[0],
- h->mb.pic.p_fdec[0], h->mb.pic.i_stride[0] );
- i_cost4 = h->pixf.satd[PIXEL_16x16]( h->mb.pic.p_fenc[0], h->mb.pic.i_stride[0],
- h->mb.pic.p_fdec[0], h->mb.pic.i_stride[0] );
+ i_cost8 = h->pixf.sa8d[PIXEL_16x16]( h->mb.pic.p_fenc[0], FENC_STRIDE,
+ h->mb.pic.p_fdec[0], FDEC_STRIDE );
+ i_cost4 = h->pixf.satd[PIXEL_16x16]( h->mb.pic.p_fenc[0], FENC_STRIDE,
+ h->mb.pic.p_fdec[0], FDEC_STRIDE );
h->mb.b_transform_8x8 = i_cost8 < i_cost4;
}
#undef ZIG
#define ZIG(i,y,x) {\
- int o = x+y*i_stride;\
- level[i] = p_src[o] - p_dst[o];\
- p_dst[o] = p_src[o];\
+ int oe = x+y*FENC_STRIDE;\
+ int od = x+y*FDEC_STRIDE;\
+ level[i] = p_src[oe] - p_dst[od];\
+ p_dst[od] = p_src[oe];\
}
-static inline void sub_zigzag_4x4full( int level[16], const uint8_t *p_src, uint8_t *p_dst, int i_stride )
+static inline void sub_zigzag_4x4full( int level[16], const uint8_t *p_src, uint8_t *p_dst )
{
ZIG( 0,0,0) ZIG( 1,0,1) ZIG( 2,1,0) ZIG( 3,2,0)
ZIG( 4,1,1) ZIG( 5,0,2) ZIG( 6,0,3) ZIG( 7,1,2)
ZIG( 8,2,1) ZIG( 9,3,0) ZIG(10,3,1) ZIG(11,2,2)
ZIG(12,1,3) ZIG(13,2,3) ZIG(14,3,2) ZIG(15,3,3)
}
-static inline void sub_zigzag_4x4( int level[15], const uint8_t *p_src, uint8_t *p_dst, int i_stride )
+static inline void sub_zigzag_4x4( int level[15], const uint8_t *p_src, uint8_t *p_dst )
{
ZIG( 0,0,1) ZIG( 1,1,0) ZIG( 2,2,0)
ZIG( 3,1,1) ZIG( 4,0,2) ZIG( 5,0,3) ZIG( 6,1,2)
void x264_mb_encode_i4x4( x264_t *h, int idx, int i_qscale )
{
- const int i_stride = h->mb.pic.i_stride[0];
- const int i_offset = 4 * block_idx_x[idx] + 4 * block_idx_y[idx] * i_stride;
- uint8_t *p_src = &h->mb.pic.p_fenc[0][i_offset];
- uint8_t *p_dst = &h->mb.pic.p_fdec[0][i_offset];
+ int x = 4 * block_idx_x[idx];
+ int y = 4 * block_idx_y[idx];
+ uint8_t *p_src = &h->mb.pic.p_fenc[0][x+y*FENC_STRIDE];
+ uint8_t *p_dst = &h->mb.pic.p_fdec[0][x+y*FDEC_STRIDE];
int16_t dct4x4[4][4];
if( h->mb.b_lossless )
{
- sub_zigzag_4x4full( h->dct.block[idx].luma4x4, p_src, p_dst, i_stride );
+ sub_zigzag_4x4full( h->dct.block[idx].luma4x4, p_src, p_dst );
return;
}
- h->dctf.sub4x4_dct( dct4x4, p_src, i_stride, p_dst, i_stride );
+ h->dctf.sub4x4_dct( dct4x4, p_src, FENC_STRIDE, p_dst, FDEC_STRIDE );
if( h->mb.b_noise_reduction )
x264_denoise_dct( h, (int16_t*)dct4x4 );
h->quantf.dequant_4x4( dct4x4, h->dequant4_mf[CQM_4IY], i_qscale );
/* output samples to fdec */
- h->dctf.add4x4_idct( p_dst, i_stride, dct4x4 );
+ h->dctf.add4x4_idct( p_dst, FDEC_STRIDE, dct4x4 );
}
void x264_mb_encode_i8x8( x264_t *h, int idx, int i_qscale )
{
- const int i_stride = h->mb.pic.i_stride[0];
- const int i_offset = 8 * (idx&1) + 8 * (idx>>1) * i_stride;
- uint8_t *p_src = &h->mb.pic.p_fenc[0][i_offset];
- uint8_t *p_dst = &h->mb.pic.p_fdec[0][i_offset];
+ int x = 8 * (idx&1);
+ int y = 8 * (idx>>1);
+ uint8_t *p_src = &h->mb.pic.p_fenc[0][x+y*FENC_STRIDE];
+ uint8_t *p_dst = &h->mb.pic.p_fdec[0][x+y*FDEC_STRIDE];
int16_t dct8x8[8][8];
- h->dctf.sub8x8_dct8( dct8x8, p_src, i_stride, p_dst, i_stride );
+ h->dctf.sub8x8_dct8( dct8x8, p_src, FENC_STRIDE, p_dst, FDEC_STRIDE );
if( h->mb.b_noise_reduction )
x264_denoise_dct( h, (int16_t*)dct8x8 );
scan_zigzag_8x8full( h->dct.luma8x8[idx], dct8x8 );
h->quantf.dequant_8x8( dct8x8, h->dequant8_mf[CQM_8IY], i_qscale );
- h->dctf.add8x8_idct8( p_dst, i_stride, dct8x8 );
+ h->dctf.add8x8_idct8( p_dst, FDEC_STRIDE, dct8x8 );
}
static void x264_mb_encode_i16x16( x264_t *h, int i_qscale )
{
- const int i_stride = h->mb.pic.i_stride[0];
uint8_t *p_src = h->mb.pic.p_fenc[0];
uint8_t *p_dst = h->mb.pic.p_fdec[0];
{
for( i = 0; i < 16; i++ )
{
- int o = block_idx_x[i]*4 + block_idx_y[i]*4*i_stride;
- sub_zigzag_4x4( h->dct.block[i].residual_ac, p_src+o, p_dst+o, i_stride );
- dct4x4[0][block_idx_y[i]][block_idx_x[i]] = p_src[o] - p_dst[o];
- p_dst[o] = p_src[o];
+ int oe = block_idx_x[i]*4 + block_idx_y[i]*4*FENC_STRIDE;
+ int od = block_idx_x[i]*4 + block_idx_y[i]*4*FDEC_STRIDE;
+ sub_zigzag_4x4( h->dct.block[i].residual_ac, p_src+oe, p_dst+od );
+ dct4x4[0][block_idx_y[i]][block_idx_x[i]] = p_src[oe] - p_dst[od];
+ p_dst[od] = p_src[oe];
}
scan_zigzag_4x4full( h->dct.luma16x16_dc, dct4x4[0] );
return;
}
- h->dctf.sub16x16_dct( &dct4x4[1], p_src, i_stride, p_dst, i_stride );
+ h->dctf.sub16x16_dct( &dct4x4[1], p_src, FENC_STRIDE, p_dst, FDEC_STRIDE );
for( i = 0; i < 16; i++ )
{
/* copy dc coeff */
dct4x4[1+i][0][0] = dct4x4[0][block_idx_y[i]][block_idx_x[i]];
}
/* put pixels to fdec */
- h->dctf.add16x16_idct( p_dst, i_stride, &dct4x4[1] );
+ h->dctf.add16x16_idct( p_dst, FDEC_STRIDE, &dct4x4[1] );
}
static void x264_mb_encode_8x8_chroma( x264_t *h, int b_inter, int i_qscale )
for( ch = 0; ch < 2; ch++ )
{
- const int i_stride = h->mb.pic.i_stride[1+ch];
uint8_t *p_src = h->mb.pic.p_fenc[1+ch];
uint8_t *p_dst = h->mb.pic.p_fdec[1+ch];
int i_decimate_score = 0;
{
for( i = 0; i < 4; i++ )
{
- int o = block_idx_x[i]*4 + block_idx_y[i]*4*i_stride;
- sub_zigzag_4x4( h->dct.block[16+i+ch*4].residual_ac, p_src+o, p_dst+o, i_stride );
- h->dct.chroma_dc[ch][i] = p_src[o] - p_dst[o];
- p_dst[o] = p_src[o];
+ int oe = block_idx_x[i]*4 + block_idx_y[i]*4*FENC_STRIDE;
+ int od = block_idx_x[i]*4 + block_idx_y[i]*4*FDEC_STRIDE;
+ sub_zigzag_4x4( h->dct.block[16+i+ch*4].residual_ac, p_src+oe, p_dst+od );
+ h->dct.chroma_dc[ch][i] = p_src[oe] - p_dst[od];
+ p_dst[od] = p_src[oe];
}
continue;
}
- h->dctf.sub8x8_dct( dct4x4, p_src, i_stride, p_dst, i_stride );
+ h->dctf.sub8x8_dct( dct4x4, p_src, FENC_STRIDE, p_dst, FDEC_STRIDE );
/* calculate dct coeffs */
for( i = 0; i < 4; i++ )
{
/* copy dc coeff */
dct4x4[i][0][0] = dct2x2[0][i];
}
- h->dctf.add8x8_idct( p_dst, i_stride, dct4x4 );
+ h->dctf.add8x8_idct( p_dst, FDEC_STRIDE, dct4x4 );
}
}
/* Motion compensation XXX probably unneeded */
h->mc.mc_luma( h->mb.pic.p_fref[0][0], h->mb.pic.i_stride[0],
- h->mb.pic.p_fdec[0], h->mb.pic.i_stride[0],
- mvx, mvy, 16, 16 );
+ h->mb.pic.p_fdec[0], FDEC_STRIDE,
+ mvx, mvy, 16, 16 );
/* Chroma MC */
h->mc.mc_chroma( h->mb.pic.p_fref[0][0][4], h->mb.pic.i_stride[1],
- h->mb.pic.p_fdec[1], h->mb.pic.i_stride[1],
- mvx, mvy, 8, 8 );
+ h->mb.pic.p_fdec[1], FDEC_STRIDE,
+ mvx, mvy, 8, 8 );
h->mc.mc_chroma( h->mb.pic.p_fref[0][0][5], h->mb.pic.i_stride[2],
- h->mb.pic.p_fdec[2], h->mb.pic.i_stride[2],
- mvx, mvy, 8, 8 );
+ h->mb.pic.p_fdec[2], FDEC_STRIDE,
+ mvx, mvy, 8, 8 );
x264_macroblock_encode_skip( h );
}
const int i_mode = h->mb.i_intra16x16_pred_mode;
h->mb.b_transform_8x8 = 0;
/* do the right prediction */
- h->predict_16x16[i_mode]( h->mb.pic.p_fdec[0], h->mb.pic.i_stride[0] );
+ h->predict_16x16[i_mode]( h->mb.pic.p_fdec[0], FDEC_STRIDE );
/* encode the 16x16 macroblock */
x264_mb_encode_i16x16( h, i_qp );
h->mb.b_transform_8x8 = 1;
for( i = 0; i < 4; i++ )
{
- const int i_dst = h->mb.pic.i_stride[0];
+ const int i_dst = FDEC_STRIDE;
uint8_t *p_dst = &h->mb.pic.p_fdec[0][8 * (i&1) + 8 * (i>>1) * i_dst];
int i_mode = h->mb.cache.intra4x4_pred_mode[x264_scan8[4*i]];
h->mb.b_transform_8x8 = 0;
for( i = 0; i < 16; i++ )
{
- const int i_dst = h->mb.pic.i_stride[0];
+ const int i_dst = FDEC_STRIDE;
uint8_t *p_dst = &h->mb.pic.p_fdec[0][4 * block_idx_x[i] + 4 * block_idx_y[i] * i_dst];
int i_mode = h->mb.cache.intra4x4_pred_mode[x264_scan8[i]];
{
for( i4x4 = 0; i4x4 < 16; i4x4++ )
{
- int o = block_idx_x[i4x4]*4 + block_idx_y[i4x4]*4 * h->mb.pic.i_stride[0];
- sub_zigzag_4x4full( h->dct.block[i4x4].luma4x4, h->mb.pic.p_fenc[0]+o, h->mb.pic.p_fdec[0]+o, h->mb.pic.i_stride[0] );
+ int x = 4*block_idx_x[i4x4];
+ int y = 4*block_idx_y[i4x4];
+ sub_zigzag_4x4full( h->dct.block[i4x4].luma4x4,
+ h->mb.pic.p_fenc[0]+x+y*FENC_STRIDE,
+ h->mb.pic.p_fdec[0]+x+y*FDEC_STRIDE );
}
}
else if( h->mb.b_transform_8x8 )
int16_t dct8x8[4][8][8];
int nnz8x8[4] = {1,1,1,1};
h->dctf.sub16x16_dct8( dct8x8,
- h->mb.pic.p_fenc[0], h->mb.pic.i_stride[0],
- h->mb.pic.p_fdec[0], h->mb.pic.i_stride[0] );
+ h->mb.pic.p_fenc[0], FENC_STRIDE,
+ h->mb.pic.p_fdec[0], FDEC_STRIDE );
for( idx = 0; idx < 4; idx++ )
{
memset( h->dct.luma8x8, 0, sizeof( h->dct.luma8x8 ) );
else
{
- const int stride = h->mb.pic.i_stride[0];
for( idx = 0; idx < 4; idx++ )
if( nnz8x8[idx] )
{
h->quantf.dequant_8x8( dct8x8[idx], h->dequant8_mf[CQM_8PY], i_qp );
- h->dctf.add8x8_idct8( &h->mb.pic.p_fdec[0][(idx&1)*8 + (idx>>1)*8*stride], stride, dct8x8[idx] );
+ h->dctf.add8x8_idct8( &h->mb.pic.p_fdec[0][(idx&1)*8 + (idx>>1)*8*FDEC_STRIDE], FDEC_STRIDE, dct8x8[idx] );
}
}
}
int16_t dct4x4[16][4][4];
int nnz8x8[4] = {1,1,1,1};
h->dctf.sub16x16_dct( dct4x4,
- h->mb.pic.p_fenc[0], h->mb.pic.i_stride[0],
- h->mb.pic.p_fdec[0], h->mb.pic.i_stride[0] );
+ h->mb.pic.p_fenc[0], FENC_STRIDE,
+ h->mb.pic.p_fdec[0], FDEC_STRIDE );
for( i8x8 = 0; i8x8 < 4; i8x8++ )
{
memset( h->dct.block, 0, 16 * sizeof( *h->dct.block ) );
else
{
- const int stride = h->mb.pic.i_stride[0];
for( i8x8 = 0; i8x8 < 4; i8x8++ )
if( nnz8x8[i8x8] )
{
for( i = 0; i < 4; i++ )
h->quantf.dequant_4x4( dct4x4[i8x8*4+i], h->dequant4_mf[CQM_4PY], i_qp );
- h->dctf.add8x8_idct( &h->mb.pic.p_fdec[0][(i8x8&1)*8 + (i8x8>>1)*8*stride], stride, &dct4x4[i8x8*4] );
+ h->dctf.add8x8_idct( &h->mb.pic.p_fdec[0][(i8x8&1)*8 + (i8x8>>1)*8*FDEC_STRIDE], FDEC_STRIDE, &dct4x4[i8x8*4] );
}
}
}
if( IS_INTRA( h->mb.i_type ) )
{
const int i_mode = h->mb.i_chroma_pred_mode;
- h->predict_8x8c[i_mode]( h->mb.pic.p_fdec[1], h->mb.pic.i_stride[1] );
- h->predict_8x8c[i_mode]( h->mb.pic.p_fdec[2], h->mb.pic.i_stride[2] );
+ h->predict_8x8c[i_mode]( h->mb.pic.p_fdec[1], FDEC_STRIDE );
+ h->predict_8x8c[i_mode]( h->mb.pic.p_fdec[2], FDEC_STRIDE );
}
/* encode the 8x8 blocks */
/* Motion compensation */
h->mc.mc_luma( h->mb.pic.p_fref[0][0], h->mb.pic.i_stride[0],
- h->mb.pic.p_fdec[0], h->mb.pic.i_stride[0],
- mvp[0], mvp[1], 16, 16 );
+ h->mb.pic.p_fdec[0], FDEC_STRIDE,
+ mvp[0], mvp[1], 16, 16 );
}
/* get luma diff */
- h->dctf.sub16x16_dct( dct4x4, h->mb.pic.p_fenc[0], h->mb.pic.i_stride[0],
- h->mb.pic.p_fdec[0], h->mb.pic.i_stride[0] );
+ h->dctf.sub16x16_dct( dct4x4, h->mb.pic.p_fenc[0], FENC_STRIDE,
+ h->mb.pic.p_fdec[0], FDEC_STRIDE );
for( i8x8 = 0, i_decimate_mb = 0; i8x8 < 4; i8x8++ )
{
for( ch = 0; ch < 2; ch++ )
{
- const int i_stride = h->mb.pic.i_stride[1+ch];
uint8_t *p_src = h->mb.pic.p_fenc[1+ch];
uint8_t *p_dst = h->mb.pic.p_fdec[1+ch];
if( !b_bidir )
{
- h->mc.mc_chroma( h->mb.pic.p_fref[0][0][4+ch], i_stride,
- h->mb.pic.p_fdec[1+ch], i_stride,
- mvp[0], mvp[1], 8, 8 );
+ h->mc.mc_chroma( h->mb.pic.p_fref[0][0][4+ch], h->mb.pic.i_stride[1+ch],
+ h->mb.pic.p_fdec[1+ch], FDEC_STRIDE,
+ mvp[0], mvp[1], 8, 8 );
}
- h->dctf.sub8x8_dct( dct4x4, p_src, i_stride, p_dst, i_stride );
+ h->dctf.sub8x8_dct( dct4x4, p_src, FENC_STRIDE, p_dst, FDEC_STRIDE );
/* calculate dct DC */
dct2x2[0][0] = dct4x4[0][0][0];
#define COST_MV_INT( mx, my, bd, d ) \
{ \
- int cost = h->pixf.sad[i_pixel]( m->p_fenc[0], m->i_stride[0], \
+ int cost = h->pixf.sad[i_pixel]( m->p_fenc[0], FENC_STRIDE, \
&p_fref[(my)*m->i_stride[0]+(mx)], m->i_stride[0] ) \
+ p_cost_mvx[ (mx)<<2 ] \
+ p_cost_mvy[ (my)<<2 ]; \
#define COST_MV_PDE( mx, my ) \
{ \
- int cost = h->pixf.sad_pde[i_pixel]( m->p_fenc[0], m->i_stride[0], \
+ int cost = h->pixf.sad_pde[i_pixel]( m->p_fenc[0], FENC_STRIDE, \
&p_fref[(my)*m->i_stride[0]+(mx)], m->i_stride[0], \
bcost - p_cost_mvx[ (mx)<<2 ] - p_cost_mvy[ (my)<<2 ] ); \
if( cost < bcost - p_cost_mvx[ (mx)<<2 ] - p_cost_mvy[ (my)<<2 ] ) \
const int dw = x264_pixel_size[i_pixel].w;
const int dh = x264_pixel_size[i_pixel].h * stride;
static uint8_t zero[16*16] = {0,};
- const int enc_dc = h->pixf.sad[i_pixel]( m->p_fenc[0], stride, zero, 16 );
+ const int enc_dc = h->pixf.sad[i_pixel]( m->p_fenc[0], FENC_STRIDE, zero, 16 );
const uint16_t *integral_base = &m->integral[ -1 - 1*stride ];
if( h->pixf.sad_pde[i_pixel] )
{ \
int stride = 16; \
uint8_t *src = h->mc.get_ref( m->p_fref, m->i_stride[0], pix, &stride, mx, my, bw, bh ); \
- int cost = h->pixf.sad[i_pixel]( m->p_fenc[0], m->i_stride[0], src, stride ) \
+ int cost = h->pixf.sad[i_pixel]( m->p_fenc[0], FENC_STRIDE, src, stride ) \
+ p_cost_mvx[ mx ] + p_cost_mvy[ my ]; \
if( cost < bcost ) \
{ \
{ \
int stride = 16; \
uint8_t *src = h->mc.get_ref( m->p_fref, m->i_stride[0], pix, &stride, mx, my, bw, bh ); \
- int cost = h->pixf.mbcmp[i_pixel]( m->p_fenc[0], m->i_stride[0], src, stride ) \
+ int cost = h->pixf.mbcmp[i_pixel]( m->p_fenc[0], FENC_STRIDE, src, stride ) \
+ p_cost_mvx[ mx ] + p_cost_mvy[ my ]; \
if( b_chroma_me && cost < bcost ) \
{ \
h->mc.mc_chroma( m->p_fref[4], m->i_stride[1], pix, 8, mx, my, bw/2, bh/2 ); \
- cost += h->pixf.mbcmp[i_pixel+3]( m->p_fenc[1], m->i_stride[1], pix, 8 ); \
+ cost += h->pixf.mbcmp[i_pixel+3]( m->p_fenc[1], FENC_STRIDE, pix, 8 ); \
if( cost < bcost ) \
{ \
h->mc.mc_chroma( m->p_fref[5], m->i_stride[1], pix, 8, mx, my, bw/2, bh/2 ); \
- cost += h->pixf.mbcmp[i_pixel+3]( m->p_fenc[2], m->i_stride[1], pix, 8 ); \
+ cost += h->pixf.mbcmp[i_pixel+3]( m->p_fenc[2], FENC_STRIDE, pix, 8 ); \
} \
} \
if( cost < bcost ) \
h->mc.avg[i_pixel]( pix, bw, pix1[i1], bw ); \
else \
h->mc.avg_weight[i_pixel]( pix, bw, pix1[i1], bw, i_weight ); \
- cost = h->pixf.mbcmp[i_pixel]( m0->p_fenc[0], m0->i_stride[0], pix, bw ) \
+ cost = h->pixf.mbcmp[i_pixel]( m0->p_fenc[0], FENC_STRIDE, pix, bw ) \
+ p_cost_m0x[ m0x ] + p_cost_m0y[ m0y ] \
+ p_cost_m1x[ m1x ] + p_cost_m1y[ m1y ]; \
if( cost < bcost ) \
static int ssd_mb( x264_t *h )
{
- return h->pixf.ssd[PIXEL_16x16]( h->mb.pic.p_fenc[0], h->mb.pic.i_stride[0],
- h->mb.pic.p_fdec[0], h->mb.pic.i_stride[0] )
- + h->pixf.ssd[PIXEL_8x8]( h->mb.pic.p_fenc[1], h->mb.pic.i_stride[1],
- h->mb.pic.p_fdec[1], h->mb.pic.i_stride[1] )
- + h->pixf.ssd[PIXEL_8x8]( h->mb.pic.p_fenc[2], h->mb.pic.i_stride[2],
- h->mb.pic.p_fdec[2], h->mb.pic.i_stride[2] );
+ return h->pixf.ssd[PIXEL_16x16]( h->mb.pic.p_fenc[0], FENC_STRIDE,
+ h->mb.pic.p_fdec[0], FDEC_STRIDE )
+ + h->pixf.ssd[PIXEL_8x8]( h->mb.pic.p_fenc[1], FENC_STRIDE,
+ h->mb.pic.p_fdec[1], FDEC_STRIDE )
+ + h->pixf.ssd[PIXEL_8x8]( h->mb.pic.p_fenc[2], FENC_STRIDE,
+ h->mb.pic.p_fdec[2], FDEC_STRIDE );
}
static int x264_rd_cost_mb( x264_t *h, int i_lambda2 )
int i_cost_bak;
int l, i;
+ h->mb.pic.p_fenc[0] = h->mb.pic.fenc_buf;
+ h->mc.copy[PIXEL_8x8]( h->mb.pic.p_fenc[0], FENC_STRIDE, &fenc->lowres[0][i_pel_offset], i_stride, 8 );
+
if( !p0 && !p1 && !b )
goto lowres_intra_mb;
(mv1)[0], (mv1)[1], 8, 8 ); \
h->mc.avg[PIXEL_8x8]( pix1, 8, src2, stride2 ); \
i_cost = penalty + h->pixf.mbcmp[PIXEL_8x8]( \
- m[0].p_fenc[0], m[0].i_stride[0], pix1, 8 ); \
+ m[0].p_fenc[0], FENC_STRIDE, pix1, 8 ); \
if( i_bcost > i_cost ) \
{ \
i_bcost = i_cost; \
m[0].i_pixel = PIXEL_8x8;
m[0].p_cost_mv = a->p_cost_mv;
m[0].i_stride[0] = i_stride;
- m[0].p_fenc[0] = &fenc->lowres[0][ i_pel_offset ];
+ m[0].p_fenc[0] = h->mb.pic.p_fenc[0];
LOAD_HPELS_LUMA( m[0].p_fref, fref0->lowres );
if( b_bidir )
lowres_intra_mb:
{
- uint8_t *src = &fenc->lowres[0][ i_pel_offset ];
+ uint8_t pix_buf[9*FDEC_STRIDE];
+ uint8_t *pix = &pix_buf[8+FDEC_STRIDE - 1];
+ uint8_t *src = &fenc->lowres[0][i_pel_offset - 1];
int intra_penalty = 5 + 10 * b_bidir;
i_cost_bak = i_bcost;
- memcpy( pix1, src-i_stride-1, 9 );
+ memcpy( pix-FDEC_STRIDE, src-i_stride, 9 );
for( i=0; i<8; i++ )
- pix1[(i+1)*9] = src[-1+i*i_stride];
+ pix[i*FDEC_STRIDE] = src[i*i_stride];
+ pix++;
for( i = I_PRED_CHROMA_DC; i <= I_PRED_CHROMA_P; i++ )
{
int i_cost;
- h->predict_8x8c[i]( &pix1[10], 9 );
- i_cost = h->pixf.mbcmp[PIXEL_8x8]( &pix1[10], 9, src, i_stride ) + intra_penalty;
+ h->predict_8x8c[i]( pix, FDEC_STRIDE );
+ i_cost = h->pixf.mbcmp[PIXEL_8x8]( pix, FDEC_STRIDE, h->mb.pic.p_fenc[0], FENC_STRIDE ) + intra_penalty;
i_bcost = X264_MIN( i_bcost, i_cost );
}
if( i_bcost != i_cost_bak )