#define VP10_SYNC_CODE_1 0x83
#define VP10_SYNC_CODE_2 0x43
-#define VP9_FRAME_MARKER 0x2
+#define VPX_FRAME_MARKER 0x2
#ifdef __cplusplus
} // extern "C"
/* ADST in vertical */
TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
- VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
break;
case DCT_ADST:
/* ADST in horizontal */
- VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
/* DCT in vertical */
TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7,
break;
case ADST_ADST:
/* ADST in horizontal */
- VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
/* ADST in vertical */
TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
- VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
break;
default:
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif // CONFIG_VP9_HIGHBITDEPTH
- VP9_ENC_BORDER_IN_PIXELS,
+ VPX_ENC_BORDER_IN_PIXELS,
cm->byte_alignment) < 0) {
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate MFQE framebuffer");
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_DEC_BORDER_IN_PIXELS, cm->byte_alignment,
+ VPX_DEC_BORDER_IN_PIXELS, cm->byte_alignment,
NULL, NULL, NULL) < 0)
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate post-processing buffer");
// If the MV points so far into the UMV border that no visible pixels
// are used for reconstruction, the subpel part of the MV can be
// discarded and the MV limited to 16 pixels with equivalent results.
- const int spel_left = (VP9_INTERP_EXTEND + bw) << SUBPEL_BITS;
+ const int spel_left = (VPX_INTERP_EXTEND + bw) << SUBPEL_BITS;
const int spel_right = spel_left - SUBPEL_SHIFTS;
- const int spel_top = (VP9_INTERP_EXTEND + bh) << SUBPEL_BITS;
+ const int spel_top = (VPX_INTERP_EXTEND + bh) << SUBPEL_BITS;
const int spel_bottom = spel_top - SUBPEL_SHIFTS;
MV clamped_mv = {
src_mv->row * (1 << (1 - ss_y)),
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_DEC_BORDER_IN_PIXELS, cm->byte_alignment,
+ VPX_DEC_BORDER_IN_PIXELS, cm->byte_alignment,
NULL, NULL, NULL) < 0)
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate tmp restoration buffer");
/* Encoder forward decls */
struct macroblock;
-struct vp10_variance_vtable;
+struct vpx_variance_vtable;
struct search_site_config;
struct mv;
union int_mv;
#
# Motion search
#
-add_proto qw/int vp10_full_search_sad/, "const struct macroblock *x, const struct mv *ref_mv, int sad_per_bit, int distance, const struct vp10_variance_vtable *fn_ptr, const struct mv *center_mv, struct mv *best_mv";
+add_proto qw/int vp10_full_search_sad/, "const struct macroblock *x, const struct mv *ref_mv, int sad_per_bit, int distance, const struct vpx_variance_vtable *fn_ptr, const struct mv *center_mv, struct mv *best_mv";
specialize qw/vp10_full_search_sad sse3 sse4_1/;
$vp10_full_search_sad_sse3=vp10_full_search_sadx3;
$vp10_full_search_sad_sse4_1=vp10_full_search_sadx8;
-add_proto qw/int vp10_diamond_search_sad/, "struct macroblock *x, const struct search_site_config *cfg, struct mv *ref_mv, struct mv *best_mv, int search_param, int sad_per_bit, int *num00, const struct vp10_variance_vtable *fn_ptr, const struct mv *center_mv";
+add_proto qw/int vp10_diamond_search_sad/, "struct macroblock *x, const struct search_site_config *cfg, struct mv *ref_mv, struct mv *best_mv, int search_param, int sad_per_bit, int *num00, const struct vpx_variance_vtable *fn_ptr, const struct mv *center_mv";
specialize qw/vp10_diamond_search_sad/;
-add_proto qw/int vp10_full_range_search/, "const struct macroblock *x, const struct search_site_config *cfg, struct mv *ref_mv, struct mv *best_mv, int search_param, int sad_per_bit, int *num00, const struct vp10_variance_vtable *fn_ptr, const struct mv *center_mv";
+add_proto qw/int vp10_full_range_search/, "const struct macroblock *x, const struct search_site_config *cfg, struct mv *ref_mv, struct mv *best_mv, int search_param, int sad_per_bit, int *num00, const struct vpx_variance_vtable *fn_ptr, const struct mv *center_mv";
specialize qw/vp10_full_range_search/;
add_proto qw/void vp10_temporal_filter_apply/, "uint8_t *frame1, unsigned int stride, uint8_t *frame2, unsigned int block_width, unsigned int block_height, int strength, int filter_weight, unsigned int *accumulator, uint16_t *count";
#include "vp10/decoder/decoder.h"
#include "vp10/decoder/dsubexp.h"
-#define MAX_VP9_HEADER_SIZE 80
+#define MAX_VPX_HEADER_SIZE 80
static int is_compound_reference_allowed(const VP10_COMMON *cm) {
int i;
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_DEC_BORDER_IN_PIXELS,
+ VPX_DEC_BORDER_IN_PIXELS,
cm->byte_alignment,
&pool->frame_bufs[cm->new_fb_idx].raw_frame_buffer, pool->get_fb_cb,
pool->cb_priv)) {
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_DEC_BORDER_IN_PIXELS,
+ VPX_DEC_BORDER_IN_PIXELS,
cm->byte_alignment,
&pool->frame_bufs[cm->new_fb_idx].raw_frame_buffer, pool->get_fb_cb,
pool->cb_priv)) {
cm->is_reference_frame = 1;
#endif // CONFIG_EXT_REFS
- if (vpx_rb_read_literal(rb, 2) != VP9_FRAME_MARKER)
+ if (vpx_rb_read_literal(rb, 2) != VPX_FRAME_MARKER)
vpx_internal_error(&cm->error, VPX_CODEC_UNSUP_BITSTREAM,
"Invalid frame marker");
struct vpx_read_bit_buffer *rb,
const uint8_t *data,
const uint8_t *data_end,
- uint8_t clear_data[MAX_VP9_HEADER_SIZE]) {
+ uint8_t clear_data[MAX_VPX_HEADER_SIZE]) {
rb->bit_offset = 0;
rb->error_handler = error_handler;
rb->error_handler_data = &pbi->common;
if (pbi->decrypt_cb) {
- const int n = (int)VPXMIN(MAX_VP9_HEADER_SIZE, data_end - data);
+ const int n = (int)VPXMIN(MAX_VPX_HEADER_SIZE, data_end - data);
pbi->decrypt_cb(pbi->decrypt_state, data, clear_data, n);
rb->bit_buffer = clear_data;
rb->bit_buffer_end = clear_data + n;
MACROBLOCKD *const xd = &pbi->mb;
struct vpx_read_bit_buffer rb;
int context_updated = 0;
- uint8_t clear_data[MAX_VP9_HEADER_SIZE];
+ uint8_t clear_data[MAX_VPX_HEADER_SIZE];
const size_t first_partition_size = read_uncompressed_header(pbi,
init_read_bit_buffer(pbi, &rb, data, data_end, clear_data));
YV12_BUFFER_CONFIG *const new_fb = get_frame_new_buffer(cm);
VP10_COMMON *const cm = &cpi->common;
MACROBLOCKD *const xd = &cpi->td.mb.e_mbd;
- vpx_wb_write_literal(wb, VP9_FRAME_MARKER, 2);
+ vpx_wb_write_literal(wb, VPX_FRAME_MARKER, 2);
write_profile(cm->profile, wb);
// Set up limit values for MV components.
// Mv beyond the range do not produce new/different prediction block.
- x->mv_row_min = -(((mi_row + mi_height) * MI_SIZE) + VP9_INTERP_EXTEND);
- x->mv_col_min = -(((mi_col + mi_width) * MI_SIZE) + VP9_INTERP_EXTEND);
- x->mv_row_max = (cm->mi_rows - mi_row) * MI_SIZE + VP9_INTERP_EXTEND;
- x->mv_col_max = (cm->mi_cols - mi_col) * MI_SIZE + VP9_INTERP_EXTEND;
+ x->mv_row_min = -(((mi_row + mi_height) * MI_SIZE) + VPX_INTERP_EXTEND);
+ x->mv_col_min = -(((mi_col + mi_width) * MI_SIZE) + VPX_INTERP_EXTEND);
+ x->mv_row_max = (cm->mi_rows - mi_row) * MI_SIZE + VPX_INTERP_EXTEND;
+ x->mv_col_max = (cm->mi_cols - mi_col) * MI_SIZE + VPX_INTERP_EXTEND;
// Set up distance of MB to edge of frame in 1/8th pel units.
assert(!(mi_col & (mi_width - 1)) && !(mi_row & (mi_height - 1)));
// Set up limit values for MV components.
// Mv beyond the range do not produce new/different prediction block.
- x->mv_row_min = -(((mi_row_pred + mi_height) * MI_SIZE) + VP9_INTERP_EXTEND);
- x->mv_col_min = -(((mi_col_pred + mi_width) * MI_SIZE) + VP9_INTERP_EXTEND);
- x->mv_row_max = (cm->mi_rows - mi_row_pred) * MI_SIZE + VP9_INTERP_EXTEND;
- x->mv_col_max = (cm->mi_cols - mi_col_pred) * MI_SIZE + VP9_INTERP_EXTEND;
+ x->mv_row_min = -(((mi_row_pred + mi_height) * MI_SIZE) + VPX_INTERP_EXTEND);
+ x->mv_col_min = -(((mi_col_pred + mi_width) * MI_SIZE) + VPX_INTERP_EXTEND);
+ x->mv_row_max = (cm->mi_rows - mi_row_pred) * MI_SIZE + VPX_INTERP_EXTEND;
+ x->mv_col_max = (cm->mi_cols - mi_col_pred) * MI_SIZE + VPX_INTERP_EXTEND;
// Set up distance of MB to edge of frame in 1/8th pel units.
assert(!(mi_col_pred & (mi_width - 1)) && !(mi_row_pred & (mi_height - 1)));
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
+ VPX_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate altref buffer");
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
+ VPX_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate last frame buffer");
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
+ VPX_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate last frame deblocked buffer");
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
+ VPX_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate scaled source buffer");
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
+ VPX_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate scaled last source buffer");
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- (VP9_ENC_BORDER_IN_PIXELS << 3),
+ (VPX_ENC_BORDER_IN_PIXELS << 3),
cm->byte_alignment,
NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
if (vpx_realloc_frame_buffer(&new_fb_ptr->buf, cm->width, cm->height,
cm->subsampling_x, cm->subsampling_y,
cm->use_highbitdepth,
- VP9_ENC_BORDER_IN_PIXELS,
+ VPX_ENC_BORDER_IN_PIXELS,
cm->byte_alignment, NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate frame buffer");
new_fb_ptr->buf.y_crop_height != cm->height) {
if (vpx_realloc_frame_buffer(&new_fb_ptr->buf, cm->width, cm->height,
cm->subsampling_x, cm->subsampling_y,
- VP9_ENC_BORDER_IN_PIXELS,
+ VPX_ENC_BORDER_IN_PIXELS,
cm->byte_alignment, NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate frame buffer");
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- (VP9_ENC_BORDER_IN_PIXELS << 3),
+ (VPX_ENC_BORDER_IN_PIXELS << 3),
cm->byte_alignment,
NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
+ VPX_ENC_BORDER_IN_PIXELS, cm->byte_alignment,
NULL, NULL, NULL))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate frame buffer");
#if CONFIG_VP9_HIGHBITDEPTH
cm->use_highbitdepth,
#endif
- VP9_ENC_BORDER_IN_PIXELS))
+ VPX_ENC_BORDER_IN_PIXELS))
vpx_internal_error(&cm->error, VPX_CODEC_MEM_ERROR,
"Failed to allocate denoiser");
}
fractional_mv_step_fp *find_fractional_mv_step;
vp10_full_search_fn_t full_search_sad; // It is currently unused.
vp10_diamond_search_fn_t diamond_search_sad;
- vp10_variance_fn_ptr_t fn_ptr[BLOCK_SIZES];
+ vpx_variance_fn_ptr_t fn_ptr[BLOCK_SIZES];
uint64_t time_receive_data;
uint64_t time_compress_data;
uint64_t time_pick_lpf;
MV ref_mv_full = {ref_mv->row >> 3, ref_mv->col >> 3};
int num00, tmp_err, n;
const BLOCK_SIZE bsize = xd->mi[0]->mbmi.sb_type;
- vp10_variance_fn_ptr_t v_fn_ptr = cpi->fn_ptr[bsize];
+ vpx_variance_fn_ptr_t v_fn_ptr = cpi->fn_ptr[bsize];
const int new_mv_mode_penalty = NEW_MV_MODE_PENALTY;
int step_param = 3;
#if CONFIG_VP9_HIGHBITDEPTH
use_highbitdepth,
#endif
- VP9_ENC_BORDER_IN_PIXELS,
+ VPX_ENC_BORDER_IN_PIXELS,
legacy_byte_alignment))
goto bail;
}
#if CONFIG_VP9_HIGHBITDEPTH
use_highbitdepth,
#endif
- VP9_ENC_BORDER_IN_PIXELS,
+ VPX_ENC_BORDER_IN_PIXELS,
0))
return 1;
vpx_free_frame_buffer(&buf->img);
MACROBLOCK *const x = &cpi->td.mb;
MACROBLOCKD *const xd = &x->e_mbd;
const MV_SPEED_FEATURES *const mv_sf = &cpi->sf.mv;
- const vp10_variance_fn_ptr_t v_fn_ptr = cpi->fn_ptr[BLOCK_16X16];
+ const vpx_variance_fn_ptr_t v_fn_ptr = cpi->fn_ptr[BLOCK_16X16];
const int tmp_col_min = x->mv_col_min;
const int tmp_col_max = x->mv_col_max;
const MV *bestmv,
const MV *ref_mv,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
const uint8_t *const src,
const int src_stride,
const uint8_t *const y,
const MV *ref_mv,
int allow_hp,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop,
int iters_per_step,
int *cost_list,
const MV *ref_mv,
int allow_hp,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop,
int iters_per_step,
int *cost_list,
const MV *ref_mv,
int allow_hp,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop,
int iters_per_step,
int *cost_list,
};
static int upsampled_pref_error(const MACROBLOCKD *xd,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
const uint8_t *const src, const int src_stride,
const uint8_t *const y, int y_stride,
const uint8_t *second_pred,
static unsigned int upsampled_setup_center_error(
const MACROBLOCKD *xd, const MV *bestmv, const MV *ref_mv,
- int error_per_bit, const vp10_variance_fn_ptr_t *vfp,
+ int error_per_bit, const vpx_variance_fn_ptr_t *vfp,
const uint8_t *const src, const int src_stride,
const uint8_t *const y, int y_stride, const uint8_t *second_pred,
int w, int h, int offset, int *mvjcost, int *mvcost[2],
const MV *ref_mv,
int allow_hp,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop,
int iters_per_step,
int *cost_list,
static INLINE void calc_int_cost_list(const MACROBLOCK *x,
const MV *const ref_mv,
int sadpb,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *best_mv,
int *cost_list) {
static const MV neighbors[4] = {{0, -1}, {1, 0}, {0, 1}, {-1, 0}};
static INLINE void calc_int_sad_list(const MACROBLOCK *x,
const MV *const ref_mv,
int sadpb,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *best_mv, int *cost_list,
const int use_mvcost,
const int bestsad) {
int sad_per_bit,
int do_init_search,
int *cost_list,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost,
const MV *center_mv,
const int num_candidates[MAX_PATTERN_SCALES],
int vp10_get_mvpred_var(const MACROBLOCK *x,
const MV *best_mv, const MV *center_mv,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost) {
const MACROBLOCKD *const xd = &x->e_mbd;
const struct buf_2d *const what = &x->plane[0].src;
int vp10_get_mvpred_av_var(const MACROBLOCK *x,
const MV *best_mv, const MV *center_mv,
const uint8_t *second_pred,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost) {
const MACROBLOCKD *const xd = &x->e_mbd;
const struct buf_2d *const what = &x->plane[0].src;
int sad_per_bit,
int do_init_search,
int *cost_list,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost,
const MV *center_mv) {
// First scale has 8-closest points, the rest have 6 points in hex shape
int sad_per_bit,
int do_init_search,
int *cost_list,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost,
const MV *center_mv) {
// First scale has 4-closest points, the rest have 8 points in diamond
int sad_per_bit,
int do_init_search,
int *cost_list,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost,
const MV *center_mv) {
// All scales have 8 closest points in square shape
int sad_per_bit,
int do_init_search, // must be zero for fast_hex
int *cost_list,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost,
const MV *center_mv) {
return vp10_hex_search(
int sad_per_bit,
int do_init_search,
int *cost_list,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost,
const MV *center_mv) {
return bigdia_search(
static int exhuastive_mesh_search(MACROBLOCK *x,
MV *ref_mv, MV *best_mv,
int range, int step, int sad_per_bit,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv) {
const MACROBLOCKD *const xd = &x->e_mbd;
const struct buf_2d *const what = &x->plane[0].src;
const search_site_config *cfg,
MV *ref_mv, MV *best_mv, int search_param,
int sad_per_bit, int *num00,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv) {
int i, j, step;
MV *mvp_full, int step_param,
int sadpb, int further_steps, int do_refine,
int *cost_list,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *ref_mv) {
MV temp_mv;
int thissme, n, num00 = 0;
static int full_pixel_exhaustive(VP10_COMP *cpi, MACROBLOCK *x,
const MV *centre_mv_full, int sadpb,
int *cost_list,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *ref_mv, MV *dst_mv) {
const SPEED_FEATURES *const sf = &cpi->sf;
MV temp_mv = {centre_mv_full->row, centre_mv_full->col};
int vp10_full_search_sad_c(const MACROBLOCK *x, const MV *ref_mv,
int sad_per_bit, int distance,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv, MV *best_mv) {
int r, c;
const MACROBLOCKD *const xd = &x->e_mbd;
int vp10_full_search_sadx3(const MACROBLOCK *x, const MV *ref_mv,
int sad_per_bit, int distance,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv, MV *best_mv) {
int r;
const MACROBLOCKD *const xd = &x->e_mbd;
int vp10_full_search_sadx8(const MACROBLOCK *x, const MV *ref_mv,
int sad_per_bit, int distance,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv, MV *best_mv) {
int r;
const MACROBLOCKD *const xd = &x->e_mbd;
int vp10_refining_search_sad(MACROBLOCK *x,
MV *ref_mv, int error_per_bit,
int search_range,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv) {
const MACROBLOCKD *const xd = &x->e_mbd;
const MV neighbors[4] = {{ -1, 0}, {0, -1}, {0, 1}, {1, 0}};
int vp10_refining_search_8p_c(MACROBLOCK *x,
int error_per_bit,
int search_range,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv,
const uint8_t *second_pred) {
const MV neighbors[8] = {{-1, 0}, {0, -1}, {0, 1}, {1, 0},
int var_max, int rd) {
const SPEED_FEATURES *const sf = &cpi->sf;
const SEARCH_METHODS method = sf->mv.search_method;
- vp10_variance_fn_ptr_t *fn_ptr = &cpi->fn_ptr[bsize];
+ vpx_variance_fn_ptr_t *fn_ptr = &cpi->fn_ptr[bsize];
int var = 0;
if (cost_list) {
MV *bestmv, const MV *ref_mv,
int allow_hp,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop,
int iters_per_step,
int *mvjcost, int *mvcost[2],
const MV *bestmv,
const MV *ref_mv,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
const uint8_t *const src,
const int src_stride,
const uint8_t *const y,
static int upsampled_masked_pref_error(const MACROBLOCKD *xd,
const uint8_t *mask,
int mask_stride,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
const uint8_t *const src,
const int src_stride,
const uint8_t *const y, int y_stride,
const MACROBLOCKD *xd,
const uint8_t *mask, int mask_stride,
const MV *bestmv, const MV *ref_mv,
- int error_per_bit, const vp10_variance_fn_ptr_t *vfp,
+ int error_per_bit, const vpx_variance_fn_ptr_t *vfp,
const uint8_t *const src, const int src_stride,
const uint8_t *const y, int y_stride,
int w, int h, int offset, int *mvjcost, int *mvcost[2],
MV *bestmv, const MV *ref_mv,
int allow_hp,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop,
int iters_per_step,
int *mvjcost, int *mvcost[2],
static int get_masked_mvpred_var(const MACROBLOCK *x,
const uint8_t *mask, int mask_stride,
const MV *best_mv, const MV *center_mv,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost, int is_second) {
const MACROBLOCKD *const xd = &x->e_mbd;
const struct buf_2d *const what = &x->plane[0].src;
const uint8_t *mask, int mask_stride,
MV *ref_mv, int error_per_bit,
int search_range,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv, int is_second) {
const MV neighbors[4] = {{ -1, 0}, {0, -1}, {0, 1}, {1, 0}};
const MACROBLOCKD *const xd = &x->e_mbd;
MV *ref_mv, MV *best_mv,
int search_param,
int sad_per_bit, int *num00,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv, int is_second) {
const MACROBLOCKD *const xd = &x->e_mbd;
const struct buf_2d *const what = &x->plane[0].src;
const uint8_t *mask, int mask_stride,
MV *mvp_full, int step_param,
int sadpb, int further_steps, int do_refine,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *ref_mv, MV *dst_mv,
int is_second) {
MV temp_mv;
const MV *bestmv,
const MV *ref_mv,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
const int32_t *const wsrc,
const uint8_t *const y,
int y_stride,
static int upsampled_obmc_pref_error(const MACROBLOCKD *xd,
const int32_t *mask,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
const int32_t *const wsrc,
const uint8_t *const y, int y_stride,
int w, int h, unsigned int *sse) {
const MACROBLOCKD *xd,
const int32_t *mask,
const MV *bestmv, const MV *ref_mv,
- int error_per_bit, const vp10_variance_fn_ptr_t *vfp,
+ int error_per_bit, const vpx_variance_fn_ptr_t *vfp,
const int32_t *const wsrc,
const uint8_t *const y, int y_stride,
int w, int h, int offset, int *mvjcost, int *mvcost[2],
int mi_row, int mi_col,
MV *bestmv, const MV *ref_mv,
int allow_hp, int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop, int iters_per_step,
int *mvjcost, int *mvcost[2],
int *distortion, unsigned int *sse1,
const int32_t *wsrc,
const int32_t *mask,
const MV *best_mv, const MV *center_mv,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost, int is_second) {
const MACROBLOCKD *const xd = &x->e_mbd;
const struct buf_2d *const in_what = &xd->plane[0].pre[is_second];
const int32_t *mask,
MV *ref_mv, int error_per_bit,
int search_range,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv, int is_second) {
const MV neighbors[4] = {{ -1, 0}, {0, -1}, {0, 1}, {1, 0}};
const MACROBLOCKD *const xd = &x->e_mbd;
MV *ref_mv, MV *best_mv,
int search_param,
int sad_per_bit, int *num00,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv, int is_second) {
const MACROBLOCKD *const xd = &x->e_mbd;
const struct buf_2d *const in_what = &xd->plane[0].pre[is_second];
const int32_t *wsrc, const int32_t *mask,
MV *mvp_full, int step_param,
int sadpb, int further_steps, int do_refine,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *ref_mv, MV *dst_mv,
int is_second) {
MV temp_mv;
#define MAX_FIRST_STEP (1 << (MAX_MVSEARCH_STEPS-1))
// Allowed motion vector pixel distance outside image border
// for Block_16x16
-#define BORDER_MV_PIXELS_B16 (16 + VP9_INTERP_EXTEND)
+#define BORDER_MV_PIXELS_B16 (16 + VPX_INTERP_EXTEND)
// motion search site
typedef struct search_site {
// Utility to compute variance + MV rate cost for a given MV
int vp10_get_mvpred_var(const MACROBLOCK *x,
const MV *best_mv, const MV *center_mv,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost);
int vp10_get_mvpred_av_var(const MACROBLOCK *x,
const MV *best_mv, const MV *center_mv,
const uint8_t *second_pred,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost);
struct VP10_COMP;
int vp10_refining_search_sad(struct macroblock *x,
struct mv *ref_mv,
int sad_per_bit, int distance,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const struct mv *center_mv);
// Runs sequence of diamond searches in smaller steps for RD.
MV *mvp_full, int step_param,
int sadpb, int further_steps, int do_refine,
int *cost_list,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *ref_mv, MV *dst_mv);
// Perform integral projection based motion estimation.
int sad_per_bit,
int do_init_search,
int *cost_list,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int use_mvcost,
const MV *center_mv);
const MV *ref_mv,
int allow_hp,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop, // 0 - full, 1 - qtr only, 2 - half only
int iters_per_step,
int *cost_list,
typedef int (*vp10_full_search_fn_t)(const MACROBLOCK *x,
const MV *ref_mv, int sad_per_bit,
int distance,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv, MV *best_mv);
typedef int (*vp10_diamond_search_fn_t)(MACROBLOCK *x,
MV *ref_mv, MV *best_mv,
int search_param, int sad_per_bit,
int *num00,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv);
int vp10_refining_search_8p_c(MACROBLOCK *x,
int error_per_bit,
int search_range,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *center_mv, const uint8_t *second_pred);
struct VP10_COMP;
MV *bestmv, const MV *ref_mv,
int allow_hp,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop,
int iters_per_step,
int *mvjcost, int *mvcost[2],
MV *bestmv, const MV *ref_mv,
int allow_hp,
int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop,
int iters_per_step,
int *mvjcost, int *mvcost[2],
const uint8_t *mask, int mask_stride,
MV *mvp_full, int step_param,
int sadpb, int further_steps, int do_refine,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *ref_mv, MV *dst_mv,
int is_second);
#endif // CONFIG_EXT_INTER
const int32_t *mask,
MV *mvp_full, int step_param,
int sadpb, int further_steps, int do_refine,
- const vp10_variance_fn_ptr_t *fn_ptr,
+ const vpx_variance_fn_ptr_t *fn_ptr,
const MV *ref_mv, MV *dst_mv,
int is_second);
int vp10_find_best_obmc_sub_pixel_tree_up(struct VP10_COMP *cpi, MACROBLOCK *x,
int mi_row, int mi_col,
MV *bestmv, const MV *ref_mv,
int allow_hp, int error_per_bit,
- const vp10_variance_fn_ptr_t *vfp,
+ const vpx_variance_fn_ptr_t *vfp,
int forced_stop, int iters_per_step,
int *mvjcost, int *mvcost[2],
int *distortion, unsigned int *sse1,
VP9_FDCT4(in0, in1, in2, in3, in0, in1, in2, in3);
break;
case ADST_DCT:
- VP9_FADST4(in0, in1, in2, in3, in0, in1, in2, in3);
+ VPX_FADST4(in0, in1, in2, in3, in0, in1, in2, in3);
TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3);
VP9_FDCT4(in0, in1, in2, in3, in0, in1, in2, in3);
break;
case DCT_ADST:
VP9_FDCT4(in0, in1, in2, in3, in0, in1, in2, in3);
TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3);
- VP9_FADST4(in0, in1, in2, in3, in0, in1, in2, in3);
+ VPX_FADST4(in0, in1, in2, in3, in0, in1, in2, in3);
break;
case ADST_ADST:
- VP9_FADST4(in0, in1, in2, in3, in0, in1, in2, in3);
+ VPX_FADST4(in0, in1, in2, in3, in0, in1, in2, in3);
TRANSPOSE4x4_SH_SH(in0, in1, in2, in3, in0, in1, in2, in3);
- VP9_FADST4(in0, in1, in2, in3, in0, in1, in2, in3);
+ VPX_FADST4(in0, in1, in2, in3, in0, in1, in2, in3);
break;
default:
assert(0);
switch (tx_type) {
case DCT_DCT:
- VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
- VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
break;
case ADST_DCT:
- VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
- VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
break;
case DCT_ADST:
- VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
- VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
break;
case ADST_ADST:
- VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
- VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
break;
default:
#include "vpx_dsp/mips/txfm_macros_msa.h"
#include "vpx_ports/mem.h"
-#define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \
+#define VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \
out0, out1, out2, out3, out4, out5, out6, out7) { \
v8i16 cnst0_m, cnst1_m, cnst2_m, cnst3_m, cnst4_m; \
v8i16 vec0_m, vec1_m, vec2_m, vec3_m, s0_m, s1_m; \
out5 = -out5; \
}
-#define VP9_FADST4(in0, in1, in2, in3, out0, out1, out2, out3) { \
+#define VPX_FADST4(in0, in1, in2, in3, out0, out1, out2, out3) { \
v4i32 s0_m, s1_m, s2_m, s3_m, constant_m; \
v4i32 in0_r_m, in1_r_m, in2_r_m, in3_r_m; \
\
(mode_mv[NEARMV][ref_frame].as_int == INVALID_MV)));
}
-#define LEFT_TOP_MARGIN ((VP9_ENC_BORDER_IN_PIXELS - VP9_INTERP_EXTEND) << 3)
-#define RIGHT_BOTTOM_MARGIN ((VP9_ENC_BORDER_IN_PIXELS -\
- VP9_INTERP_EXTEND) << 3)
+#define LEFT_TOP_MARGIN ((VPX_ENC_BORDER_IN_PIXELS - VPX_INTERP_EXTEND) << 3)
+#define RIGHT_BOTTOM_MARGIN ((VPX_ENC_BORDER_IN_PIXELS -\
+ VPX_INTERP_EXTEND) << 3)
// TODO(jingning): this mv clamping function should be block size dependent.
static INLINE void clamp_mv2(MV *mv, const MACROBLOCKD *xd) {
memset(&skinmap, 0, sizeof(YV12_BUFFER_CONFIG));
if (vpx_alloc_frame_buffer(&skinmap, cm->width, cm->height,
cm->subsampling_x, cm->subsampling_y,
- VP9_ENC_BORDER_IN_PIXELS, cm->byte_alignment)) {
+ VPX_ENC_BORDER_IN_PIXELS, cm->byte_alignment)) {
vpx_free_frame_buffer(&skinmap);
return;
}
// L/A/G reference frames that have a border of 32 (VP9ENCBORDERINPIXELS)
// A 6/8 tap filter is used for motion search. This requires 2 pixels
// before and 3 pixels after. So the largest Y mv on a border would
- // then be 16 - VP9_INTERP_EXTEND. The UV blocks are half the size of the
+ // then be 16 - VPX_INTERP_EXTEND. The UV blocks are half the size of the
// Y and therefore only extended by 8. The largest mv that a UV block
- // can support is 8 - VP9_INTERP_EXTEND. A UV mv is half of a Y mv.
- // (16 - VP9_INTERP_EXTEND) >> 1 which is greater than
- // 8 - VP9_INTERP_EXTEND.
+ // can support is 8 - VPX_INTERP_EXTEND. A UV mv is half of a Y mv.
+ // (16 - VPX_INTERP_EXTEND) >> 1 which is greater than
+ // 8 - VPX_INTERP_EXTEND.
// To keep the mv in play for both Y and UV planes the max that it
- // can be on a border is therefore 16 - (2*VP9_INTERP_EXTEND+1).
- cpi->td.mb.mv_row_min = -((mb_row * 16) + (17 - 2 * VP9_INTERP_EXTEND));
+ // can be on a border is therefore 16 - (2*VPX_INTERP_EXTEND+1).
+ cpi->td.mb.mv_row_min = -((mb_row * 16) + (17 - 2 * VPX_INTERP_EXTEND));
cpi->td.mb.mv_row_max = ((mb_rows - 1 - mb_row) * 16)
- + (17 - 2 * VP9_INTERP_EXTEND);
+ + (17 - 2 * VPX_INTERP_EXTEND);
for (mb_col = 0; mb_col < mb_cols; mb_col++) {
int i, j, k;
memset(accumulator, 0, 16 * 16 * 3 * sizeof(accumulator[0]));
memset(count, 0, 16 * 16 * 3 * sizeof(count[0]));
- cpi->td.mb.mv_col_min = -((mb_col * 16) + (17 - 2 * VP9_INTERP_EXTEND));
+ cpi->td.mb.mv_col_min = -((mb_col * 16) + (17 - 2 * VPX_INTERP_EXTEND));
cpi->td.mb.mv_col_max = ((mb_cols - 1 - mb_col) * 16)
- + (17 - 2 * VP9_INTERP_EXTEND);
+ + (17 - 2 * VPX_INTERP_EXTEND);
for (frame = 0; frame < frame_count; frame++) {
const int thresh_low = 10000;
const int frame_marker = vpx_rb_read_literal(&rb, 2);
const BITSTREAM_PROFILE profile = vp10_read_profile(&rb);
- if (frame_marker != VP9_FRAME_MARKER)
+ if (frame_marker != VPX_FRAME_MARKER)
return VPX_CODEC_UNSUP_BITSTREAM;
if (profile >= MAX_PROFILES)
img->range = yv12->color_range;
img->bit_depth = 8;
img->w = yv12->y_stride;
- img->h = ALIGN_POWER_OF_TWO(yv12->y_height + 2 * VP9_ENC_BORDER_IN_PIXELS, 3);
+ img->h = ALIGN_POWER_OF_TWO(yv12->y_height + 2 * VPX_ENC_BORDER_IN_PIXELS, 3);
img->d_w = yv12->y_crop_width;
img->d_h = yv12->y_crop_height;
img->r_w = yv12->render_width;
LD_SH8(input, src_stride, in0, in1, in2, in3, in4, in5, in6, in7);
SLLI_4V(in0, in1, in2, in3, 2);
SLLI_4V(in4, in5, in6, in7, 2);
- VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
- VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
+ VPX_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7,
in0, in1, in2, in3, in4, in5, in6, in7);
in4, in5, in6, in7); \
}
-#define VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, \
+#define VPX_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, \
out0, out1, out2, out3, out4, out5, out6, out7) { \
v8i16 s0_m, s1_m, s2_m, s3_m, s4_m, s5_m, s6_m; \
v8i16 s7_m, x0_m, x1_m, x2_m, x3_m; \
#include "vpx_dsp/mips/txfm_macros_msa.h"
#include "vpx_dsp/txfm_common.h"
-#define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \
+#define VPX_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \
out0, out1, out2, out3, out4, out5, out6, out7) { \
v8i16 cnst0_m, cnst1_m, cnst2_m, cnst3_m, cnst4_m; \
v8i16 vec0_m, vec1_m, vec2_m, vec3_m, s0_m, s1_m; \
#endif // CONFIG_VP10 && CONFIG_OBMC
#if CONFIG_VP10
-typedef struct vp10_variance_vtable {
+typedef struct vpx_variance_vtable {
vpx_sad_fn_t sdf;
vpx_sad_avg_fn_t sdaf;
vpx_variance_fn_t vf;
vpx_obmc_variance_fn_t ovf;
vpx_obmc_subpixvariance_fn_t osvf;
#endif // CONFIG_OBMC
-} vp10_variance_fn_ptr_t;
+} vpx_variance_fn_ptr_t;
#endif // CONFIG_VP10
void vpx_highbd_var_filter_block2d_bil_first_pass(
}
void vpx_extend_frame_inner_borders_c(YV12_BUFFER_CONFIG *ybf) {
- const int inner_bw = (ybf->border > VP9INNERBORDERINPIXELS) ?
- VP9INNERBORDERINPIXELS : ybf->border;
+ const int inner_bw = (ybf->border > VPXINNERBORDERINPIXELS) ?
+ VPXINNERBORDERINPIXELS : ybf->border;
extend_frame(ybf, inner_bw);
}
}
void vpx_extend_frame_inner_borders_dspr2(YV12_BUFFER_CONFIG *ybf) {
- const int inner_bw = (ybf->border > VP9INNERBORDERINPIXELS) ?
- VP9INNERBORDERINPIXELS : ybf->border;
+ const int inner_bw = (ybf->border > VPXINNERBORDERINPIXELS) ?
+ VPXINNERBORDERINPIXELS : ybf->border;
extend_frame(ybf, inner_bw);
}
#endif
#define VP8BORDERINPIXELS 32
#if CONFIG_EXT_PARTITION
-# define VP9INNERBORDERINPIXELS 160
+# define VPXINNERBORDERINPIXELS 160
#else
-# define VP9INNERBORDERINPIXELS 96
+# define VPXINNERBORDERINPIXELS 96
#endif // CONFIG_EXT_PARTITION
-#define VP9_INTERP_EXTEND 4
-#define VP9_ENC_BORDER_IN_PIXELS 160
-#define VP9_DEC_BORDER_IN_PIXELS 160
+#define VPX_INTERP_EXTEND 4
+#define VPX_ENC_BORDER_IN_PIXELS 160
+#define VPX_DEC_BORDER_IN_PIXELS 160
typedef struct yv12_buffer_config {
int y_width;