diff options
Diffstat (limited to 'vp8/encoder/ppc/csystemdependent.c')
-rw-r--r-- | vp8/encoder/ppc/csystemdependent.c | 168 |
1 files changed, 168 insertions, 0 deletions
diff --git a/vp8/encoder/ppc/csystemdependent.c b/vp8/encoder/ppc/csystemdependent.c new file mode 100644 index 000000000..f99277f99 --- /dev/null +++ b/vp8/encoder/ppc/csystemdependent.c @@ -0,0 +1,168 @@ +/* + * Copyright (c) 2010 The VP8 project authors. All Rights Reserved. + * + * Use of this source code is governed by a BSD-style license and patent + * grant that can be found in the LICENSE file in the root of the source + * tree. All contributing project authors may be found in the AUTHORS + * file in the root of the source tree. + */ + + +#include "variance.h" +#include "onyx_int.h" + +SADFunction *vp8_sad16x16; +SADFunction *vp8_sad16x8; +SADFunction *vp8_sad8x16; +SADFunction *vp8_sad8x8; +SADFunction *vp8_sad4x4; + +variance_function *vp8_variance4x4; +variance_function *vp8_variance8x8; +variance_function *vp8_variance8x16; +variance_function *vp8_variance16x8; +variance_function *vp8_variance16x16; + +variance_function *vp8_mse16x16; + +sub_pixel_variance_function *vp8_sub_pixel_variance4x4; +sub_pixel_variance_function *vp8_sub_pixel_variance8x8; +sub_pixel_variance_function *vp8_sub_pixel_variance8x16; +sub_pixel_variance_function *vp8_sub_pixel_variance16x8; +sub_pixel_variance_function *vp8_sub_pixel_variance16x16; + +int (*vp8_block_error)(short *coeff, short *dqcoeff); +int (*vp8_mbblock_error)(MACROBLOCK *mb, int dc); + +int (*vp8_mbuverror)(MACROBLOCK *mb); +unsigned int (*vp8_get_mb_ss)(short *); +void (*vp8_short_fdct4x4)(short *input, short *output, int pitch); +void (*vp8_short_fdct8x4)(short *input, short *output, int pitch); +void (*vp8_fast_fdct4x4)(short *input, short *output, int pitch); +void (*vp8_fast_fdct8x4)(short *input, short *output, int pitch); +void (*short_walsh4x4)(short *input, short *output, int pitch); + +void (*vp8_subtract_b)(BLOCK *be, BLOCKD *bd, int pitch); +void (*vp8_subtract_mby)(short *diff, unsigned char *src, unsigned char *pred, int stride); +void (*vp8_subtract_mbuv)(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride); +void (*vp8_fast_quantize_b)(BLOCK *b, BLOCKD *d); + +unsigned int (*vp8_get16x16pred_error)(unsigned char *src_ptr, int src_stride, unsigned char *ref_ptr, int ref_stride); +unsigned int (*vp8_get8x8var)(unsigned char *src_ptr, int source_stride, unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); +unsigned int (*vp8_get16x16var)(unsigned char *src_ptr, int source_stride, unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); +unsigned int (*vp8_get4x4sse_cs)(unsigned char *src_ptr, int source_stride, unsigned char *ref_ptr, int recon_stride); + +// c imports +extern int block_error_c(short *coeff, short *dqcoeff); +extern int vp8_mbblock_error_c(MACROBLOCK *mb, int dc); + +extern int vp8_mbuverror_c(MACROBLOCK *mb); +extern unsigned int vp8_get8x8var_c(unsigned char *src_ptr, int source_stride, unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); +extern void short_fdct4x4_c(short *input, short *output, int pitch); +extern void short_fdct8x4_c(short *input, short *output, int pitch); +extern void vp8_short_walsh4x4_c(short *input, short *output, int pitch); + +extern void vp8_subtract_b_c(BLOCK *be, BLOCKD *bd, int pitch); +extern void subtract_mby_c(short *diff, unsigned char *src, unsigned char *pred, int stride); +extern void subtract_mbuv_c(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride); +extern void vp8_fast_quantize_b_c(BLOCK *b, BLOCKD *d); + +extern SADFunction sad16x16_c; +extern SADFunction sad16x8_c; +extern SADFunction sad8x16_c; +extern SADFunction sad8x8_c; +extern SADFunction sad4x4_c; + +extern variance_function variance16x16_c; +extern variance_function variance8x16_c; +extern variance_function variance16x8_c; +extern variance_function variance8x8_c; +extern variance_function variance4x4_c; +extern variance_function mse16x16_c; + +extern sub_pixel_variance_function sub_pixel_variance4x4_c; +extern sub_pixel_variance_function sub_pixel_variance8x8_c; +extern sub_pixel_variance_function sub_pixel_variance8x16_c; +extern sub_pixel_variance_function sub_pixel_variance16x8_c; +extern sub_pixel_variance_function sub_pixel_variance16x16_c; + +extern unsigned int vp8_get_mb_ss_c(short *); +extern unsigned int vp8_get16x16pred_error_c(unsigned char *src_ptr, int src_stride, unsigned char *ref_ptr, int ref_stride); +extern unsigned int vp8_get8x8var_c(unsigned char *src_ptr, int source_stride, unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); +extern unsigned int vp8_get16x16var_c(unsigned char *src_ptr, int source_stride, unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); +extern unsigned int vp8_get4x4sse_cs_c(unsigned char *src_ptr, int source_stride, unsigned char *ref_ptr, int recon_stride); + +// ppc +extern int vp8_block_error_ppc(short *coeff, short *dqcoeff); + +extern void vp8_short_fdct4x4_ppc(short *input, short *output, int pitch); +extern void vp8_short_fdct8x4_ppc(short *input, short *output, int pitch); + +extern void vp8_subtract_mby_ppc(short *diff, unsigned char *src, unsigned char *pred, int stride); +extern void vp8_subtract_mbuv_ppc(short *diff, unsigned char *usrc, unsigned char *vsrc, unsigned char *pred, int stride); + +extern SADFunction vp8_sad16x16_ppc; +extern SADFunction vp8_sad16x8_ppc; +extern SADFunction vp8_sad8x16_ppc; +extern SADFunction vp8_sad8x8_ppc; +extern SADFunction vp8_sad4x4_ppc; + +extern variance_function vp8_variance16x16_ppc; +extern variance_function vp8_variance8x16_ppc; +extern variance_function vp8_variance16x8_ppc; +extern variance_function vp8_variance8x8_ppc; +extern variance_function vp8_variance4x4_ppc; +extern variance_function vp8_mse16x16_ppc; + +extern sub_pixel_variance_function vp8_sub_pixel_variance4x4_ppc; +extern sub_pixel_variance_function vp8_sub_pixel_variance8x8_ppc; +extern sub_pixel_variance_function vp8_sub_pixel_variance8x16_ppc; +extern sub_pixel_variance_function vp8_sub_pixel_variance16x8_ppc; +extern sub_pixel_variance_function vp8_sub_pixel_variance16x16_ppc; + +extern unsigned int vp8_get8x8var_ppc(unsigned char *src_ptr, int source_stride, unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); +extern unsigned int vp8_get16x16var_ppc(unsigned char *src_ptr, int source_stride, unsigned char *ref_ptr, int recon_stride, unsigned int *SSE, int *Sum); + +void vp8_cmachine_specific_config(void) +{ + // Pure C: + vp8_mbuverror = vp8_mbuverror_c; + vp8_fast_quantize_b = vp8_fast_quantize_b_c; + vp8_short_fdct4x4 = vp8_short_fdct4x4_ppc; + vp8_short_fdct8x4 = vp8_short_fdct8x4_ppc; + vp8_fast_fdct4x4 = vp8_short_fdct4x4_ppc; + vp8_fast_fdct8x4 = vp8_short_fdct8x4_ppc; + short_walsh4x4 = vp8_short_walsh4x4_c; + + vp8_variance4x4 = vp8_variance4x4_ppc; + vp8_variance8x8 = vp8_variance8x8_ppc; + vp8_variance8x16 = vp8_variance8x16_ppc; + vp8_variance16x8 = vp8_variance16x8_ppc; + vp8_variance16x16 = vp8_variance16x16_ppc; + vp8_mse16x16 = vp8_mse16x16_ppc; + + vp8_sub_pixel_variance4x4 = vp8_sub_pixel_variance4x4_ppc; + vp8_sub_pixel_variance8x8 = vp8_sub_pixel_variance8x8_ppc; + vp8_sub_pixel_variance8x16 = vp8_sub_pixel_variance8x16_ppc; + vp8_sub_pixel_variance16x8 = vp8_sub_pixel_variance16x8_ppc; + vp8_sub_pixel_variance16x16 = vp8_sub_pixel_variance16x16_ppc; + + vp8_get_mb_ss = vp8_get_mb_ss_c; + vp8_get16x16pred_error = vp8_get16x16pred_error_c; + vp8_get8x8var = vp8_get8x8var_ppc; + vp8_get16x16var = vp8_get16x16var_ppc; + vp8_get4x4sse_cs = vp8_get4x4sse_cs_c; + + vp8_sad16x16 = vp8_sad16x16_ppc; + vp8_sad16x8 = vp8_sad16x8_ppc; + vp8_sad8x16 = vp8_sad8x16_ppc; + vp8_sad8x8 = vp8_sad8x8_ppc; + vp8_sad4x4 = vp8_sad4x4_ppc; + + vp8_block_error = vp8_block_error_ppc; + vp8_mbblock_error = vp8_mbblock_error_c; + + vp8_subtract_b = vp8_subtract_b_c; + vp8_subtract_mby = vp8_subtract_mby_ppc; + vp8_subtract_mbuv = vp8_subtract_mbuv_ppc; +} |