use asm_offsets with vp8_fast_quantize_b_sse3
on the same order as the sse2 fast quantize change: ~2%
except for 32bit. only a slight improvment there.
Change-Id: Iff80e5f1ce7e646eebfdc8871405458ff911986b
diff --git a/vp8/encoder/x86/x86_csystemdependent.c b/vp8/encoder/x86/x86_csystemdependent.c
index 8bceece..8f2774b 100644
--- a/vp8/encoder/x86/x86_csystemdependent.c
+++ b/vp8/encoder/x86/x86_csystemdependent.c
@@ -112,21 +112,6 @@
#endif
#if HAVE_SSSE3
-int vp8_fast_quantize_b_impl_ssse3(short *coeff_ptr,
- short *qcoeff_ptr, short *dequant_ptr,
- short *round_ptr,
- short *quant_ptr, short *dqcoeff_ptr);
-static void fast_quantize_b_ssse3(BLOCK *b, BLOCKD *d)
-{
- d->eob = vp8_fast_quantize_b_impl_ssse3(
- b->coeff,
- d->qcoeff,
- d->dequant,
- b->round,
- b->quant_fast,
- d->dqcoeff
- );
-}
#if CONFIG_PSNR
#if ARCH_X86_64
typedef void ssimpf
@@ -307,7 +292,7 @@
cpi->rtcd.variance.subpixvar16x8 = vp8_sub_pixel_variance16x8_ssse3;
cpi->rtcd.variance.subpixvar16x16 = vp8_sub_pixel_variance16x16_ssse3;
- cpi->rtcd.quantize.fastquantb = fast_quantize_b_ssse3;
+ cpi->rtcd.quantize.fastquantb = vp8_fast_quantize_b_ssse3;
#if CONFIG_PSNR
#if ARCH_X86_64