| /* |
| * Copyright 2011 The LibYuv Project Authors. All rights reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| |
| #include "libyuv/scale.h" |
| |
| #include <assert.h> |
| #include <string.h> |
| |
| #include "libyuv/cpu_id.h" |
| #include "libyuv/planar_functions.h" // For CopyARGB |
| #include "libyuv/row.h" |
| #include "libyuv/scale_row.h" |
| |
| #ifdef __cplusplus |
| namespace libyuv { |
| extern "C" { |
| #endif |
| |
| static __inline int Abs(int v) { |
| return v >= 0 ? v : -v; |
| } |
| |
| // ScaleARGB ARGB, 1/2 |
| // This is an optimized version for scaling down a ARGB to 1/2 of |
| // its original size. |
| static void ScaleARGBDown2(int src_width, int src_height, |
| int dst_width, int dst_height, |
| int src_stride, int dst_stride, |
| const uint8* src_argb, uint8* dst_argb, |
| int x, int dx, int y, int dy, |
| enum FilterMode filtering) { |
| int j; |
| int row_stride = src_stride * (dy >> 16); |
| void (*ScaleARGBRowDown2)(const uint8* src_argb, ptrdiff_t src_stride, |
| uint8* dst_argb, int dst_width) = |
| filtering == kFilterNone ? ScaleARGBRowDown2_C : |
| (filtering == kFilterLinear ? ScaleARGBRowDown2Linear_C : |
| ScaleARGBRowDown2Box_C); |
| assert(dx == 65536 * 2); // Test scale factor of 2. |
| assert((dy & 0x1ffff) == 0); // Test vertical scale is multiple of 2. |
| // Advance to odd row, even column. |
| if (filtering == kFilterBilinear) { |
| src_argb += (y >> 16) * src_stride + (x >> 16) * 4; |
| } else { |
| src_argb += (y >> 16) * src_stride + ((x >> 16) - 1) * 4; |
| } |
| |
| #if defined(HAS_SCALEARGBROWDOWN2_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2)) { |
| ScaleARGBRowDown2 = filtering == kFilterNone ? ScaleARGBRowDown2_Any_SSE2 : |
| (filtering == kFilterLinear ? ScaleARGBRowDown2Linear_Any_SSE2 : |
| ScaleARGBRowDown2Box_Any_SSE2); |
| if (IS_ALIGNED(dst_width, 4)) { |
| ScaleARGBRowDown2 = filtering == kFilterNone ? ScaleARGBRowDown2_SSE2 : |
| (filtering == kFilterLinear ? ScaleARGBRowDown2Linear_SSE2 : |
| ScaleARGBRowDown2Box_SSE2); |
| } |
| } |
| #endif |
| #if defined(HAS_SCALEARGBROWDOWN2_NEON) |
| if (TestCpuFlag(kCpuHasNEON)) { |
| ScaleARGBRowDown2 = filtering == kFilterNone ? ScaleARGBRowDown2_Any_NEON : |
| (filtering == kFilterLinear ? ScaleARGBRowDown2Linear_Any_NEON : |
| ScaleARGBRowDown2Box_Any_NEON); |
| if (IS_ALIGNED(dst_width, 8)) { |
| ScaleARGBRowDown2 = filtering == kFilterNone ? ScaleARGBRowDown2_NEON : |
| (filtering == kFilterLinear ? ScaleARGBRowDown2Linear_NEON : |
| ScaleARGBRowDown2Box_NEON); |
| } |
| } |
| #endif |
| |
| if (filtering == kFilterLinear) { |
| src_stride = 0; |
| } |
| for (j = 0; j < dst_height; ++j) { |
| ScaleARGBRowDown2(src_argb, src_stride, dst_argb, dst_width); |
| src_argb += row_stride; |
| dst_argb += dst_stride; |
| } |
| } |
| |
| // ScaleARGB ARGB, 1/4 |
| // This is an optimized version for scaling down a ARGB to 1/4 of |
| // its original size. |
| static void ScaleARGBDown4Box(int src_width, int src_height, |
| int dst_width, int dst_height, |
| int src_stride, int dst_stride, |
| const uint8* src_argb, uint8* dst_argb, |
| int x, int dx, int y, int dy) { |
| int j; |
| // Allocate 2 rows of ARGB. |
| const int kRowSize = (dst_width * 2 * 4 + 31) & ~31; |
| align_buffer_64(row, kRowSize * 2); |
| int row_stride = src_stride * (dy >> 16); |
| void (*ScaleARGBRowDown2)(const uint8* src_argb, ptrdiff_t src_stride, |
| uint8* dst_argb, int dst_width) = ScaleARGBRowDown2Box_C; |
| // Advance to odd row, even column. |
| src_argb += (y >> 16) * src_stride + (x >> 16) * 4; |
| assert(dx == 65536 * 4); // Test scale factor of 4. |
| assert((dy & 0x3ffff) == 0); // Test vertical scale is multiple of 4. |
| #if defined(HAS_SCALEARGBROWDOWN2_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2)) { |
| ScaleARGBRowDown2 = ScaleARGBRowDown2Box_Any_SSE2; |
| if (IS_ALIGNED(dst_width, 4)) { |
| ScaleARGBRowDown2 = ScaleARGBRowDown2Box_SSE2; |
| } |
| } |
| #endif |
| #if defined(HAS_SCALEARGBROWDOWN2_NEON) |
| if (TestCpuFlag(kCpuHasNEON)) { |
| ScaleARGBRowDown2 = ScaleARGBRowDown2Box_Any_NEON; |
| if (IS_ALIGNED(dst_width, 8)) { |
| ScaleARGBRowDown2 = ScaleARGBRowDown2Box_NEON; |
| } |
| } |
| #endif |
| |
| for (j = 0; j < dst_height; ++j) { |
| ScaleARGBRowDown2(src_argb, src_stride, row, dst_width * 2); |
| ScaleARGBRowDown2(src_argb + src_stride * 2, src_stride, |
| row + kRowSize, dst_width * 2); |
| ScaleARGBRowDown2(row, kRowSize, dst_argb, dst_width); |
| src_argb += row_stride; |
| dst_argb += dst_stride; |
| } |
| free_aligned_buffer_64(row); |
| } |
| |
| // ScaleARGB ARGB Even |
| // This is an optimized version for scaling down a ARGB to even |
| // multiple of its original size. |
| static void ScaleARGBDownEven(int src_width, int src_height, |
| int dst_width, int dst_height, |
| int src_stride, int dst_stride, |
| const uint8* src_argb, uint8* dst_argb, |
| int x, int dx, int y, int dy, |
| enum FilterMode filtering) { |
| int j; |
| int col_step = dx >> 16; |
| int row_stride = (dy >> 16) * src_stride; |
| void (*ScaleARGBRowDownEven)(const uint8* src_argb, ptrdiff_t src_stride, |
| int src_step, uint8* dst_argb, int dst_width) = |
| filtering ? ScaleARGBRowDownEvenBox_C : ScaleARGBRowDownEven_C; |
| assert(IS_ALIGNED(src_width, 2)); |
| assert(IS_ALIGNED(src_height, 2)); |
| src_argb += (y >> 16) * src_stride + (x >> 16) * 4; |
| #if defined(HAS_SCALEARGBROWDOWNEVEN_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2)) { |
| ScaleARGBRowDownEven = filtering ? ScaleARGBRowDownEvenBox_Any_SSE2 : |
| ScaleARGBRowDownEven_Any_SSE2; |
| if (IS_ALIGNED(dst_width, 4)) { |
| ScaleARGBRowDownEven = filtering ? ScaleARGBRowDownEvenBox_SSE2 : |
| ScaleARGBRowDownEven_SSE2; |
| } |
| } |
| #endif |
| #if defined(HAS_SCALEARGBROWDOWNEVEN_NEON) |
| if (TestCpuFlag(kCpuHasNEON)) { |
| ScaleARGBRowDownEven = filtering ? ScaleARGBRowDownEvenBox_Any_NEON : |
| ScaleARGBRowDownEven_Any_NEON; |
| if (IS_ALIGNED(dst_width, 4)) { |
| ScaleARGBRowDownEven = filtering ? ScaleARGBRowDownEvenBox_NEON : |
| ScaleARGBRowDownEven_NEON; |
| } |
| } |
| #endif |
| |
| if (filtering == kFilterLinear) { |
| src_stride = 0; |
| } |
| for (j = 0; j < dst_height; ++j) { |
| ScaleARGBRowDownEven(src_argb, src_stride, col_step, dst_argb, dst_width); |
| src_argb += row_stride; |
| dst_argb += dst_stride; |
| } |
| } |
| |
| // Scale ARGB down with bilinear interpolation. |
| static void ScaleARGBBilinearDown(int src_width, int src_height, |
| int dst_width, int dst_height, |
| int src_stride, int dst_stride, |
| const uint8* src_argb, uint8* dst_argb, |
| int x, int dx, int y, int dy, |
| enum FilterMode filtering) { |
| int j; |
| void (*InterpolateRow)(uint8* dst_argb, const uint8* src_argb, |
| ptrdiff_t src_stride, int dst_width, int source_y_fraction) = |
| InterpolateRow_C; |
| void (*ScaleARGBFilterCols)(uint8* dst_argb, const uint8* src_argb, |
| int dst_width, int x, int dx) = |
| (src_width >= 32768) ? ScaleARGBFilterCols64_C : ScaleARGBFilterCols_C; |
| int64 xlast = x + (int64)(dst_width - 1) * dx; |
| int64 xl = (dx >= 0) ? x : xlast; |
| int64 xr = (dx >= 0) ? xlast : x; |
| int clip_src_width; |
| xl = (xl >> 16) & ~3; // Left edge aligned. |
| xr = (xr >> 16) + 1; // Right most pixel used. Bilinear uses 2 pixels. |
| xr = (xr + 1 + 3) & ~3; // 1 beyond 4 pixel aligned right most pixel. |
| if (xr > src_width) { |
| xr = src_width; |
| } |
| clip_src_width = (int)(xr - xl) * 4; // Width aligned to 4. |
| src_argb += xl * 4; |
| x -= (int)(xl << 16); |
| #if defined(HAS_INTERPOLATEROW_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2)) { |
| InterpolateRow = InterpolateRow_Any_SSE2; |
| if (IS_ALIGNED(clip_src_width, 16)) { |
| InterpolateRow = InterpolateRow_SSE2; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_SSSE3) |
| if (TestCpuFlag(kCpuHasSSSE3)) { |
| InterpolateRow = InterpolateRow_Any_SSSE3; |
| if (IS_ALIGNED(clip_src_width, 16)) { |
| InterpolateRow = InterpolateRow_SSSE3; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_AVX2) |
| if (TestCpuFlag(kCpuHasAVX2)) { |
| InterpolateRow = InterpolateRow_Any_AVX2; |
| if (IS_ALIGNED(clip_src_width, 32)) { |
| InterpolateRow = InterpolateRow_AVX2; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_NEON) |
| if (TestCpuFlag(kCpuHasNEON)) { |
| InterpolateRow = InterpolateRow_Any_NEON; |
| if (IS_ALIGNED(clip_src_width, 16)) { |
| InterpolateRow = InterpolateRow_NEON; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_MIPS_DSPR2) |
| if (TestCpuFlag(kCpuHasMIPS_DSPR2) && |
| IS_ALIGNED(src_argb, 4) && IS_ALIGNED(src_stride, 4)) { |
| InterpolateRow = InterpolateRow_Any_MIPS_DSPR2; |
| if (IS_ALIGNED(clip_src_width, 4)) { |
| InterpolateRow = InterpolateRow_MIPS_DSPR2; |
| } |
| } |
| #endif |
| #if defined(HAS_SCALEARGBFILTERCOLS_SSSE3) |
| if (TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) { |
| ScaleARGBFilterCols = ScaleARGBFilterCols_SSSE3; |
| } |
| #endif |
| #if defined(HAS_SCALEARGBFILTERCOLS_NEON) |
| if (TestCpuFlag(kCpuHasNEON)) { |
| ScaleARGBFilterCols = ScaleARGBFilterCols_Any_NEON; |
| if (IS_ALIGNED(dst_width, 4)) { |
| ScaleARGBFilterCols = ScaleARGBFilterCols_NEON; |
| } |
| } |
| #endif |
| // TODO(fbarchard): Consider not allocating row buffer for kFilterLinear. |
| // Allocate a row of ARGB. |
| { |
| align_buffer_64(row, clip_src_width * 4); |
| |
| const int max_y = (src_height - 1) << 16; |
| if (y > max_y) { |
| y = max_y; |
| } |
| for (j = 0; j < dst_height; ++j) { |
| int yi = y >> 16; |
| const uint8* src = src_argb + yi * src_stride; |
| if (filtering == kFilterLinear) { |
| ScaleARGBFilterCols(dst_argb, src, dst_width, x, dx); |
| } else { |
| int yf = (y >> 8) & 255; |
| InterpolateRow(row, src, src_stride, clip_src_width, yf); |
| ScaleARGBFilterCols(dst_argb, row, dst_width, x, dx); |
| } |
| dst_argb += dst_stride; |
| y += dy; |
| if (y > max_y) { |
| y = max_y; |
| } |
| } |
| free_aligned_buffer_64(row); |
| } |
| } |
| |
| // Scale ARGB up with bilinear interpolation. |
| static void ScaleARGBBilinearUp(int src_width, int src_height, |
| int dst_width, int dst_height, |
| int src_stride, int dst_stride, |
| const uint8* src_argb, uint8* dst_argb, |
| int x, int dx, int y, int dy, |
| enum FilterMode filtering) { |
| int j; |
| void (*InterpolateRow)(uint8* dst_argb, const uint8* src_argb, |
| ptrdiff_t src_stride, int dst_width, int source_y_fraction) = |
| InterpolateRow_C; |
| void (*ScaleARGBFilterCols)(uint8* dst_argb, const uint8* src_argb, |
| int dst_width, int x, int dx) = |
| filtering ? ScaleARGBFilterCols_C : ScaleARGBCols_C; |
| const int max_y = (src_height - 1) << 16; |
| #if defined(HAS_INTERPOLATEROW_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2)) { |
| InterpolateRow = InterpolateRow_Any_SSE2; |
| if (IS_ALIGNED(dst_width, 4)) { |
| InterpolateRow = InterpolateRow_SSE2; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_SSSE3) |
| if (TestCpuFlag(kCpuHasSSSE3)) { |
| InterpolateRow = InterpolateRow_Any_SSSE3; |
| if (IS_ALIGNED(dst_width, 4)) { |
| InterpolateRow = InterpolateRow_SSSE3; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_AVX2) |
| if (TestCpuFlag(kCpuHasAVX2)) { |
| InterpolateRow = InterpolateRow_Any_AVX2; |
| if (IS_ALIGNED(dst_width, 8)) { |
| InterpolateRow = InterpolateRow_AVX2; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_NEON) |
| if (TestCpuFlag(kCpuHasNEON)) { |
| InterpolateRow = InterpolateRow_Any_NEON; |
| if (IS_ALIGNED(dst_width, 4)) { |
| InterpolateRow = InterpolateRow_NEON; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_MIPS_DSPR2) |
| if (TestCpuFlag(kCpuHasMIPS_DSPR2) && |
| IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride, 4)) { |
| InterpolateRow = InterpolateRow_MIPS_DSPR2; |
| } |
| #endif |
| if (src_width >= 32768) { |
| ScaleARGBFilterCols = filtering ? |
| ScaleARGBFilterCols64_C : ScaleARGBCols64_C; |
| } |
| #if defined(HAS_SCALEARGBFILTERCOLS_SSSE3) |
| if (filtering && TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) { |
| ScaleARGBFilterCols = ScaleARGBFilterCols_SSSE3; |
| } |
| #endif |
| #if defined(HAS_SCALEARGBFILTERCOLS_NEON) |
| if (filtering && TestCpuFlag(kCpuHasNEON)) { |
| ScaleARGBFilterCols = ScaleARGBFilterCols_Any_NEON; |
| if (IS_ALIGNED(dst_width, 4)) { |
| ScaleARGBFilterCols = ScaleARGBFilterCols_NEON; |
| } |
| } |
| #endif |
| #if defined(HAS_SCALEARGBCOLS_SSE2) |
| if (!filtering && TestCpuFlag(kCpuHasSSE2) && src_width < 32768) { |
| ScaleARGBFilterCols = ScaleARGBCols_SSE2; |
| } |
| #endif |
| #if defined(HAS_SCALEARGBCOLS_NEON) |
| if (!filtering && TestCpuFlag(kCpuHasNEON)) { |
| ScaleARGBFilterCols = ScaleARGBCols_Any_NEON; |
| if (IS_ALIGNED(dst_width, 8)) { |
| ScaleARGBFilterCols = ScaleARGBCols_NEON; |
| } |
| } |
| #endif |
| if (!filtering && src_width * 2 == dst_width && x < 0x8000) { |
| ScaleARGBFilterCols = ScaleARGBColsUp2_C; |
| #if defined(HAS_SCALEARGBCOLSUP2_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 8)) { |
| ScaleARGBFilterCols = ScaleARGBColsUp2_SSE2; |
| } |
| #endif |
| } |
| |
| if (y > max_y) { |
| y = max_y; |
| } |
| |
| { |
| int yi = y >> 16; |
| const uint8* src = src_argb + yi * src_stride; |
| |
| // Allocate 2 rows of ARGB. |
| const int kRowSize = (dst_width * 4 + 31) & ~31; |
| align_buffer_64(row, kRowSize * 2); |
| |
| uint8* rowptr = row; |
| int rowstride = kRowSize; |
| int lasty = yi; |
| |
| ScaleARGBFilterCols(rowptr, src, dst_width, x, dx); |
| if (src_height > 1) { |
| src += src_stride; |
| } |
| ScaleARGBFilterCols(rowptr + rowstride, src, dst_width, x, dx); |
| src += src_stride; |
| |
| for (j = 0; j < dst_height; ++j) { |
| yi = y >> 16; |
| if (yi != lasty) { |
| if (y > max_y) { |
| y = max_y; |
| yi = y >> 16; |
| src = src_argb + yi * src_stride; |
| } |
| if (yi != lasty) { |
| ScaleARGBFilterCols(rowptr, src, dst_width, x, dx); |
| rowptr += rowstride; |
| rowstride = -rowstride; |
| lasty = yi; |
| src += src_stride; |
| } |
| } |
| if (filtering == kFilterLinear) { |
| InterpolateRow(dst_argb, rowptr, 0, dst_width * 4, 0); |
| } else { |
| int yf = (y >> 8) & 255; |
| InterpolateRow(dst_argb, rowptr, rowstride, dst_width * 4, yf); |
| } |
| dst_argb += dst_stride; |
| y += dy; |
| } |
| free_aligned_buffer_64(row); |
| } |
| } |
| |
| #ifdef YUVSCALEUP |
| // Scale YUV to ARGB up with bilinear interpolation. |
| static void ScaleYUVToARGBBilinearUp(int src_width, int src_height, |
| int dst_width, int dst_height, |
| int src_stride_y, |
| int src_stride_u, |
| int src_stride_v, |
| int dst_stride_argb, |
| const uint8* src_y, |
| const uint8* src_u, |
| const uint8* src_v, |
| uint8* dst_argb, |
| int x, int dx, int y, int dy, |
| enum FilterMode filtering) { |
| int j; |
| void (*I422ToARGBRow)(const uint8* y_buf, |
| const uint8* u_buf, |
| const uint8* v_buf, |
| uint8* rgb_buf, |
| int width) = I422ToARGBRow_C; |
| #if defined(HAS_I422TOARGBROW_SSSE3) |
| if (TestCpuFlag(kCpuHasSSSE3)) { |
| I422ToARGBRow = I422ToARGBRow_Any_SSSE3; |
| if (IS_ALIGNED(src_width, 8)) { |
| I422ToARGBRow = I422ToARGBRow_SSSE3; |
| } |
| } |
| #endif |
| #if defined(HAS_I422TOARGBROW_AVX2) |
| if (TestCpuFlag(kCpuHasAVX2)) { |
| I422ToARGBRow = I422ToARGBRow_Any_AVX2; |
| if (IS_ALIGNED(src_width, 16)) { |
| I422ToARGBRow = I422ToARGBRow_AVX2; |
| } |
| } |
| #endif |
| #if defined(HAS_I422TOARGBROW_NEON) |
| if (TestCpuFlag(kCpuHasNEON)) { |
| I422ToARGBRow = I422ToARGBRow_Any_NEON; |
| if (IS_ALIGNED(src_width, 8)) { |
| I422ToARGBRow = I422ToARGBRow_NEON; |
| } |
| } |
| #endif |
| #if defined(HAS_I422TOARGBROW_MIPS_DSPR2) |
| if (TestCpuFlag(kCpuHasMIPS_DSPR2) && IS_ALIGNED(src_width, 4) && |
| IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) && |
| IS_ALIGNED(src_u, 2) && IS_ALIGNED(src_stride_u, 2) && |
| IS_ALIGNED(src_v, 2) && IS_ALIGNED(src_stride_v, 2) && |
| IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride_argb, 4)) { |
| I422ToARGBRow = I422ToARGBRow_MIPS_DSPR2; |
| } |
| #endif |
| |
| void (*InterpolateRow)(uint8* dst_argb, const uint8* src_argb, |
| ptrdiff_t src_stride, int dst_width, int source_y_fraction) = |
| InterpolateRow_C; |
| #if defined(HAS_INTERPOLATEROW_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2)) { |
| InterpolateRow = InterpolateRow_Any_SSE2; |
| if (IS_ALIGNED(dst_width, 4)) { |
| InterpolateRow = InterpolateRow_SSE2; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_SSSE3) |
| if (TestCpuFlag(kCpuHasSSSE3)) { |
| InterpolateRow = InterpolateRow_Any_SSSE3; |
| if (IS_ALIGNED(dst_width, 4)) { |
| InterpolateRow = InterpolateRow_SSSE3; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_AVX2) |
| if (TestCpuFlag(kCpuHasAVX2)) { |
| InterpolateRow = InterpolateRow_Any_AVX2; |
| if (IS_ALIGNED(dst_width, 8)) { |
| InterpolateRow = InterpolateRow_AVX2; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_NEON) |
| if (TestCpuFlag(kCpuHasNEON)) { |
| InterpolateRow = InterpolateRow_Any_NEON; |
| if (IS_ALIGNED(dst_width, 4)) { |
| InterpolateRow = InterpolateRow_NEON; |
| } |
| } |
| #endif |
| #if defined(HAS_INTERPOLATEROW_MIPS_DSPR2) |
| if (TestCpuFlag(kCpuHasMIPS_DSPR2) && |
| IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride_argb, 4)) { |
| InterpolateRow = InterpolateRow_MIPS_DSPR2; |
| } |
| #endif |
| |
| void (*ScaleARGBFilterCols)(uint8* dst_argb, const uint8* src_argb, |
| int dst_width, int x, int dx) = |
| filtering ? ScaleARGBFilterCols_C : ScaleARGBCols_C; |
| if (src_width >= 32768) { |
| ScaleARGBFilterCols = filtering ? |
| ScaleARGBFilterCols64_C : ScaleARGBCols64_C; |
| } |
| #if defined(HAS_SCALEARGBFILTERCOLS_SSSE3) |
| if (filtering && TestCpuFlag(kCpuHasSSSE3) && src_width < 32768) { |
| ScaleARGBFilterCols = ScaleARGBFilterCols_SSSE3; |
| } |
| #endif |
| #if defined(HAS_SCALEARGBFILTERCOLS_NEON) |
| if (filtering && TestCpuFlag(kCpuHasNEON)) { |
| ScaleARGBFilterCols = ScaleARGBFilterCols_Any_NEON; |
| if (IS_ALIGNED(dst_width, 4)) { |
| ScaleARGBFilterCols = ScaleARGBFilterCols_NEON; |
| } |
| } |
| #endif |
| #if defined(HAS_SCALEARGBCOLS_SSE2) |
| if (!filtering && TestCpuFlag(kCpuHasSSE2) && src_width < 32768) { |
| ScaleARGBFilterCols = ScaleARGBCols_SSE2; |
| } |
| #endif |
| #if defined(HAS_SCALEARGBCOLS_NEON) |
| if (!filtering && TestCpuFlag(kCpuHasNEON)) { |
| ScaleARGBFilterCols = ScaleARGBCols_Any_NEON; |
| if (IS_ALIGNED(dst_width, 8)) { |
| ScaleARGBFilterCols = ScaleARGBCols_NEON; |
| } |
| } |
| #endif |
| if (!filtering && src_width * 2 == dst_width && x < 0x8000) { |
| ScaleARGBFilterCols = ScaleARGBColsUp2_C; |
| #if defined(HAS_SCALEARGBCOLSUP2_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 8)) { |
| ScaleARGBFilterCols = ScaleARGBColsUp2_SSE2; |
| } |
| #endif |
| } |
| |
| const int max_y = (src_height - 1) << 16; |
| if (y > max_y) { |
| y = max_y; |
| } |
| const int kYShift = 1; // Shift Y by 1 to convert Y plane to UV coordinate. |
| int yi = y >> 16; |
| int uv_yi = yi >> kYShift; |
| const uint8* src_row_y = src_y + yi * src_stride_y; |
| const uint8* src_row_u = src_u + uv_yi * src_stride_u; |
| const uint8* src_row_v = src_v + uv_yi * src_stride_v; |
| |
| // Allocate 2 rows of ARGB. |
| const int kRowSize = (dst_width * 4 + 31) & ~31; |
| align_buffer_64(row, kRowSize * 2); |
| |
| // Allocate 1 row of ARGB for source conversion. |
| align_buffer_64(argb_row, src_width * 4); |
| |
| uint8* rowptr = row; |
| int rowstride = kRowSize; |
| int lasty = yi; |
| |
| // TODO(fbarchard): Convert first 2 rows of YUV to ARGB. |
| ScaleARGBFilterCols(rowptr, src_row_y, dst_width, x, dx); |
| if (src_height > 1) { |
| src_row_y += src_stride_y; |
| if (yi & 1) { |
| src_row_u += src_stride_u; |
| src_row_v += src_stride_v; |
| } |
| } |
| ScaleARGBFilterCols(rowptr + rowstride, src_row_y, dst_width, x, dx); |
| if (src_height > 2) { |
| src_row_y += src_stride_y; |
| if (!(yi & 1)) { |
| src_row_u += src_stride_u; |
| src_row_v += src_stride_v; |
| } |
| } |
| |
| for (j = 0; j < dst_height; ++j) { |
| yi = y >> 16; |
| if (yi != lasty) { |
| if (y > max_y) { |
| y = max_y; |
| yi = y >> 16; |
| uv_yi = yi >> kYShift; |
| src_row_y = src_y + yi * src_stride_y; |
| src_row_u = src_u + uv_yi * src_stride_u; |
| src_row_v = src_v + uv_yi * src_stride_v; |
| } |
| if (yi != lasty) { |
| // TODO(fbarchard): Convert the clipped region of row. |
| I422ToARGBRow(src_row_y, src_row_u, src_row_v, argb_row, src_width); |
| ScaleARGBFilterCols(rowptr, argb_row, dst_width, x, dx); |
| rowptr += rowstride; |
| rowstride = -rowstride; |
| lasty = yi; |
| src_row_y += src_stride_y; |
| if (yi & 1) { |
| src_row_u += src_stride_u; |
| src_row_v += src_stride_v; |
| } |
| } |
| } |
| if (filtering == kFilterLinear) { |
| InterpolateRow(dst_argb, rowptr, 0, dst_width * 4, 0); |
| } else { |
| int yf = (y >> 8) & 255; |
| InterpolateRow(dst_argb, rowptr, rowstride, dst_width * 4, yf); |
| } |
| dst_argb += dst_stride_argb; |
| y += dy; |
| } |
| free_aligned_buffer_64(row); |
| free_aligned_buffer_64(row_argb); |
| } |
| #endif |
| |
| // Scale ARGB to/from any dimensions, without interpolation. |
| // Fixed point math is used for performance: The upper 16 bits |
| // of x and dx is the integer part of the source position and |
| // the lower 16 bits are the fixed decimal part. |
| |
| static void ScaleARGBSimple(int src_width, int src_height, |
| int dst_width, int dst_height, |
| int src_stride, int dst_stride, |
| const uint8* src_argb, uint8* dst_argb, |
| int x, int dx, int y, int dy) { |
| int j; |
| void (*ScaleARGBCols)(uint8* dst_argb, const uint8* src_argb, |
| int dst_width, int x, int dx) = |
| (src_width >= 32768) ? ScaleARGBCols64_C : ScaleARGBCols_C; |
| #if defined(HAS_SCALEARGBCOLS_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2) && src_width < 32768) { |
| ScaleARGBCols = ScaleARGBCols_SSE2; |
| } |
| #endif |
| #if defined(HAS_SCALEARGBCOLS_NEON) |
| if (TestCpuFlag(kCpuHasNEON)) { |
| ScaleARGBCols = ScaleARGBCols_Any_NEON; |
| if (IS_ALIGNED(dst_width, 8)) { |
| ScaleARGBCols = ScaleARGBCols_NEON; |
| } |
| } |
| #endif |
| if (src_width * 2 == dst_width && x < 0x8000) { |
| ScaleARGBCols = ScaleARGBColsUp2_C; |
| #if defined(HAS_SCALEARGBCOLSUP2_SSE2) |
| if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 8)) { |
| ScaleARGBCols = ScaleARGBColsUp2_SSE2; |
| } |
| #endif |
| } |
| |
| for (j = 0; j < dst_height; ++j) { |
| ScaleARGBCols(dst_argb, src_argb + (y >> 16) * src_stride, |
| dst_width, x, dx); |
| dst_argb += dst_stride; |
| y += dy; |
| } |
| } |
| |
| // ScaleARGB a ARGB. |
| // This function in turn calls a scaling function |
| // suitable for handling the desired resolutions. |
| static void ScaleARGB(const uint8* src, int src_stride, |
| int src_width, int src_height, |
| uint8* dst, int dst_stride, |
| int dst_width, int dst_height, |
| int clip_x, int clip_y, int clip_width, int clip_height, |
| enum FilterMode filtering) { |
| // Initial source x/y coordinate and step values as 16.16 fixed point. |
| int x = 0; |
| int y = 0; |
| int dx = 0; |
| int dy = 0; |
| // ARGB does not support box filter yet, but allow the user to pass it. |
| // Simplify filtering when possible. |
| filtering = ScaleFilterReduce(src_width, src_height, |
| dst_width, dst_height, |
| filtering); |
| |
| // Negative src_height means invert the image. |
| if (src_height < 0) { |
| src_height = -src_height; |
| src = src + (src_height - 1) * src_stride; |
| src_stride = -src_stride; |
| } |
| ScaleSlope(src_width, src_height, dst_width, dst_height, filtering, |
| &x, &y, &dx, &dy); |
| src_width = Abs(src_width); |
| if (clip_x) { |
| int64 clipf = (int64)(clip_x) * dx; |
| x += (clipf & 0xffff); |
| src += (clipf >> 16) * 4; |
| dst += clip_x * 4; |
| } |
| if (clip_y) { |
| int64 clipf = (int64)(clip_y) * dy; |
| y += (clipf & 0xffff); |
| src += (clipf >> 16) * src_stride; |
| dst += clip_y * dst_stride; |
| } |
| |
| // Special case for integer step values. |
| if (((dx | dy) & 0xffff) == 0) { |
| if (!dx || !dy) { // 1 pixel wide and/or tall. |
| filtering = kFilterNone; |
| } else { |
| // Optimized even scale down. ie 2, 4, 6, 8, 10x. |
| if (!(dx & 0x10000) && !(dy & 0x10000)) { |
| if (dx == 0x20000) { |
| // Optimized 1/2 downsample. |
| ScaleARGBDown2(src_width, src_height, |
| clip_width, clip_height, |
| src_stride, dst_stride, src, dst, |
| x, dx, y, dy, filtering); |
| return; |
| } |
| if (dx == 0x40000 && filtering == kFilterBox) { |
| // Optimized 1/4 box downsample. |
| ScaleARGBDown4Box(src_width, src_height, |
| clip_width, clip_height, |
| src_stride, dst_stride, src, dst, |
| x, dx, y, dy); |
| return; |
| } |
| ScaleARGBDownEven(src_width, src_height, |
| clip_width, clip_height, |
| src_stride, dst_stride, src, dst, |
| x, dx, y, dy, filtering); |
| return; |
| } |
| // Optimized odd scale down. ie 3, 5, 7, 9x. |
| if ((dx & 0x10000) && (dy & 0x10000)) { |
| filtering = kFilterNone; |
| if (dx == 0x10000 && dy == 0x10000) { |
| // Straight copy. |
| ARGBCopy(src + (y >> 16) * src_stride + (x >> 16) * 4, src_stride, |
| dst, dst_stride, clip_width, clip_height); |
| return; |
| } |
| } |
| } |
| } |
| if (dx == 0x10000 && (x & 0xffff) == 0) { |
| // Arbitrary scale vertically, but unscaled vertically. |
| ScalePlaneVertical(src_height, |
| clip_width, clip_height, |
| src_stride, dst_stride, src, dst, |
| x, y, dy, 4, filtering); |
| return; |
| } |
| if (filtering && dy < 65536) { |
| ScaleARGBBilinearUp(src_width, src_height, |
| clip_width, clip_height, |
| src_stride, dst_stride, src, dst, |
| x, dx, y, dy, filtering); |
| return; |
| } |
| if (filtering) { |
| ScaleARGBBilinearDown(src_width, src_height, |
| clip_width, clip_height, |
| src_stride, dst_stride, src, dst, |
| x, dx, y, dy, filtering); |
| return; |
| } |
| ScaleARGBSimple(src_width, src_height, clip_width, clip_height, |
| src_stride, dst_stride, src, dst, |
| x, dx, y, dy); |
| } |
| |
| LIBYUV_API |
| int ARGBScaleClip(const uint8* src_argb, int src_stride_argb, |
| int src_width, int src_height, |
| uint8* dst_argb, int dst_stride_argb, |
| int dst_width, int dst_height, |
| int clip_x, int clip_y, int clip_width, int clip_height, |
| enum FilterMode filtering) { |
| if (!src_argb || src_width == 0 || src_height == 0 || |
| !dst_argb || dst_width <= 0 || dst_height <= 0 || |
| clip_x < 0 || clip_y < 0 || |
| clip_width > 32768 || clip_height > 32768 || |
| (clip_x + clip_width) > dst_width || |
| (clip_y + clip_height) > dst_height) { |
| return -1; |
| } |
| ScaleARGB(src_argb, src_stride_argb, src_width, src_height, |
| dst_argb, dst_stride_argb, dst_width, dst_height, |
| clip_x, clip_y, clip_width, clip_height, filtering); |
| return 0; |
| } |
| |
| // Scale an ARGB image. |
| LIBYUV_API |
| int ARGBScale(const uint8* src_argb, int src_stride_argb, |
| int src_width, int src_height, |
| uint8* dst_argb, int dst_stride_argb, |
| int dst_width, int dst_height, |
| enum FilterMode filtering) { |
| if (!src_argb || src_width == 0 || src_height == 0 || |
| src_width > 32768 || src_height > 32768 || |
| !dst_argb || dst_width <= 0 || dst_height <= 0) { |
| return -1; |
| } |
| ScaleARGB(src_argb, src_stride_argb, src_width, src_height, |
| dst_argb, dst_stride_argb, dst_width, dst_height, |
| 0, 0, dst_width, dst_height, filtering); |
| return 0; |
| } |
| |
| #ifdef __cplusplus |
| } // extern "C" |
| } // namespace libyuv |
| #endif |