blob: e8eaec7a940e2f9bfa298be3f8ab17fa07e1fa27 [file] [log] [blame]
Yaowu Xuc27fc142016-08-22 16:08:15 -07001/*
Yaowu Xu9c01aa12016-09-01 14:32:49 -07002 * Copyright (c) 2016, Alliance for Open Media. All rights reserved
Yaowu Xuc27fc142016-08-22 16:08:15 -07003 *
Yaowu Xu9c01aa12016-09-01 14:32:49 -07004 * This source code is subject to the terms of the BSD 2 Clause License and
5 * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License
6 * was not distributed with this source code in the LICENSE file, you can
7 * obtain it at www.aomedia.org/license/software. If the Alliance for Open
8 * Media Patent License 1.0 was not distributed with this source code in the
9 * PATENTS file, you can obtain it at www.aomedia.org/license/patent.
Yaowu Xuc27fc142016-08-22 16:08:15 -070010 */
11
Yaowu Xuf883b422016-08-30 14:01:10 -070012#include "./aom_dsp_rtcd.h"
Yaowu Xuc27fc142016-08-22 16:08:15 -070013#include "aom_dsp/mips/macros_msa.h"
14
15#define IPRED_SUBS_UH2_UH(in0, in1, out0, out1) \
16 { \
17 out0 = __msa_subs_u_h(out0, in0); \
18 out1 = __msa_subs_u_h(out1, in1); \
19 }
20
21static void intra_predict_vert_4x4_msa(const uint8_t *src, uint8_t *dst,
22 int32_t dst_stride) {
23 uint32_t src_data;
24
25 src_data = LW(src);
26
27 SW4(src_data, src_data, src_data, src_data, dst, dst_stride);
28}
29
30static void intra_predict_vert_8x8_msa(const uint8_t *src, uint8_t *dst,
31 int32_t dst_stride) {
32 uint32_t row;
33 uint32_t src_data1, src_data2;
34
35 src_data1 = LW(src);
36 src_data2 = LW(src + 4);
37
38 for (row = 8; row--;) {
39 SW(src_data1, dst);
40 SW(src_data2, (dst + 4));
41 dst += dst_stride;
42 }
43}
44
45static void intra_predict_vert_16x16_msa(const uint8_t *src, uint8_t *dst,
46 int32_t dst_stride) {
47 uint32_t row;
48 v16u8 src0;
49
50 src0 = LD_UB(src);
51
52 for (row = 16; row--;) {
53 ST_UB(src0, dst);
54 dst += dst_stride;
55 }
56}
57
58static void intra_predict_vert_32x32_msa(const uint8_t *src, uint8_t *dst,
59 int32_t dst_stride) {
60 uint32_t row;
61 v16u8 src1, src2;
62
63 src1 = LD_UB(src);
64 src2 = LD_UB(src + 16);
65
66 for (row = 32; row--;) {
67 ST_UB2(src1, src2, dst, 16);
68 dst += dst_stride;
69 }
70}
71
72static void intra_predict_horiz_4x4_msa(const uint8_t *src, uint8_t *dst,
73 int32_t dst_stride) {
74 uint32_t out0, out1, out2, out3;
75
76 out0 = src[0] * 0x01010101;
77 out1 = src[1] * 0x01010101;
78 out2 = src[2] * 0x01010101;
79 out3 = src[3] * 0x01010101;
80
81 SW4(out0, out1, out2, out3, dst, dst_stride);
82}
83
84static void intra_predict_horiz_8x8_msa(const uint8_t *src, uint8_t *dst,
85 int32_t dst_stride) {
86 uint64_t out0, out1, out2, out3, out4, out5, out6, out7;
87
88 out0 = src[0] * 0x0101010101010101ull;
89 out1 = src[1] * 0x0101010101010101ull;
90 out2 = src[2] * 0x0101010101010101ull;
91 out3 = src[3] * 0x0101010101010101ull;
92 out4 = src[4] * 0x0101010101010101ull;
93 out5 = src[5] * 0x0101010101010101ull;
94 out6 = src[6] * 0x0101010101010101ull;
95 out7 = src[7] * 0x0101010101010101ull;
96
97 SD4(out0, out1, out2, out3, dst, dst_stride);
98 dst += (4 * dst_stride);
99 SD4(out4, out5, out6, out7, dst, dst_stride);
100}
101
102static void intra_predict_horiz_16x16_msa(const uint8_t *src, uint8_t *dst,
103 int32_t dst_stride) {
104 uint32_t row;
105 uint8_t inp0, inp1, inp2, inp3;
106 v16u8 src0, src1, src2, src3;
107
108 for (row = 4; row--;) {
109 inp0 = src[0];
110 inp1 = src[1];
111 inp2 = src[2];
112 inp3 = src[3];
113 src += 4;
114
115 src0 = (v16u8)__msa_fill_b(inp0);
116 src1 = (v16u8)__msa_fill_b(inp1);
117 src2 = (v16u8)__msa_fill_b(inp2);
118 src3 = (v16u8)__msa_fill_b(inp3);
119
120 ST_UB4(src0, src1, src2, src3, dst, dst_stride);
121 dst += (4 * dst_stride);
122 }
123}
124
125static void intra_predict_horiz_32x32_msa(const uint8_t *src, uint8_t *dst,
126 int32_t dst_stride) {
127 uint32_t row;
128 uint8_t inp0, inp1, inp2, inp3;
129 v16u8 src0, src1, src2, src3;
130
131 for (row = 8; row--;) {
132 inp0 = src[0];
133 inp1 = src[1];
134 inp2 = src[2];
135 inp3 = src[3];
136 src += 4;
137
138 src0 = (v16u8)__msa_fill_b(inp0);
139 src1 = (v16u8)__msa_fill_b(inp1);
140 src2 = (v16u8)__msa_fill_b(inp2);
141 src3 = (v16u8)__msa_fill_b(inp3);
142
143 ST_UB2(src0, src0, dst, 16);
144 dst += dst_stride;
145 ST_UB2(src1, src1, dst, 16);
146 dst += dst_stride;
147 ST_UB2(src2, src2, dst, 16);
148 dst += dst_stride;
149 ST_UB2(src3, src3, dst, 16);
150 dst += dst_stride;
151 }
152}
153
154static void intra_predict_dc_4x4_msa(const uint8_t *src_top,
155 const uint8_t *src_left, uint8_t *dst,
156 int32_t dst_stride) {
157 uint32_t val0, val1;
158 v16i8 store, src = { 0 };
159 v8u16 sum_h;
160 v4u32 sum_w;
161 v2u64 sum_d;
162
163 val0 = LW(src_top);
164 val1 = LW(src_left);
165 INSERT_W2_SB(val0, val1, src);
166 sum_h = __msa_hadd_u_h((v16u8)src, (v16u8)src);
167 sum_w = __msa_hadd_u_w(sum_h, sum_h);
168 sum_d = __msa_hadd_u_d(sum_w, sum_w);
169 sum_w = (v4u32)__msa_srari_w((v4i32)sum_d, 3);
170 store = __msa_splati_b((v16i8)sum_w, 0);
171 val0 = __msa_copy_u_w((v4i32)store, 0);
172
173 SW4(val0, val0, val0, val0, dst, dst_stride);
174}
175
176static void intra_predict_dc_tl_4x4_msa(const uint8_t *src, uint8_t *dst,
177 int32_t dst_stride) {
178 uint32_t val0;
179 v16i8 store, data = { 0 };
180 v8u16 sum_h;
181 v4u32 sum_w;
182
183 val0 = LW(src);
184 data = (v16i8)__msa_insert_w((v4i32)data, 0, val0);
185 sum_h = __msa_hadd_u_h((v16u8)data, (v16u8)data);
186 sum_w = __msa_hadd_u_w(sum_h, sum_h);
187 sum_w = (v4u32)__msa_srari_w((v4i32)sum_w, 2);
188 store = __msa_splati_b((v16i8)sum_w, 0);
189 val0 = __msa_copy_u_w((v4i32)store, 0);
190
191 SW4(val0, val0, val0, val0, dst, dst_stride);
192}
193
194static void intra_predict_128dc_4x4_msa(uint8_t *dst, int32_t dst_stride) {
195 uint32_t out;
196 const v16i8 store = __msa_ldi_b(128);
197
198 out = __msa_copy_u_w((v4i32)store, 0);
199
200 SW4(out, out, out, out, dst, dst_stride);
201}
202
203static void intra_predict_dc_8x8_msa(const uint8_t *src_top,
204 const uint8_t *src_left, uint8_t *dst,
205 int32_t dst_stride) {
206 uint64_t val0, val1;
207 v16i8 store;
208 v16u8 src = { 0 };
209 v8u16 sum_h;
210 v4u32 sum_w;
211 v2u64 sum_d;
212
213 val0 = LD(src_top);
214 val1 = LD(src_left);
215 INSERT_D2_UB(val0, val1, src);
216 sum_h = __msa_hadd_u_h(src, src);
217 sum_w = __msa_hadd_u_w(sum_h, sum_h);
218 sum_d = __msa_hadd_u_d(sum_w, sum_w);
219 sum_w = (v4u32)__msa_pckev_w((v4i32)sum_d, (v4i32)sum_d);
220 sum_d = __msa_hadd_u_d(sum_w, sum_w);
221 sum_w = (v4u32)__msa_srari_w((v4i32)sum_d, 4);
222 store = __msa_splati_b((v16i8)sum_w, 0);
223 val0 = __msa_copy_u_d((v2i64)store, 0);
224
225 SD4(val0, val0, val0, val0, dst, dst_stride);
226 dst += (4 * dst_stride);
227 SD4(val0, val0, val0, val0, dst, dst_stride);
228}
229
230static void intra_predict_dc_tl_8x8_msa(const uint8_t *src, uint8_t *dst,
231 int32_t dst_stride) {
232 uint64_t val0;
233 v16i8 store;
234 v16u8 data = { 0 };
235 v8u16 sum_h;
236 v4u32 sum_w;
237 v2u64 sum_d;
238
239 val0 = LD(src);
240 data = (v16u8)__msa_insert_d((v2i64)data, 0, val0);
241 sum_h = __msa_hadd_u_h(data, data);
242 sum_w = __msa_hadd_u_w(sum_h, sum_h);
243 sum_d = __msa_hadd_u_d(sum_w, sum_w);
244 sum_w = (v4u32)__msa_srari_w((v4i32)sum_d, 3);
245 store = __msa_splati_b((v16i8)sum_w, 0);
246 val0 = __msa_copy_u_d((v2i64)store, 0);
247
248 SD4(val0, val0, val0, val0, dst, dst_stride);
249 dst += (4 * dst_stride);
250 SD4(val0, val0, val0, val0, dst, dst_stride);
251}
252
253static void intra_predict_128dc_8x8_msa(uint8_t *dst, int32_t dst_stride) {
254 uint64_t out;
255 const v16i8 store = __msa_ldi_b(128);
256
257 out = __msa_copy_u_d((v2i64)store, 0);
258
259 SD4(out, out, out, out, dst, dst_stride);
260 dst += (4 * dst_stride);
261 SD4(out, out, out, out, dst, dst_stride);
262}
263
264static void intra_predict_dc_16x16_msa(const uint8_t *src_top,
265 const uint8_t *src_left, uint8_t *dst,
266 int32_t dst_stride) {
267 v16u8 top, left, out;
268 v8u16 sum_h, sum_top, sum_left;
269 v4u32 sum_w;
270 v2u64 sum_d;
271
272 top = LD_UB(src_top);
273 left = LD_UB(src_left);
274 HADD_UB2_UH(top, left, sum_top, sum_left);
275 sum_h = sum_top + sum_left;
276 sum_w = __msa_hadd_u_w(sum_h, sum_h);
277 sum_d = __msa_hadd_u_d(sum_w, sum_w);
278 sum_w = (v4u32)__msa_pckev_w((v4i32)sum_d, (v4i32)sum_d);
279 sum_d = __msa_hadd_u_d(sum_w, sum_w);
280 sum_w = (v4u32)__msa_srari_w((v4i32)sum_d, 5);
281 out = (v16u8)__msa_splati_b((v16i8)sum_w, 0);
282
283 ST_UB8(out, out, out, out, out, out, out, out, dst, dst_stride);
284 dst += (8 * dst_stride);
285 ST_UB8(out, out, out, out, out, out, out, out, dst, dst_stride);
286}
287
288static void intra_predict_dc_tl_16x16_msa(const uint8_t *src, uint8_t *dst,
289 int32_t dst_stride) {
290 v16u8 data, out;
291 v8u16 sum_h;
292 v4u32 sum_w;
293 v2u64 sum_d;
294
295 data = LD_UB(src);
296 sum_h = __msa_hadd_u_h(data, data);
297 sum_w = __msa_hadd_u_w(sum_h, sum_h);
298 sum_d = __msa_hadd_u_d(sum_w, sum_w);
299 sum_w = (v4u32)__msa_pckev_w((v4i32)sum_d, (v4i32)sum_d);
300 sum_d = __msa_hadd_u_d(sum_w, sum_w);
301 sum_w = (v4u32)__msa_srari_w((v4i32)sum_d, 4);
302 out = (v16u8)__msa_splati_b((v16i8)sum_w, 0);
303
304 ST_UB8(out, out, out, out, out, out, out, out, dst, dst_stride);
305 dst += (8 * dst_stride);
306 ST_UB8(out, out, out, out, out, out, out, out, dst, dst_stride);
307}
308
309static void intra_predict_128dc_16x16_msa(uint8_t *dst, int32_t dst_stride) {
310 const v16u8 out = (v16u8)__msa_ldi_b(128);
311
312 ST_UB8(out, out, out, out, out, out, out, out, dst, dst_stride);
313 dst += (8 * dst_stride);
314 ST_UB8(out, out, out, out, out, out, out, out, dst, dst_stride);
315}
316
317static void intra_predict_dc_32x32_msa(const uint8_t *src_top,
318 const uint8_t *src_left, uint8_t *dst,
319 int32_t dst_stride) {
320 uint32_t row;
321 v16u8 top0, top1, left0, left1, out;
322 v8u16 sum_h, sum_top0, sum_top1, sum_left0, sum_left1;
323 v4u32 sum_w;
324 v2u64 sum_d;
325
326 LD_UB2(src_top, 16, top0, top1);
327 LD_UB2(src_left, 16, left0, left1);
328 HADD_UB2_UH(top0, top1, sum_top0, sum_top1);
329 HADD_UB2_UH(left0, left1, sum_left0, sum_left1);
330 sum_h = sum_top0 + sum_top1;
331 sum_h += sum_left0 + sum_left1;
332 sum_w = __msa_hadd_u_w(sum_h, sum_h);
333 sum_d = __msa_hadd_u_d(sum_w, sum_w);
334 sum_w = (v4u32)__msa_pckev_w((v4i32)sum_d, (v4i32)sum_d);
335 sum_d = __msa_hadd_u_d(sum_w, sum_w);
336 sum_w = (v4u32)__msa_srari_w((v4i32)sum_d, 6);
337 out = (v16u8)__msa_splati_b((v16i8)sum_w, 0);
338
339 for (row = 16; row--;) {
340 ST_UB2(out, out, dst, 16);
341 dst += dst_stride;
342 ST_UB2(out, out, dst, 16);
343 dst += dst_stride;
344 }
345}
346
347static void intra_predict_dc_tl_32x32_msa(const uint8_t *src, uint8_t *dst,
348 int32_t dst_stride) {
349 uint32_t row;
350 v16u8 data0, data1, out;
351 v8u16 sum_h, sum_data0, sum_data1;
352 v4u32 sum_w;
353 v2u64 sum_d;
354
355 LD_UB2(src, 16, data0, data1);
356 HADD_UB2_UH(data0, data1, sum_data0, sum_data1);
357 sum_h = sum_data0 + sum_data1;
358 sum_w = __msa_hadd_u_w(sum_h, sum_h);
359 sum_d = __msa_hadd_u_d(sum_w, sum_w);
360 sum_w = (v4u32)__msa_pckev_w((v4i32)sum_d, (v4i32)sum_d);
361 sum_d = __msa_hadd_u_d(sum_w, sum_w);
362 sum_w = (v4u32)__msa_srari_w((v4i32)sum_d, 5);
363 out = (v16u8)__msa_splati_b((v16i8)sum_w, 0);
364
365 for (row = 16; row--;) {
366 ST_UB2(out, out, dst, 16);
367 dst += dst_stride;
368 ST_UB2(out, out, dst, 16);
369 dst += dst_stride;
370 }
371}
372
373static void intra_predict_128dc_32x32_msa(uint8_t *dst, int32_t dst_stride) {
374 uint32_t row;
375 const v16u8 out = (v16u8)__msa_ldi_b(128);
376
377 for (row = 16; row--;) {
378 ST_UB2(out, out, dst, 16);
379 dst += dst_stride;
380 ST_UB2(out, out, dst, 16);
381 dst += dst_stride;
382 }
383}
384
385static void intra_predict_tm_4x4_msa(const uint8_t *src_top_ptr,
386 const uint8_t *src_left, uint8_t *dst,
387 int32_t dst_stride) {
388 uint32_t val;
389 uint8_t top_left = src_top_ptr[-1];
390 v16i8 src_left0, src_left1, src_left2, src_left3, tmp0, tmp1, src_top = { 0 };
391 v16u8 src0, src1, src2, src3;
392 v8u16 src_top_left, vec0, vec1, vec2, vec3;
393
394 src_top_left = (v8u16)__msa_fill_h(top_left);
395 val = LW(src_top_ptr);
396 src_top = (v16i8)__msa_insert_w((v4i32)src_top, 0, val);
397
398 src_left0 = __msa_fill_b(src_left[0]);
399 src_left1 = __msa_fill_b(src_left[1]);
400 src_left2 = __msa_fill_b(src_left[2]);
401 src_left3 = __msa_fill_b(src_left[3]);
402
403 ILVR_B4_UB(src_left0, src_top, src_left1, src_top, src_left2, src_top,
404 src_left3, src_top, src0, src1, src2, src3);
405 HADD_UB4_UH(src0, src1, src2, src3, vec0, vec1, vec2, vec3);
406 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, vec0, vec1);
407 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, vec2, vec3);
408 SAT_UH4_UH(vec0, vec1, vec2, vec3, 7);
409 PCKEV_B2_SB(vec1, vec0, vec3, vec2, tmp0, tmp1);
410 ST4x4_UB(tmp0, tmp1, 0, 2, 0, 2, dst, dst_stride);
411}
412
413static void intra_predict_tm_8x8_msa(const uint8_t *src_top_ptr,
414 const uint8_t *src_left, uint8_t *dst,
415 int32_t dst_stride) {
416 uint64_t val;
417 uint8_t top_left = src_top_ptr[-1];
418 uint32_t loop_cnt;
419 v16i8 src_left0, src_left1, src_left2, src_left3, tmp0, tmp1, src_top = { 0 };
420 v8u16 src_top_left, vec0, vec1, vec2, vec3;
421 v16u8 src0, src1, src2, src3;
422
423 val = LD(src_top_ptr);
424 src_top = (v16i8)__msa_insert_d((v2i64)src_top, 0, val);
425 src_top_left = (v8u16)__msa_fill_h(top_left);
426
427 for (loop_cnt = 2; loop_cnt--;) {
428 src_left0 = __msa_fill_b(src_left[0]);
429 src_left1 = __msa_fill_b(src_left[1]);
430 src_left2 = __msa_fill_b(src_left[2]);
431 src_left3 = __msa_fill_b(src_left[3]);
432 src_left += 4;
433
434 ILVR_B4_UB(src_left0, src_top, src_left1, src_top, src_left2, src_top,
435 src_left3, src_top, src0, src1, src2, src3);
436 HADD_UB4_UH(src0, src1, src2, src3, vec0, vec1, vec2, vec3);
437 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, vec0, vec1);
438 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, vec2, vec3);
439 SAT_UH4_UH(vec0, vec1, vec2, vec3, 7);
440 PCKEV_B2_SB(vec1, vec0, vec3, vec2, tmp0, tmp1);
441 ST8x4_UB(tmp0, tmp1, dst, dst_stride);
442 dst += (4 * dst_stride);
443 }
444}
445
446static void intra_predict_tm_16x16_msa(const uint8_t *src_top_ptr,
447 const uint8_t *src_left, uint8_t *dst,
448 int32_t dst_stride) {
449 uint8_t top_left = src_top_ptr[-1];
450 uint32_t loop_cnt;
451 v16i8 src_top, src_left0, src_left1, src_left2, src_left3;
452 v8u16 src_top_left, res_r, res_l;
453
454 src_top = LD_SB(src_top_ptr);
455 src_top_left = (v8u16)__msa_fill_h(top_left);
456
457 for (loop_cnt = 4; loop_cnt--;) {
458 src_left0 = __msa_fill_b(src_left[0]);
459 src_left1 = __msa_fill_b(src_left[1]);
460 src_left2 = __msa_fill_b(src_left[2]);
461 src_left3 = __msa_fill_b(src_left[3]);
462 src_left += 4;
463
464 ILVRL_B2_UH(src_left0, src_top, res_r, res_l);
465 HADD_UB2_UH(res_r, res_l, res_r, res_l);
466 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r, res_l);
467
468 SAT_UH2_UH(res_r, res_l, 7);
469 PCKEV_ST_SB(res_r, res_l, dst);
470 dst += dst_stride;
471
472 ILVRL_B2_UH(src_left1, src_top, res_r, res_l);
473 HADD_UB2_UH(res_r, res_l, res_r, res_l);
474 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r, res_l);
475 SAT_UH2_UH(res_r, res_l, 7);
476 PCKEV_ST_SB(res_r, res_l, dst);
477 dst += dst_stride;
478
479 ILVRL_B2_UH(src_left2, src_top, res_r, res_l);
480 HADD_UB2_UH(res_r, res_l, res_r, res_l);
481 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r, res_l);
482 SAT_UH2_UH(res_r, res_l, 7);
483 PCKEV_ST_SB(res_r, res_l, dst);
484 dst += dst_stride;
485
486 ILVRL_B2_UH(src_left3, src_top, res_r, res_l);
487 HADD_UB2_UH(res_r, res_l, res_r, res_l);
488 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r, res_l);
489 SAT_UH2_UH(res_r, res_l, 7);
490 PCKEV_ST_SB(res_r, res_l, dst);
491 dst += dst_stride;
492 }
493}
494
495static void intra_predict_tm_32x32_msa(const uint8_t *src_top,
496 const uint8_t *src_left, uint8_t *dst,
497 int32_t dst_stride) {
498 uint8_t top_left = src_top[-1];
499 uint32_t loop_cnt;
500 v16i8 src_top0, src_top1, src_left0, src_left1, src_left2, src_left3;
501 v8u16 src_top_left, res_r0, res_r1, res_l0, res_l1;
502
503 LD_SB2(src_top, 16, src_top0, src_top1);
504 src_top_left = (v8u16)__msa_fill_h(top_left);
505
506 for (loop_cnt = 8; loop_cnt--;) {
507 src_left0 = __msa_fill_b(src_left[0]);
508 src_left1 = __msa_fill_b(src_left[1]);
509 src_left2 = __msa_fill_b(src_left[2]);
510 src_left3 = __msa_fill_b(src_left[3]);
511 src_left += 4;
512
513 ILVR_B2_UH(src_left0, src_top0, src_left0, src_top1, res_r0, res_r1);
514 ILVL_B2_UH(src_left0, src_top0, src_left0, src_top1, res_l0, res_l1);
515 HADD_UB4_UH(res_r0, res_l0, res_r1, res_l1, res_r0, res_l0, res_r1, res_l1);
516 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r0, res_l0);
517 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r1, res_l1);
518 SAT_UH4_UH(res_r0, res_l0, res_r1, res_l1, 7);
519 PCKEV_ST_SB(res_r0, res_l0, dst);
520 PCKEV_ST_SB(res_r1, res_l1, dst + 16);
521 dst += dst_stride;
522
523 ILVR_B2_UH(src_left1, src_top0, src_left1, src_top1, res_r0, res_r1);
524 ILVL_B2_UH(src_left1, src_top0, src_left1, src_top1, res_l0, res_l1);
525 HADD_UB4_UH(res_r0, res_l0, res_r1, res_l1, res_r0, res_l0, res_r1, res_l1);
526 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r0, res_l0);
527 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r1, res_l1);
528 SAT_UH4_UH(res_r0, res_l0, res_r1, res_l1, 7);
529 PCKEV_ST_SB(res_r0, res_l0, dst);
530 PCKEV_ST_SB(res_r1, res_l1, dst + 16);
531 dst += dst_stride;
532
533 ILVR_B2_UH(src_left2, src_top0, src_left2, src_top1, res_r0, res_r1);
534 ILVL_B2_UH(src_left2, src_top0, src_left2, src_top1, res_l0, res_l1);
535 HADD_UB4_UH(res_r0, res_l0, res_r1, res_l1, res_r0, res_l0, res_r1, res_l1);
536 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r0, res_l0);
537 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r1, res_l1);
538 SAT_UH4_UH(res_r0, res_l0, res_r1, res_l1, 7);
539 PCKEV_ST_SB(res_r0, res_l0, dst);
540 PCKEV_ST_SB(res_r1, res_l1, dst + 16);
541 dst += dst_stride;
542
543 ILVR_B2_UH(src_left3, src_top0, src_left3, src_top1, res_r0, res_r1);
544 ILVL_B2_UH(src_left3, src_top0, src_left3, src_top1, res_l0, res_l1);
545 HADD_UB4_UH(res_r0, res_l0, res_r1, res_l1, res_r0, res_l0, res_r1, res_l1);
546 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r0, res_l0);
547 IPRED_SUBS_UH2_UH(src_top_left, src_top_left, res_r1, res_l1);
548 SAT_UH4_UH(res_r0, res_l0, res_r1, res_l1, 7);
549 PCKEV_ST_SB(res_r0, res_l0, dst);
550 PCKEV_ST_SB(res_r1, res_l1, dst + 16);
551 dst += dst_stride;
552 }
553}
554
Yaowu Xuf883b422016-08-30 14:01:10 -0700555void aom_v_predictor_4x4_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700556 const uint8_t *above, const uint8_t *left) {
557 (void)left;
558
559 intra_predict_vert_4x4_msa(above, dst, y_stride);
560}
561
Yaowu Xuf883b422016-08-30 14:01:10 -0700562void aom_v_predictor_8x8_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700563 const uint8_t *above, const uint8_t *left) {
564 (void)left;
565
566 intra_predict_vert_8x8_msa(above, dst, y_stride);
567}
568
Yaowu Xuf883b422016-08-30 14:01:10 -0700569void aom_v_predictor_16x16_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700570 const uint8_t *above, const uint8_t *left) {
571 (void)left;
572
573 intra_predict_vert_16x16_msa(above, dst, y_stride);
574}
575
Yaowu Xuf883b422016-08-30 14:01:10 -0700576void aom_v_predictor_32x32_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700577 const uint8_t *above, const uint8_t *left) {
578 (void)left;
579
580 intra_predict_vert_32x32_msa(above, dst, y_stride);
581}
582
Yaowu Xuf883b422016-08-30 14:01:10 -0700583void aom_h_predictor_4x4_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700584 const uint8_t *above, const uint8_t *left) {
585 (void)above;
586
587 intra_predict_horiz_4x4_msa(left, dst, y_stride);
588}
589
Yaowu Xuf883b422016-08-30 14:01:10 -0700590void aom_h_predictor_8x8_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700591 const uint8_t *above, const uint8_t *left) {
592 (void)above;
593
594 intra_predict_horiz_8x8_msa(left, dst, y_stride);
595}
596
Yaowu Xuf883b422016-08-30 14:01:10 -0700597void aom_h_predictor_16x16_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700598 const uint8_t *above, const uint8_t *left) {
599 (void)above;
600
601 intra_predict_horiz_16x16_msa(left, dst, y_stride);
602}
603
Yaowu Xuf883b422016-08-30 14:01:10 -0700604void aom_h_predictor_32x32_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700605 const uint8_t *above, const uint8_t *left) {
606 (void)above;
607
608 intra_predict_horiz_32x32_msa(left, dst, y_stride);
609}
610
Yaowu Xuf883b422016-08-30 14:01:10 -0700611void aom_dc_predictor_4x4_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700612 const uint8_t *above, const uint8_t *left) {
613 intra_predict_dc_4x4_msa(above, left, dst, y_stride);
614}
615
Yaowu Xuf883b422016-08-30 14:01:10 -0700616void aom_dc_predictor_8x8_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700617 const uint8_t *above, const uint8_t *left) {
618 intra_predict_dc_8x8_msa(above, left, dst, y_stride);
619}
620
Yaowu Xuf883b422016-08-30 14:01:10 -0700621void aom_dc_predictor_16x16_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700622 const uint8_t *above, const uint8_t *left) {
623 intra_predict_dc_16x16_msa(above, left, dst, y_stride);
624}
625
Yaowu Xuf883b422016-08-30 14:01:10 -0700626void aom_dc_predictor_32x32_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700627 const uint8_t *above, const uint8_t *left) {
628 intra_predict_dc_32x32_msa(above, left, dst, y_stride);
629}
630
Yaowu Xuf883b422016-08-30 14:01:10 -0700631void aom_dc_top_predictor_4x4_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700632 const uint8_t *above, const uint8_t *left) {
633 (void)left;
634
635 intra_predict_dc_tl_4x4_msa(above, dst, y_stride);
636}
637
Yaowu Xuf883b422016-08-30 14:01:10 -0700638void aom_dc_top_predictor_8x8_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700639 const uint8_t *above, const uint8_t *left) {
640 (void)left;
641
642 intra_predict_dc_tl_8x8_msa(above, dst, y_stride);
643}
644
Yaowu Xuf883b422016-08-30 14:01:10 -0700645void aom_dc_top_predictor_16x16_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700646 const uint8_t *above, const uint8_t *left) {
647 (void)left;
648
649 intra_predict_dc_tl_16x16_msa(above, dst, y_stride);
650}
651
Yaowu Xuf883b422016-08-30 14:01:10 -0700652void aom_dc_top_predictor_32x32_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700653 const uint8_t *above, const uint8_t *left) {
654 (void)left;
655
656 intra_predict_dc_tl_32x32_msa(above, dst, y_stride);
657}
658
Yaowu Xuf883b422016-08-30 14:01:10 -0700659void aom_dc_left_predictor_4x4_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700660 const uint8_t *above, const uint8_t *left) {
661 (void)above;
662
663 intra_predict_dc_tl_4x4_msa(left, dst, y_stride);
664}
665
Yaowu Xuf883b422016-08-30 14:01:10 -0700666void aom_dc_left_predictor_8x8_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700667 const uint8_t *above, const uint8_t *left) {
668 (void)above;
669
670 intra_predict_dc_tl_8x8_msa(left, dst, y_stride);
671}
672
Yaowu Xuf883b422016-08-30 14:01:10 -0700673void aom_dc_left_predictor_16x16_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700674 const uint8_t *above,
675 const uint8_t *left) {
676 (void)above;
677
678 intra_predict_dc_tl_16x16_msa(left, dst, y_stride);
679}
680
Yaowu Xuf883b422016-08-30 14:01:10 -0700681void aom_dc_left_predictor_32x32_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700682 const uint8_t *above,
683 const uint8_t *left) {
684 (void)above;
685
686 intra_predict_dc_tl_32x32_msa(left, dst, y_stride);
687}
688
Yaowu Xuf883b422016-08-30 14:01:10 -0700689void aom_dc_128_predictor_4x4_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700690 const uint8_t *above, const uint8_t *left) {
691 (void)above;
692 (void)left;
693
694 intra_predict_128dc_4x4_msa(dst, y_stride);
695}
696
Yaowu Xuf883b422016-08-30 14:01:10 -0700697void aom_dc_128_predictor_8x8_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700698 const uint8_t *above, const uint8_t *left) {
699 (void)above;
700 (void)left;
701
702 intra_predict_128dc_8x8_msa(dst, y_stride);
703}
704
Yaowu Xuf883b422016-08-30 14:01:10 -0700705void aom_dc_128_predictor_16x16_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700706 const uint8_t *above, const uint8_t *left) {
707 (void)above;
708 (void)left;
709
710 intra_predict_128dc_16x16_msa(dst, y_stride);
711}
712
Yaowu Xuf883b422016-08-30 14:01:10 -0700713void aom_dc_128_predictor_32x32_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700714 const uint8_t *above, const uint8_t *left) {
715 (void)above;
716 (void)left;
717
718 intra_predict_128dc_32x32_msa(dst, y_stride);
719}
720
Yaowu Xuf883b422016-08-30 14:01:10 -0700721void aom_tm_predictor_4x4_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700722 const uint8_t *above, const uint8_t *left) {
723 intra_predict_tm_4x4_msa(above, left, dst, y_stride);
724}
725
Yaowu Xuf883b422016-08-30 14:01:10 -0700726void aom_tm_predictor_8x8_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700727 const uint8_t *above, const uint8_t *left) {
728 intra_predict_tm_8x8_msa(above, left, dst, y_stride);
729}
730
Yaowu Xuf883b422016-08-30 14:01:10 -0700731void aom_tm_predictor_16x16_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700732 const uint8_t *above, const uint8_t *left) {
733 intra_predict_tm_16x16_msa(above, left, dst, y_stride);
734}
735
Yaowu Xuf883b422016-08-30 14:01:10 -0700736void aom_tm_predictor_32x32_msa(uint8_t *dst, ptrdiff_t y_stride,
Yaowu Xuc27fc142016-08-22 16:08:15 -0700737 const uint8_t *above, const uint8_t *left) {
738 intra_predict_tm_32x32_msa(above, left, dst, y_stride);
739}