27 v8i16 weights = { 0, 22725, 21407, 19266, 16383, 12873, 8867, 4520 };
28 v8i16 in0, in1, in2, in3, in4, in5, in6, in7;
30 v8i16 const0, const1, const2, const3, const4, const5, const6, const7;
31 v4i32 temp0_r, temp1_r, temp2_r, temp3_r;
32 v4i32 temp0_l, temp1_l, temp2_l, temp3_l;
33 v4i32 a0_r, a1_r, a2_r, a3_r, a0_l, a1_l, a2_l, a3_l;
34 v4i32 b0_r, b1_r, b2_r, b3_r, b0_l, b1_l, b2_l, b3_l;
36 v8i16 select_vec,
temp;
38 v4i32 const_val0 = __msa_ldi_w(1);
39 v4i32 const_val1 = __msa_ldi_w(1);
41 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7);
43 const_val = 16383 * ((1 << 19) / 16383);
44 const_val1 = __msa_insert_w(const_val0, 0, const_val);
45 const_val1 = __msa_splati_w(const_val1, 0);
47 in0, in1, in2, in3, in4, in5, in6, in7);
48 select_vec = in1 | in2 | in3 | in4 | in5 | in6 | in7;
49 select_vec = __msa_clti_u_h((v8u16) select_vec, 1);
53 w2 = (v4i32) __msa_splati_h(weights, 2);
54 w2 = (v4i32) __msa_ilvr_h(zero, (v8i16) w2);
55 w4 = (v4i32) __msa_splati_h(weights, 4);
56 w4 = (v4i32) __msa_ilvr_h(zero, (v8i16) w4);
57 w6 = (v4i32) __msa_splati_h(weights, 6);
58 w6 = (v4i32) __msa_ilvr_h(zero, (v8i16) w6);
59 MUL2(a0_r, w4, a0_l, w4, a0_r, a0_l);
60 ADD2(a0_r, const_val0, a0_l, const_val0, temp0_r, temp0_l);
61 MUL4(w2, temp3_r, w2, temp3_l, w6, temp3_r, w6, temp3_l,
62 temp1_r, temp1_l, temp2_r, temp2_l);
64 temp2_l, temp2_r, temp1_l, temp1_r,
65 a0_r, a0_l, a1_r, a1_l, a2_l, a2_r, a3_l, a3_r);
68 MUL2(temp0_r, w4, temp0_l, w4, temp0_r, temp0_l);
69 MUL4(w2, temp3_r, w2, temp3_l, w6, temp3_r, w6, temp3_l,
70 temp2_r, temp2_l, temp1_r, temp1_l);
71 ADD2(a0_r, temp0_r, a0_l, temp0_l, a0_r, a0_l);
72 SUB4(a1_r, temp0_r, a1_l, temp0_l, a2_r, temp0_r, a2_l, temp0_l,
73 a1_r, a1_l, a2_r, a2_l);
74 ADD4(a3_r, temp0_r, a3_l, temp0_l, a0_r, temp1_r, a0_l, temp1_l,
75 a3_r, a3_l, a0_r, a0_l);
76 SUB2(a1_r, temp2_r, a1_l, temp2_l, a1_r, a1_l);
77 ADD2(a2_r, temp2_r, a2_l, temp2_l, a2_r, a2_l);
78 SUB2(a3_r, temp1_r, a3_l, temp1_l, a3_r, a3_l);
83 const0, const1, const2, const3);
85 const5 = __msa_ilvod_h(-w1, -w5);
86 const7 = __msa_ilvod_h(w3, -w1);
87 DOTP_SH4_SW(b3_r, b3_r, b3_r, b3_r, const0, const1, const2, const3,
88 b0_r, b1_r, b2_r, b3_r);
90 const4, const5, const6, const7, b0_r, b1_r, b2_r, b3_r);
91 DOTP_SH4_SW(b3_l, b3_l, b3_l, b3_l, const0, const1, const2, const3,
92 b0_l, b1_l, b2_l, b3_l);
94 const4, const5, const6, const7, b0_l, b1_l, b2_l, b3_l);
95 BUTTERFLY_16(a0_r, a0_l, a1_r, a1_l, a2_r, a2_l, a3_r, a3_l,
96 b3_l, b3_r, b2_l, b2_r, b1_l, b1_r, b0_l, b0_r,
97 temp0_r, temp0_l, temp1_r, temp1_l,
98 temp2_r, temp2_l, temp3_r, temp3_l,
99 a3_l, a3_r, a2_l, a2_r, a1_l, a1_r, a0_l, a0_r);
100 SRA_4V(temp0_r, temp0_l, temp1_r, temp1_l, 11);
101 SRA_4V(temp2_r, temp2_l, temp3_r, temp3_l, 11);
103 temp2_l, temp2_r, temp3_l, temp3_r,
104 temp0_r, temp1_r, temp2_r, temp3_r);
105 in0 = (v8i16) __msa_bmnz_v((v16u8) temp0_r, (v16u8) temp,
107 in1 = (v8i16) __msa_bmnz_v((v16u8) temp1_r, (v16u8) temp,
109 in2 = (v8i16) __msa_bmnz_v((v16u8) temp2_r, (v16u8) temp,
111 in3 = (v8i16) __msa_bmnz_v((v16u8) temp3_r, (v16u8) temp,
113 SRA_4V(a3_r, a3_l, a2_r, a2_l, 11);
114 SRA_4V(a1_r, a1_l, a0_r, a0_l, 11);
115 PCKEV_H4_SW(a0_l, a0_r, a1_l, a1_r, a2_l, a2_r, a3_l, a3_r,
116 a0_r, a1_r, a2_r, a3_r);
117 in4 = (v8i16) __msa_bmnz_v((v16u8) a3_r, (v16u8) temp, (v16u8) select_vec);
118 in5 = (v8i16) __msa_bmnz_v((v16u8) a2_r, (v16u8) temp, (v16u8) select_vec);
119 in6 = (v8i16) __msa_bmnz_v((v16u8) a1_r, (v16u8) temp, (v16u8) select_vec);
120 in7 = (v8i16) __msa_bmnz_v((v16u8) a0_r, (v16u8) temp, (v16u8) select_vec);
122 in0, in1, in2, in3, in4, in5, in6, in7);
126 w2 = (v4i32) __msa_splati_h(weights, 2);
127 w2 = (v4i32) __msa_ilvr_h(zero, (v8i16) w2);
128 w4 = (v4i32) __msa_splati_h(weights, 4);
129 w4 = (v4i32) __msa_ilvr_h(zero, (v8i16) w4);
130 w6 = (v4i32) __msa_splati_h(weights, 6);
131 w6 = (v4i32) __msa_ilvr_h(zero, (v8i16) w6);
132 MUL2(a0_r, w4, a0_l, w4, a0_r, a0_l);
133 ADD2(a0_r, const_val1, a0_l, const_val1, temp0_r, temp0_l);
134 MUL4(w2, temp3_r, w2, temp3_l, w6, temp3_r, w6, temp3_l,
135 temp1_r, temp1_l, temp2_r, temp2_l);
137 temp2_l, temp2_r, temp1_l, temp1_r,
138 a0_r, a0_l, a1_r, a1_l, a2_l, a2_r, a3_l, a3_r);
141 MUL2(temp0_r, w4, temp0_l, w4, temp0_r, temp0_l);
142 MUL4(w2, temp3_r, w2, temp3_l, w6, temp3_r, w6, temp3_l,
143 temp2_r, temp2_l, temp1_r, temp1_l);
144 ADD2(a0_r, temp0_r, a0_l, temp0_l, a0_r, a0_l);
145 SUB4(a1_r, temp0_r, a1_l, temp0_l, a2_r, temp0_r, a2_l, temp0_l,
146 a1_r, a1_l, a2_r, a2_l);
147 ADD4(a3_r, temp0_r, a3_l, temp0_l, a0_r, temp1_r, a0_l, temp1_l,
148 a3_r, a3_l, a0_r, a0_l);
149 SUB2(a1_r, temp2_r, a1_l, temp2_l, a1_r, a1_l);
150 ADD2(a2_r, temp2_r, a2_l, temp2_l, a2_r, a2_l);
151 SUB2(a3_r, temp1_r, a3_l, temp1_l, a3_r, a3_l);
155 const0, const1, const2, const3);
156 DOTP_SH4_SW(b3_r, b3_r, b3_r, b3_r, const0, const1, const2, const3,
157 b0_r, b1_r, b2_r, b3_r);
158 DOTP_SH4_SW(b3_l, b3_l, b3_l, b3_l, const0, const1, const2, const3,
159 b0_l, b1_l, b2_l, b3_l);
162 const5 = __msa_ilvod_h(-w1, -w5);
163 const7 = __msa_ilvod_h(w3, -w1);
165 const4, const5, const6, const7, b0_r, b1_r, b2_r, b3_r);
167 const4, const5, const6, const7, b0_l, b1_l, b2_l, b3_l);
168 BUTTERFLY_16(a0_r, a0_l, a1_r, a1_l, a2_r, a2_l, a3_r, a3_l,
169 b3_l, b3_r, b2_l, b2_r, b1_l, b1_r, b0_l, b0_r,
170 temp0_r, temp0_l, temp1_r, temp1_l,
171 temp2_r, temp2_l, temp3_r, temp3_l,
172 a3_l, a3_r, a2_l, a2_r, a1_l, a1_r, a0_l, a0_r);
173 SRA_4V(temp0_r, temp0_l, temp1_r, temp1_l, 20);
174 SRA_4V(temp2_r, temp2_l, temp3_r, temp3_l, 20);
175 PCKEV_H4_SW(temp0_l, temp0_r, temp1_l, temp1_r, temp2_l, temp2_r,
176 temp3_l, temp3_r, temp0_r, temp1_r, temp2_r, temp3_r);
177 SRA_4V(a3_r, a3_l, a2_r, a2_l, 20);
178 SRA_4V(a1_r, a1_l, a0_r, a0_l, 20);
179 PCKEV_H4_SW(a0_l, a0_r, a1_l, a1_r, a2_l, a2_r, a3_l, a3_r,
180 a0_r, a1_r, a2_r, a3_r);
181 ST_SW8(temp0_r, temp1_r, temp2_r, temp3_r, a3_r, a2_r, a1_r, a0_r,
189 uint64_t tmp0, tmp1, tmp2, tmp3;
190 v8i16 weights = { 0, 22725, 21407, 19266, 16383, 12873, 8867, 4520 };
191 v8i16 in0, in1, in2, in3, in4, in5, in6, in7;
192 v8i16 w1, w3, w5, w7;
193 v8i16 const0, const1, const2, const3, const4, const5, const6, const7;
194 v4i32 temp0_r, temp1_r, temp2_r, temp3_r;
195 v4i32 temp0_l, temp1_l, temp2_l, temp3_l;
196 v4i32 a0_r, a1_r, a2_r, a3_r, a0_l, a1_l, a2_l, a3_l;
197 v4i32 b0_r, b1_r, b2_r, b3_r, b0_l, b1_l, b2_l, b3_l;
199 v8i16 select_vec,
temp;
201 v4i32 const_val0 = __msa_ldi_w(1);
202 v4i32 const_val1 = __msa_ldi_w(1);
204 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7);
206 const_val = 16383 * ((1 << 19) / 16383);
207 const_val1 = __msa_insert_w(const_val0, 0, const_val);
208 const_val1 = __msa_splati_w(const_val1, 0);
210 in0, in1, in2, in3, in4, in5, in6, in7);
211 select_vec = in1 | in2 | in3 | in4 | in5 | in6 | in7;
212 select_vec = __msa_clti_u_h((v8u16) select_vec, 1);
216 w2 = (v4i32) __msa_splati_h(weights, 2);
217 w2 = (v4i32) __msa_ilvr_h(zero, (v8i16) w2);
218 w4 = (v4i32) __msa_splati_h(weights, 4);
219 w4 = (v4i32) __msa_ilvr_h(zero, (v8i16) w4);
220 w6 = (v4i32) __msa_splati_h(weights, 6);
221 w6 = (v4i32) __msa_ilvr_h(zero, (v8i16) w6);
222 MUL2(a0_r, w4, a0_l, w4, a0_r, a0_l);
223 ADD2(a0_r, const_val0, a0_l, const_val0, temp0_r, temp0_l);
224 MUL2(w2, temp3_r, w2, temp3_l, temp1_r, temp1_l);
225 MUL2(w6, temp3_r, w6, temp3_l, temp2_r, temp2_l);
227 temp2_l, temp2_r, temp1_l, temp1_r,
228 a0_r, a0_l, a1_r, a1_l, a2_l, a2_r, a3_l, a3_r);
231 MUL2(temp0_r, w4, temp0_l, w4, temp0_r, temp0_l);
232 MUL2(w2, temp3_r, w2, temp3_l, temp2_r, temp2_l);
233 MUL2(w6, temp3_r, w6, temp3_l, temp1_r, temp1_l);
234 ADD2(a0_r, temp0_r, a0_l, temp0_l, a0_r, a0_l);
235 SUB2(a1_r, temp0_r, a1_l, temp0_l, a1_r, a1_l);
236 SUB2(a2_r, temp0_r, a2_l, temp0_l, a2_r, a2_l);
237 ADD2(a3_r, temp0_r, a3_l, temp0_l, a3_r, a3_l);
238 ADD2(a0_r, temp1_r, a0_l, temp1_l, a0_r, a0_l);
239 SUB2(a1_r, temp2_r, a1_l, temp2_l, a1_r, a1_l);
240 ADD2(a2_r, temp2_r, a2_l, temp2_l, a2_r, a2_l);
241 SUB2(a3_r, temp1_r, a3_l, temp1_l, a3_r, a3_l);
246 const0, const1, const2, const3);
248 const5 = __msa_ilvod_h(-w1, -w5);
249 const7 = __msa_ilvod_h(w3, -w1);
250 DOTP_SH4_SW(b3_r, b3_r, b3_r, b3_r, const0, const1, const2, const3,
251 b0_r, b1_r, b2_r, b3_r);
253 const4, const5, const6, const7, b0_r, b1_r, b2_r, b3_r);
254 DOTP_SH4_SW(b3_l, b3_l, b3_l, b3_l, const0, const1, const2, const3,
255 b0_l, b1_l, b2_l, b3_l);
257 const4, const5, const6, const7, b0_l, b1_l, b2_l, b3_l);
258 BUTTERFLY_16(a0_r, a0_l, a1_r, a1_l, a2_r, a2_l, a3_r, a3_l,
259 b3_l, b3_r, b2_l, b2_r, b1_l, b1_r, b0_l, b0_r,
260 temp0_r, temp0_l, temp1_r, temp1_l,
261 temp2_r, temp2_l, temp3_r, temp3_l,
262 a3_l, a3_r, a2_l, a2_r, a1_l, a1_r, a0_l, a0_r);
263 SRA_4V(temp0_r, temp0_l, temp1_r, temp1_l, 11);
264 SRA_4V(temp2_r, temp2_l, temp3_r, temp3_l, 11);
266 temp2_l, temp2_r, temp3_l, temp3_r,
267 temp0_r, temp1_r, temp2_r, temp3_r);
268 in0 = (v8i16) __msa_bmnz_v((v16u8) temp0_r, (v16u8) temp,
270 in1 = (v8i16) __msa_bmnz_v((v16u8) temp1_r, (v16u8) temp,
272 in2 = (v8i16) __msa_bmnz_v((v16u8) temp2_r, (v16u8) temp,
274 in3 = (v8i16) __msa_bmnz_v((v16u8) temp3_r, (v16u8) temp,
276 SRA_4V(a3_r, a3_l, a2_r, a2_l, 11);
277 SRA_4V(a1_r, a1_l, a0_r, a0_l, 11);
278 PCKEV_H4_SW(a0_l, a0_r, a1_l, a1_r, a2_l, a2_r, a3_l, a3_r,
279 a0_r, a1_r, a2_r, a3_r);
280 in4 = (v8i16) __msa_bmnz_v((v16u8) a3_r, (v16u8) temp, (v16u8) select_vec);
281 in5 = (v8i16) __msa_bmnz_v((v16u8) a2_r, (v16u8) temp, (v16u8) select_vec);
282 in6 = (v8i16) __msa_bmnz_v((v16u8) a1_r, (v16u8) temp, (v16u8) select_vec);
283 in7 = (v8i16) __msa_bmnz_v((v16u8) a0_r, (v16u8) temp, (v16u8) select_vec);
285 in0, in1, in2, in3, in4, in5, in6, in7);
288 w2 = (v4i32) __msa_splati_h(weights, 2);
289 w2 = (v4i32) __msa_ilvr_h(zero, (v8i16) w2);
290 w4 = (v4i32) __msa_splati_h(weights, 4);
291 w4 = (v4i32) __msa_ilvr_h(zero, (v8i16) w4);
292 w6 = (v4i32) __msa_splati_h(weights, 6);
293 w6 = (v4i32) __msa_ilvr_h(zero, (v8i16) w6);
294 MUL2(a0_r, w4, a0_l, w4, a0_r, a0_l);
295 ADD2(a0_r, const_val1, a0_l, const_val1, temp0_r, temp0_l);
296 MUL2(w2, temp3_r, w2, temp3_l, temp1_r, temp1_l);
297 MUL2(w6, temp3_r, w6, temp3_l, temp2_r, temp2_l);
299 temp2_l, temp2_r, temp1_l, temp1_r,
300 a0_r, a0_l, a1_r, a1_l, a2_l, a2_r, a3_l, a3_r);
303 MUL2(temp0_r, w4, temp0_l, w4, temp0_r, temp0_l);
304 MUL2(w2, temp3_r, w2, temp3_l, temp2_r, temp2_l);
305 MUL2(w6, temp3_r, w6, temp3_l, temp1_r, temp1_l);
306 ADD2(a0_r, temp0_r, a0_l, temp0_l, a0_r, a0_l);
307 SUB2(a1_r, temp0_r, a1_l, temp0_l, a1_r, a1_l);
308 SUB2(a2_r, temp0_r, a2_l, temp0_l, a2_r, a2_l);
309 ADD2(a3_r, temp0_r, a3_l, temp0_l, a3_r, a3_l);
310 ADD2(a0_r, temp1_r, a0_l, temp1_l, a0_r, a0_l);
311 SUB2(a1_r, temp2_r, a1_l, temp2_l, a1_r, a1_l);
312 ADD2(a2_r, temp2_r, a2_l, temp2_l, a2_r, a2_l);
313 SUB2(a3_r, temp1_r, a3_l, temp1_l, a3_r, a3_l);
317 const0, const1, const2, const3);
318 DOTP_SH4_SW(b3_r, b3_r, b3_r, b3_r, const0, const1, const2, const3,
319 b0_r, b1_r, b2_r, b3_r);
320 DOTP_SH4_SW(b3_l, b3_l, b3_l, b3_l, const0, const1, const2, const3,
321 b0_l, b1_l, b2_l, b3_l);
324 const5 = __msa_ilvod_h(-w1, -w5);
325 const7 = __msa_ilvod_h(w3, -w1);
327 const4, const5, const6, const7, b0_r, b1_r, b2_r, b3_r);
329 const4, const5, const6, const7, b0_l, b1_l, b2_l, b3_l);
330 BUTTERFLY_16(a0_r, a0_l, a1_r, a1_l, a2_r, a2_l, a3_r, a3_l,
331 b3_l, b3_r, b2_l, b2_r, b1_l, b1_r, b0_l, b0_r,
332 temp0_r, temp0_l, temp1_r, temp1_l,
333 temp2_r, temp2_l, temp3_r, temp3_l,
334 a3_l, a3_r, a2_l, a2_r, a1_l, a1_r, a0_l, a0_r);
335 SRA_4V(temp0_r, temp0_l, temp1_r, temp1_l, 20);
336 SRA_4V(temp2_r, temp2_l, temp3_r, temp3_l, 20);
337 SRA_4V(a3_r, a3_l, a2_r, a2_l, 20);
338 SRA_4V(a1_r, a1_l, a0_r, a0_l, 20);
339 PCKEV_H4_SW(temp0_l, temp0_r, temp1_l, temp1_r, temp2_l, temp2_r,
340 temp3_l, temp3_r, temp0_r, temp1_r, temp2_r, temp3_r);
341 PCKEV_H4_SW(a0_l, a0_r, a1_l, a1_r, a2_l, a2_r, a3_l, a3_r,
342 a0_r, a1_r, a2_r, a3_r);
348 temp2_r, temp2_r, temp3_r, temp3_r,
349 temp0_r, temp1_r, temp2_r, temp3_r);
350 tmp0 = __msa_copy_u_d((v2i64) temp0_r, 1);
351 tmp1 = __msa_copy_u_d((v2i64) temp1_r, 1);
352 tmp2 = __msa_copy_u_d((v2i64) temp2_r, 1);
353 tmp3 = __msa_copy_u_d((v2i64) temp3_r, 1);
354 SD4(tmp0, tmp1, tmp2, tmp3, dst, dst_stride);
355 dst += 4 * dst_stride;
361 a2_r, a2_r, a3_r, a3_r, a0_r, a1_r, a2_r, a3_r);
362 tmp3 = __msa_copy_u_d((v2i64) a0_r, 1);
363 tmp2 = __msa_copy_u_d((v2i64) a1_r, 1);
364 tmp1 = __msa_copy_u_d((v2i64) a2_r, 1);
365 tmp0 = __msa_copy_u_d((v2i64) a3_r, 1);
366 SD4(tmp0, tmp1, tmp2, tmp3, dst, dst_stride);
367 dst += 4 * dst_stride;
374 uint64_t tmp0, tmp1, tmp2, tmp3;
375 v8i16 weights = { 0, 22725, 21407, 19266, 16383, 12873, 8867, 4520 };
376 v8i16 in0, in1, in2, in3, in4, in5, in6, in7;
377 v8i16 w1, w3, w5, w7;
378 v8i16 const0, const1, const2, const3, const4, const5, const6, const7;
379 v4i32 temp0_r, temp1_r, temp2_r, temp3_r;
380 v4i32 temp4_r, temp5_r, temp6_r, temp7_r, temp8_r;
381 v4i32 temp0_l, temp1_l, temp2_l, temp3_l;
382 v4i32 temp4_l, temp5_l, temp6_l, temp7_l, temp8_l;
383 v4i32 a0_r, a1_r, a2_r, a3_r, a0_l, a1_l, a2_l, a3_l;
384 v4i32 b0_r, b1_r, b2_r, b3_r, b0_l, b1_l, b2_l, b3_l;
386 v8i16 select_vec,
temp;
388 v4i32 const_val0 = __msa_ldi_w(1);
389 v4i32 const_val1 = __msa_ldi_w(1);
392 const_val = 16383 * ((1 << 19) / 16383);
393 const_val1 = __msa_insert_w(const_val0, 0, const_val);
394 const_val1 = __msa_splati_w(const_val1, 0);
395 LD_SH8(block, 8, in0, in1, in2, in3, in4, in5, in6, in7);
397 in0, in1, in2, in3, in4, in5, in6, in7);
399 select_vec = in1 | in2 | in3 | in4 | in5 | in6 | in7;
400 select_vec = __msa_clti_u_h((v8u16) select_vec, 1);
410 const0, const1, const2, const3);
412 const5 = __msa_ilvod_h(-w1, -w5);
413 const7 = __msa_ilvod_h(w3, -w1);
414 DOTP_SH4_SW(b3_r, b3_r, b3_r, b3_r, const0, const1, const2, const3,
415 b0_r, b1_r, b2_r, b3_r);
417 const4, const5, const6, const7, b0_r, b1_r, b2_r, b3_r);
418 DOTP_SH4_SW(b3_l, b3_l, b3_l, b3_l, const0, const1, const2, const3,
419 b0_l, b1_l, b2_l, b3_l);
421 const4, const5, const6, const7, b0_l, b1_l, b2_l, b3_l);
422 w2 = (v4i32) __msa_splati_h(weights, 2);
423 w2 = (v4i32) __msa_ilvr_h(zero, (v8i16) w2);
424 w4 = (v4i32) __msa_splati_h(weights, 4);
425 w4 = (v4i32) __msa_ilvr_h(zero, (v8i16) w4);
426 w6 = (v4i32) __msa_splati_h(weights, 6);
427 w6 = (v4i32) __msa_ilvr_h(zero, (v8i16) w6);
428 MUL2(a0_r, w4, a0_l, w4, a0_r, a0_l);
429 ADD2(a0_r, const_val0, a0_l, const_val0, temp0_r, temp0_l);
430 MUL2(w2, temp3_r, w2, temp3_l, temp1_r, temp1_l);
431 MUL2(w6, temp3_r, w6, temp3_l, temp2_r, temp2_l);
433 temp2_l, temp2_r, temp1_l, temp1_r,
434 a0_r, a0_l, a1_r, a1_l, a2_l, a2_r, a3_l, a3_r);
435 MUL2(temp4_r, w4, temp4_l, w4, temp4_r, temp4_l);
436 MUL2(temp7_r, w2, temp7_l, w2, temp6_r, temp6_l);
437 MUL2(temp7_r, w6, temp7_l, w6, temp5_r, temp5_l);
438 ADD2(a0_r, temp4_r, a0_l, temp4_l, a0_r, a0_l);
439 SUB2(a1_r, temp4_r, a1_l, temp4_l, a1_r, a1_l);
440 SUB2(a2_r, temp4_r, a2_l, temp4_l, a2_r, a2_l);
441 ADD2(a3_r, temp4_r, a3_l, temp4_l, a3_r, a3_l);
442 ADD2(a0_r, temp5_r, a0_l, temp5_l, a0_r, a0_l);
443 SUB2(a1_r, temp6_r, a1_l, temp6_l, a1_r, a1_l);
444 ADD2(a2_r, temp6_r, a2_l, temp6_l, a2_r, a2_l);
445 SUB2(a3_r, temp5_r, a3_l, temp5_l, a3_r, a3_l);
446 BUTTERFLY_16(a0_r, a0_l, a1_r, a1_l, a2_r, a2_l, a3_r, a3_l,
447 b3_l, b3_r, b2_l, b2_r, b1_l, b1_r, b0_l, b0_r,
448 temp0_r, temp0_l, temp1_r, temp1_l,
449 temp2_r, temp2_l, temp3_r, temp3_l,
450 a3_l, a3_r, a2_l, a2_r, a1_l, a1_r, a0_l, a0_r);
451 SRA_4V(temp0_r, temp0_l, temp1_r, temp1_l, 11);
452 SRA_4V(temp2_r, temp2_l, temp3_r, temp3_l, 11);
454 temp2_l, temp2_r, temp3_l, temp3_r,
455 temp0_r, temp1_r, temp2_r, temp3_r);
456 in0 = (v8i16) __msa_bmnz_v((v16u8) temp0_r, (v16u8) temp,
458 in1 = (v8i16) __msa_bmnz_v((v16u8) temp1_r, (v16u8) temp,
460 in2 = (v8i16) __msa_bmnz_v((v16u8) temp2_r, (v16u8) temp,
462 in3 = (v8i16) __msa_bmnz_v((v16u8) temp3_r, (v16u8) temp,
464 SRA_4V(a3_r, a3_l, a2_r, a2_l, 11);
465 SRA_4V(a1_r, a1_l, a0_r, a0_l, 11);
466 PCKEV_H4_SW(a0_l, a0_r, a1_l, a1_r, a2_l, a2_r, a3_l, a3_r,
467 a0_r, a1_r, a2_r, a3_r);
468 in4 = (v8i16) __msa_bmnz_v((v16u8) a3_r, (v16u8) temp, (v16u8) select_vec);
469 in5 = (v8i16) __msa_bmnz_v((v16u8) a2_r, (v16u8) temp, (v16u8) select_vec);
470 in6 = (v8i16) __msa_bmnz_v((v16u8) a1_r, (v16u8) temp, (v16u8) select_vec);
471 in7 = (v8i16) __msa_bmnz_v((v16u8) a0_r, (v16u8) temp, (v16u8) select_vec);
473 in0, in1, in2, in3, in4, in5, in6, in7);
477 MUL2(a0_r, w4, a0_l, w4, a0_r, a0_l);
478 ADD2(a0_r, const_val1, a0_l, const_val1, temp0_r, temp0_l);
479 MUL2(w2, temp3_r, w2, temp3_l, temp1_r, temp1_l);
480 MUL2(w6, temp3_r, w6, temp3_l, temp2_r, temp2_l);
482 temp2_l, temp2_r, temp1_l, temp1_r,
483 a0_r, a0_l, a1_r, a1_l, a2_l, a2_r, a3_l, a3_r);
486 MUL2(temp0_r, w4, temp0_l, w4, temp0_r, temp0_l);
487 MUL2(w2, temp3_r, w2, temp3_l, temp2_r, temp2_l);
488 MUL2(w6, temp3_r, w6, temp3_l, temp1_r, temp1_l);
489 ADD2(a0_r, temp0_r, a0_l, temp0_l, a0_r, a0_l);
490 SUB2(a1_r, temp0_r, a1_l, temp0_l, a1_r, a1_l);
491 SUB2(a2_r, temp0_r, a2_l, temp0_l, a2_r, a2_l);
492 ADD2(a3_r, temp0_r, a3_l, temp0_l, a3_r, a3_l);
493 ADD2(a0_r, temp1_r, a0_l, temp1_l, a0_r, a0_l);
494 SUB2(a1_r, temp2_r, a1_l, temp2_l, a1_r, a1_l);
495 ADD2(a2_r, temp2_r, a2_l, temp2_l, a2_r, a2_l);
496 SUB2(a3_r, temp1_r, a3_l, temp1_l, a3_r, a3_l);
499 DOTP_SH4_SW(b3_r, b3_r, b3_r, b3_r, const0, const1, const2, const3,
500 b0_r, b1_r, b2_r, b3_r);
501 DOTP_SH4_SW(b3_l, b3_l, b3_l, b3_l, const0, const1, const2, const3,
502 b0_l, b1_l, b2_l, b3_l);
504 const4, const5, const6, const7, b0_r, b1_r, b2_r, b3_r);
506 const4, const5, const6, const7, b0_l, b1_l, b2_l, b3_l);
507 BUTTERFLY_16(a0_r, a0_l, a1_r, a1_l, a2_r, a2_l, a3_r, a3_l,
508 b3_l, b3_r, b2_l, b2_r, b1_l, b1_r, b0_l, b0_r,
509 temp0_r, temp0_l, temp1_r, temp1_l,
510 temp2_r, temp2_l, temp3_r, temp3_l,
511 a3_l, a3_r, a2_l, a2_r, a1_l, a1_r, a0_l, a0_r);
512 SRA_4V(temp0_r, temp0_l, temp1_r, temp1_l, 20);
513 SRA_4V(temp2_r, temp2_l, temp3_r, temp3_l, 20);
514 LD_SH4(dst, dst_stride, in0, in1, in2, in3);
515 PCKEV_H4_SW(temp0_l, temp0_r, temp1_l, temp1_r, temp2_l, temp2_r,
516 temp3_l, temp3_r, temp0_r, temp1_r, temp2_r, temp3_r);
517 ILVR_B4_SW(zero, in0, zero, in1, zero, in2, zero, in3,
518 temp0_l, temp1_l, temp2_l, temp3_l);
519 temp0_r = (v4i32) ((v8i16) (temp0_r) + (v8i16) (temp0_l));
520 temp1_r = (v4i32) ((v8i16) (temp1_r) + (v8i16) (temp1_l));
521 temp2_r = (v4i32) ((v8i16) (temp2_r) + (v8i16) (temp2_l));
522 temp3_r = (v4i32) ((v8i16) (temp3_r) + (v8i16) (temp3_l));
528 temp2_r, temp2_r, temp3_r, temp3_r,
529 temp0_r, temp1_r, temp2_r, temp3_r);
530 tmp0 = __msa_copy_u_d((v2i64) temp0_r, 1);
531 tmp1 = __msa_copy_u_d((v2i64) temp1_r, 1);
532 tmp2 = __msa_copy_u_d((v2i64) temp2_r, 1);
533 tmp3 = __msa_copy_u_d((v2i64) temp3_r, 1);
534 SD4(tmp0, tmp1, tmp2, tmp3, dst, dst_stride);
536 SRA_4V(a3_r, a3_l, a2_r, a2_l, 20);
537 SRA_4V(a1_r, a1_l, a0_r, a0_l, 20);
538 LD_SH4(dst + 4 * dst_stride, dst_stride, in4, in5, in6, in7);
539 PCKEV_H4_SW(a0_l, a0_r, a1_l, a1_r, a2_l, a2_r, a3_l, a3_r,
540 a0_r, a1_r, a2_r, a3_r);
541 ILVR_B4_SW(zero, in4, zero, in5, zero, in6, zero, in7,
542 a3_l, a2_l, a1_l, a0_l);
543 a3_r = (v4i32) ((v8i16) (a3_r) + (v8i16) (a3_l));
544 a2_r = (v4i32) ((v8i16) (a2_r) + (v8i16) (a2_l));
545 a1_r = (v4i32) ((v8i16) (a1_r) + (v8i16) (a1_l));
546 a0_r = (v4i32) ((v8i16) (a0_r) + (v8i16) (a0_l));
552 a2_r, a2_r, a3_r, a3_r, a0_r, a1_r, a2_r, a3_r);
553 tmp0 = __msa_copy_u_d((v2i64) a3_r, 1);
554 tmp1 = __msa_copy_u_d((v2i64) a2_r, 1);
555 tmp2 = __msa_copy_u_d((v2i64) a1_r, 1);
556 tmp3 = __msa_copy_u_d((v2i64) a0_r, 1);
557 SD4(tmp0, tmp1, tmp2, tmp3, dst + 4 * dst_stride, dst_stride);
#define MUL2(in0, in1, in2, in3, out0, out1)
void ff_simple_idct_msa(int16_t *block)
#define SRA_4V(in0, in1, in2, in3, shift)
#define CLIP_SH_0_255(in)
#define SPLATI_H4_SH(...)
#define SUB4(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, out3)
static void simple_idct_msa(int16_t *block)
#define MUL4(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, out3)
#define ADD4(in0, in1, in2, in3, in4, in5, in6, in7, out0, out1, out2, out3)
#define DPADD_SH4_SW(...)
#define TRANSPOSE8x8_SH_SH(...)
#define BUTTERFLY_16(in0, in1, in2, in3, in4, in5, in6, in7,in8, in9,in10, in11, in12, in13, in14, in15,out0, out1, out2, out3, out4, out5, out6, out7,out8, out9, out10, out11, out12, out13, out14, out15)
#define BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7,out0, out1, out2, out3, out4, out5, out6, out7)
#define UNPCK_SH_SW(in, out0, out1)
static void simple_idct_add_msa(uint8_t *dst, int32_t dst_stride, int16_t *block)
#define ADD2(in0, in1, in2, in3, out0, out1)
#define SD4(in0, in1, in2, in3, pdst, stride)
void ff_simple_idct_add_msa(uint8_t *dst, ptrdiff_t dst_stride, int16_t *block)
#define SUB2(in0, in1, in2, in3, out0, out1)
static void simple_idct_put_msa(uint8_t *dst, int32_t dst_stride, int16_t *block)
void ff_simple_idct_put_msa(uint8_t *dst, ptrdiff_t dst_stride, int16_t *block)