You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
808 lines
29 KiB
808 lines
29 KiB
|
|
/* filter_msa_intrinsics.c - MSA optimised filter functions |
|
* |
|
* Copyright (c) 2018 Cosmin Truta |
|
* Copyright (c) 2016 Glenn Randers-Pehrson |
|
* Written by Mandar Sahastrabuddhe, August 2016. |
|
* |
|
* This code is released under the libpng license. |
|
* For conditions of distribution and use, see the disclaimer |
|
* and license in png.h |
|
*/ |
|
|
|
#include <stdio.h> |
|
#include <stdint.h> |
|
#include "../pngpriv.h" |
|
|
|
#ifdef PNG_READ_SUPPORTED |
|
|
|
/* This code requires -mfpu=msa on the command line: */ |
|
#if PNG_MIPS_MSA_IMPLEMENTATION == 1 /* intrinsics code from pngpriv.h */ |
|
|
|
#include <msa.h> |
|
|
|
/* libpng row pointers are not necessarily aligned to any particular boundary, |
|
* however this code will only work with appropriate alignment. mips/mips_init.c |
|
* checks for this (and will not compile unless it is done). This code uses |
|
* variants of png_aligncast to avoid compiler warnings. |
|
*/ |
|
#define png_ptr(type,pointer) png_aligncast(type *,pointer) |
|
#define png_ptrc(type,pointer) png_aligncastconst(const type *,pointer) |
|
|
|
/* The following relies on a variable 'temp_pointer' being declared with type |
|
* 'type'. This is written this way just to hide the GCC strict aliasing |
|
* warning; note that the code is safe because there never is an alias between |
|
* the input and output pointers. |
|
*/ |
|
#define png_ldr(type,pointer)\ |
|
(temp_pointer = png_ptr(type,pointer), *temp_pointer) |
|
|
|
#if PNG_MIPS_MSA_OPT > 0 |
|
|
|
#ifdef CLANG_BUILD |
|
#define MSA_SRLI_B(a, b) __msa_srli_b((v16i8) a, b) |
|
|
|
#define LW(psrc) \ |
|
( { \ |
|
uint8_t *psrc_lw_m = (uint8_t *) (psrc); \ |
|
uint32_t val_m; \ |
|
\ |
|
asm volatile ( \ |
|
"lw %[val_m], %[psrc_lw_m] \n\t" \ |
|
\ |
|
: [val_m] "=r" (val_m) \ |
|
: [psrc_lw_m] "m" (*psrc_lw_m) \ |
|
); \ |
|
\ |
|
val_m; \ |
|
} ) |
|
|
|
#define SH(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sh_m = (uint8_t *) (pdst); \ |
|
uint16_t val_m = (val); \ |
|
\ |
|
asm volatile ( \ |
|
"sh %[val_m], %[pdst_sh_m] \n\t" \ |
|
\ |
|
: [pdst_sh_m] "=m" (*pdst_sh_m) \ |
|
: [val_m] "r" (val_m) \ |
|
); \ |
|
} |
|
|
|
#define SW(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sw_m = (uint8_t *) (pdst); \ |
|
uint32_t val_m = (val); \ |
|
\ |
|
asm volatile ( \ |
|
"sw %[val_m], %[pdst_sw_m] \n\t" \ |
|
\ |
|
: [pdst_sw_m] "=m" (*pdst_sw_m) \ |
|
: [val_m] "r" (val_m) \ |
|
); \ |
|
} |
|
|
|
#if (__mips == 64) |
|
#define SD(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sd_m = (uint8_t *) (pdst); \ |
|
uint64_t val_m = (val); \ |
|
\ |
|
asm volatile ( \ |
|
"sd %[val_m], %[pdst_sd_m] \n\t" \ |
|
\ |
|
: [pdst_sd_m] "=m" (*pdst_sd_m) \ |
|
: [val_m] "r" (val_m) \ |
|
); \ |
|
} |
|
#else |
|
#define SD(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sd_m = (uint8_t *) (pdst); \ |
|
uint32_t val0_m, val1_m; \ |
|
\ |
|
val0_m = (uint32_t) ((val) & 0x00000000FFFFFFFF); \ |
|
val1_m = (uint32_t) (((val) >> 32) & 0x00000000FFFFFFFF); \ |
|
\ |
|
SW(val0_m, pdst_sd_m); \ |
|
SW(val1_m, pdst_sd_m + 4); \ |
|
} |
|
#endif |
|
#else |
|
#define MSA_SRLI_B(a, b) (a >> b) |
|
|
|
#if (__mips_isa_rev >= 6) |
|
#define LW(psrc) \ |
|
( { \ |
|
uint8_t *psrc_lw_m = (uint8_t *) (psrc); \ |
|
uint32_t val_m; \ |
|
\ |
|
asm volatile ( \ |
|
"lw %[val_m], %[psrc_lw_m] \n\t" \ |
|
\ |
|
: [val_m] "=r" (val_m) \ |
|
: [psrc_lw_m] "m" (*psrc_lw_m) \ |
|
); \ |
|
\ |
|
val_m; \ |
|
} ) |
|
|
|
#define SH(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sh_m = (uint8_t *) (pdst); \ |
|
uint16_t val_m = (val); \ |
|
\ |
|
asm volatile ( \ |
|
"sh %[val_m], %[pdst_sh_m] \n\t" \ |
|
\ |
|
: [pdst_sh_m] "=m" (*pdst_sh_m) \ |
|
: [val_m] "r" (val_m) \ |
|
); \ |
|
} |
|
|
|
#define SW(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sw_m = (uint8_t *) (pdst); \ |
|
uint32_t val_m = (val); \ |
|
\ |
|
asm volatile ( \ |
|
"sw %[val_m], %[pdst_sw_m] \n\t" \ |
|
\ |
|
: [pdst_sw_m] "=m" (*pdst_sw_m) \ |
|
: [val_m] "r" (val_m) \ |
|
); \ |
|
} |
|
|
|
#if (__mips == 64) |
|
#define SD(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sd_m = (uint8_t *) (pdst); \ |
|
uint64_t val_m = (val); \ |
|
\ |
|
asm volatile ( \ |
|
"sd %[val_m], %[pdst_sd_m] \n\t" \ |
|
\ |
|
: [pdst_sd_m] "=m" (*pdst_sd_m) \ |
|
: [val_m] "r" (val_m) \ |
|
); \ |
|
} |
|
#else |
|
#define SD(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sd_m = (uint8_t *) (pdst); \ |
|
uint32_t val0_m, val1_m; \ |
|
\ |
|
val0_m = (uint32_t) ((val) & 0x00000000FFFFFFFF); \ |
|
val1_m = (uint32_t) (((val) >> 32) & 0x00000000FFFFFFFF); \ |
|
\ |
|
SW(val0_m, pdst_sd_m); \ |
|
SW(val1_m, pdst_sd_m + 4); \ |
|
} |
|
#endif |
|
#else // !(__mips_isa_rev >= 6) |
|
#define LW(psrc) \ |
|
( { \ |
|
uint8_t *psrc_lw_m = (uint8_t *) (psrc); \ |
|
uint32_t val_m; \ |
|
\ |
|
asm volatile ( \ |
|
"ulw %[val_m], %[psrc_lw_m] \n\t" \ |
|
\ |
|
: [val_m] "=r" (val_m) \ |
|
: [psrc_lw_m] "m" (*psrc_lw_m) \ |
|
); \ |
|
\ |
|
val_m; \ |
|
} ) |
|
|
|
#define SH(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sh_m = (uint8_t *) (pdst); \ |
|
uint16_t val_m = (val); \ |
|
\ |
|
asm volatile ( \ |
|
"ush %[val_m], %[pdst_sh_m] \n\t" \ |
|
\ |
|
: [pdst_sh_m] "=m" (*pdst_sh_m) \ |
|
: [val_m] "r" (val_m) \ |
|
); \ |
|
} |
|
|
|
#define SW(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sw_m = (uint8_t *) (pdst); \ |
|
uint32_t val_m = (val); \ |
|
\ |
|
asm volatile ( \ |
|
"usw %[val_m], %[pdst_sw_m] \n\t" \ |
|
\ |
|
: [pdst_sw_m] "=m" (*pdst_sw_m) \ |
|
: [val_m] "r" (val_m) \ |
|
); \ |
|
} |
|
|
|
#define SD(val, pdst) \ |
|
{ \ |
|
uint8_t *pdst_sd_m = (uint8_t *) (pdst); \ |
|
uint32_t val0_m, val1_m; \ |
|
\ |
|
val0_m = (uint32_t) ((val) & 0x00000000FFFFFFFF); \ |
|
val1_m = (uint32_t) (((val) >> 32) & 0x00000000FFFFFFFF); \ |
|
\ |
|
SW(val0_m, pdst_sd_m); \ |
|
SW(val1_m, pdst_sd_m + 4); \ |
|
} |
|
|
|
#define SW_ZERO(pdst) \ |
|
{ \ |
|
uint8_t *pdst_m = (uint8_t *) (pdst); \ |
|
\ |
|
asm volatile ( \ |
|
"usw $0, %[pdst_m] \n\t" \ |
|
\ |
|
: [pdst_m] "=m" (*pdst_m) \ |
|
: \ |
|
); \ |
|
} |
|
#endif // (__mips_isa_rev >= 6) |
|
#endif |
|
|
|
#define LD_B(RTYPE, psrc) *((RTYPE *) (psrc)) |
|
#define LD_UB(...) LD_B(v16u8, __VA_ARGS__) |
|
#define LD_B2(RTYPE, psrc, stride, out0, out1) \ |
|
{ \ |
|
out0 = LD_B(RTYPE, (psrc)); \ |
|
out1 = LD_B(RTYPE, (psrc) + stride); \ |
|
} |
|
#define LD_UB2(...) LD_B2(v16u8, __VA_ARGS__) |
|
#define LD_B4(RTYPE, psrc, stride, out0, out1, out2, out3) \ |
|
{ \ |
|
LD_B2(RTYPE, (psrc), stride, out0, out1); \ |
|
LD_B2(RTYPE, (psrc) + 2 * stride , stride, out2, out3); \ |
|
} |
|
#define LD_UB4(...) LD_B4(v16u8, __VA_ARGS__) |
|
|
|
#define ST_B(RTYPE, in, pdst) *((RTYPE *) (pdst)) = (in) |
|
#define ST_UB(...) ST_B(v16u8, __VA_ARGS__) |
|
#define ST_B2(RTYPE, in0, in1, pdst, stride) \ |
|
{ \ |
|
ST_B(RTYPE, in0, (pdst)); \ |
|
ST_B(RTYPE, in1, (pdst) + stride); \ |
|
} |
|
#define ST_UB2(...) ST_B2(v16u8, __VA_ARGS__) |
|
#define ST_B4(RTYPE, in0, in1, in2, in3, pdst, stride) \ |
|
{ \ |
|
ST_B2(RTYPE, in0, in1, (pdst), stride); \ |
|
ST_B2(RTYPE, in2, in3, (pdst) + 2 * stride, stride); \ |
|
} |
|
#define ST_UB4(...) ST_B4(v16u8, __VA_ARGS__) |
|
|
|
#define ADD2(in0, in1, in2, in3, out0, out1) \ |
|
{ \ |
|
out0 = in0 + in1; \ |
|
out1 = in2 + in3; \ |
|
} |
|
#define ADD3(in0, in1, in2, in3, in4, in5, \ |
|
out0, out1, out2) \ |
|
{ \ |
|
ADD2(in0, in1, in2, in3, out0, out1); \ |
|
out2 = in4 + in5; \ |
|
} |
|
#define ADD4(in0, in1, in2, in3, in4, in5, in6, in7, \ |
|
out0, out1, out2, out3) \ |
|
{ \ |
|
ADD2(in0, in1, in2, in3, out0, out1); \ |
|
ADD2(in4, in5, in6, in7, out2, out3); \ |
|
} |
|
|
|
#define ILVR_B2(RTYPE, in0, in1, in2, in3, out0, out1) \ |
|
{ \ |
|
out0 = (RTYPE) __msa_ilvr_b((v16i8) in0, (v16i8) in1); \ |
|
out1 = (RTYPE) __msa_ilvr_b((v16i8) in2, (v16i8) in3); \ |
|
} |
|
#define ILVR_B2_SH(...) ILVR_B2(v8i16, __VA_ARGS__) |
|
|
|
#define HSUB_UB2(RTYPE, in0, in1, out0, out1) \ |
|
{ \ |
|
out0 = (RTYPE) __msa_hsub_u_h((v16u8) in0, (v16u8) in0); \ |
|
out1 = (RTYPE) __msa_hsub_u_h((v16u8) in1, (v16u8) in1); \ |
|
} |
|
#define HSUB_UB2_SH(...) HSUB_UB2(v8i16, __VA_ARGS__) |
|
|
|
#define SLDI_B2_0(RTYPE, in0, in1, out0, out1, slide_val) \ |
|
{ \ |
|
v16i8 zero_m = { 0 }; \ |
|
out0 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in0, slide_val); \ |
|
out1 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in1, slide_val); \ |
|
} |
|
#define SLDI_B2_0_UB(...) SLDI_B2_0(v16u8, __VA_ARGS__) |
|
|
|
#define SLDI_B3_0(RTYPE, in0, in1, in2, out0, out1, out2, slide_val) \ |
|
{ \ |
|
v16i8 zero_m = { 0 }; \ |
|
SLDI_B2_0(RTYPE, in0, in1, out0, out1, slide_val); \ |
|
out2 = (RTYPE) __msa_sldi_b((v16i8) zero_m, (v16i8) in2, slide_val); \ |
|
} |
|
#define SLDI_B3_0_UB(...) SLDI_B3_0(v16u8, __VA_ARGS__) |
|
|
|
#define ILVEV_W2(RTYPE, in0, in1, in2, in3, out0, out1) \ |
|
{ \ |
|
out0 = (RTYPE) __msa_ilvev_w((v4i32) in1, (v4i32) in0); \ |
|
out1 = (RTYPE) __msa_ilvev_w((v4i32) in3, (v4i32) in2); \ |
|
} |
|
#define ILVEV_W2_UB(...) ILVEV_W2(v16u8, __VA_ARGS__) |
|
|
|
#define ADD_ABS_H3(RTYPE, in0, in1, in2, out0, out1, out2) \ |
|
{ \ |
|
RTYPE zero = {0}; \ |
|
\ |
|
out0 = __msa_add_a_h((v8i16) zero, in0); \ |
|
out1 = __msa_add_a_h((v8i16) zero, in1); \ |
|
out2 = __msa_add_a_h((v8i16) zero, in2); \ |
|
} |
|
#define ADD_ABS_H3_SH(...) ADD_ABS_H3(v8i16, __VA_ARGS__) |
|
|
|
#define VSHF_B2(RTYPE, in0, in1, in2, in3, mask0, mask1, out0, out1) \ |
|
{ \ |
|
out0 = (RTYPE) __msa_vshf_b((v16i8) mask0, (v16i8) in1, (v16i8) in0); \ |
|
out1 = (RTYPE) __msa_vshf_b((v16i8) mask1, (v16i8) in3, (v16i8) in2); \ |
|
} |
|
#define VSHF_B2_UB(...) VSHF_B2(v16u8, __VA_ARGS__) |
|
|
|
#define CMP_AND_SELECT(inp0, inp1, inp2, inp3, inp4, inp5, out0) \ |
|
{ \ |
|
v8i16 _sel_h0, _sel_h1; \ |
|
v16u8 _sel_b0, _sel_b1; \ |
|
_sel_h0 = (v8i16) __msa_clt_u_h((v8u16) inp1, (v8u16) inp0); \ |
|
_sel_b0 = (v16u8) __msa_pckev_b((v16i8) _sel_h0, (v16i8) _sel_h0); \ |
|
inp0 = (v8i16) __msa_bmnz_v((v16u8) inp0, (v16u8) inp1, (v16u8) _sel_h0); \ |
|
inp4 = (v16u8) __msa_bmnz_v(inp3, inp4, _sel_b0); \ |
|
_sel_h1 = (v8i16) __msa_clt_u_h((v8u16) inp2, (v8u16) inp0); \ |
|
_sel_b1 = (v16u8) __msa_pckev_b((v16i8) _sel_h1, (v16i8) _sel_h1); \ |
|
inp4 = (v16u8) __msa_bmnz_v(inp4, inp5, _sel_b1); \ |
|
out0 += inp4; \ |
|
} |
|
|
|
void png_read_filter_row_up_msa(png_row_infop row_info, png_bytep row, |
|
png_const_bytep prev_row) |
|
{ |
|
size_t i, cnt, cnt16, cnt32; |
|
size_t istop = row_info->rowbytes; |
|
png_bytep rp = row; |
|
png_const_bytep pp = prev_row; |
|
v16u8 src0, src1, src2, src3, src4, src5, src6, src7; |
|
|
|
for (i = 0; i < (istop >> 6); i++) |
|
{ |
|
LD_UB4(rp, 16, src0, src1, src2, src3); |
|
LD_UB4(pp, 16, src4, src5, src6, src7); |
|
pp += 64; |
|
|
|
ADD4(src0, src4, src1, src5, src2, src6, src3, src7, |
|
src0, src1, src2, src3); |
|
|
|
ST_UB4(src0, src1, src2, src3, rp, 16); |
|
rp += 64; |
|
} |
|
|
|
if (istop & 0x3F) |
|
{ |
|
cnt32 = istop & 0x20; |
|
cnt16 = istop & 0x10; |
|
cnt = istop & 0xF; |
|
|
|
if(cnt32) |
|
{ |
|
if (cnt16 && cnt) |
|
{ |
|
LD_UB4(rp, 16, src0, src1, src2, src3); |
|
LD_UB4(pp, 16, src4, src5, src6, src7); |
|
|
|
ADD4(src0, src4, src1, src5, src2, src6, src3, src7, |
|
src0, src1, src2, src3); |
|
|
|
ST_UB4(src0, src1, src2, src3, rp, 16); |
|
rp += 64; |
|
} |
|
else if (cnt16 || cnt) |
|
{ |
|
LD_UB2(rp, 16, src0, src1); |
|
LD_UB2(pp, 16, src4, src5); |
|
pp += 32; |
|
src2 = LD_UB(rp + 32); |
|
src6 = LD_UB(pp); |
|
|
|
ADD3(src0, src4, src1, src5, src2, src6, src0, src1, src2); |
|
|
|
ST_UB2(src0, src1, rp, 16); |
|
rp += 32; |
|
ST_UB(src2, rp); |
|
rp += 16; |
|
} |
|
else |
|
{ |
|
LD_UB2(rp, 16, src0, src1); |
|
LD_UB2(pp, 16, src4, src5); |
|
|
|
ADD2(src0, src4, src1, src5, src0, src1); |
|
|
|
ST_UB2(src0, src1, rp, 16); |
|
rp += 32; |
|
} |
|
} |
|
else if (cnt16 && cnt) |
|
{ |
|
LD_UB2(rp, 16, src0, src1); |
|
LD_UB2(pp, 16, src4, src5); |
|
|
|
ADD2(src0, src4, src1, src5, src0, src1); |
|
|
|
ST_UB2(src0, src1, rp, 16); |
|
rp += 32; |
|
} |
|
else if (cnt16 || cnt) |
|
{ |
|
src0 = LD_UB(rp); |
|
src4 = LD_UB(pp); |
|
pp += 16; |
|
|
|
src0 += src4; |
|
|
|
ST_UB(src0, rp); |
|
rp += 16; |
|
} |
|
} |
|
} |
|
|
|
void png_read_filter_row_sub4_msa(png_row_infop row_info, png_bytep row, |
|
png_const_bytep prev_row) |
|
{ |
|
size_t count; |
|
size_t istop = row_info->rowbytes; |
|
png_bytep src = row; |
|
png_bytep nxt = row + 4; |
|
int32_t inp0; |
|
v16u8 src0, src1, src2, src3, src4; |
|
v16u8 dst0, dst1; |
|
v16u8 zero = { 0 }; |
|
|
|
istop -= 4; |
|
|
|
inp0 = LW(src); |
|
src += 4; |
|
src0 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp0); |
|
|
|
for (count = 0; count < istop; count += 16) |
|
{ |
|
src1 = LD_UB(src); |
|
src += 16; |
|
|
|
src2 = (v16u8) __msa_sldi_b((v16i8) zero, (v16i8) src1, 4); |
|
src3 = (v16u8) __msa_sldi_b((v16i8) zero, (v16i8) src1, 8); |
|
src4 = (v16u8) __msa_sldi_b((v16i8) zero, (v16i8) src1, 12); |
|
src1 += src0; |
|
src2 += src1; |
|
src3 += src2; |
|
src4 += src3; |
|
src0 = src4; |
|
ILVEV_W2_UB(src1, src2, src3, src4, dst0, dst1); |
|
dst0 = (v16u8) __msa_pckev_d((v2i64) dst1, (v2i64) dst0); |
|
|
|
ST_UB(dst0, nxt); |
|
nxt += 16; |
|
} |
|
} |
|
|
|
void png_read_filter_row_sub3_msa(png_row_infop row_info, png_bytep row, |
|
png_const_bytep prev_row) |
|
{ |
|
size_t count; |
|
size_t istop = row_info->rowbytes; |
|
png_bytep src = row; |
|
png_bytep nxt = row + 3; |
|
int64_t out0; |
|
int32_t inp0, out1; |
|
v16u8 src0, src1, src2, src3, src4, dst0, dst1; |
|
v16u8 zero = { 0 }; |
|
v16i8 mask0 = { 0, 1, 2, 16, 17, 18, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 }; |
|
v16i8 mask1 = { 0, 1, 2, 3, 4, 5, 16, 17, 18, 19, 20, 21, 0, 0, 0, 0 }; |
|
|
|
istop -= 3; |
|
|
|
inp0 = LW(src); |
|
src += 3; |
|
src0 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp0); |
|
|
|
for (count = 0; count < istop; count += 12) |
|
{ |
|
src1 = LD_UB(src); |
|
src += 12; |
|
|
|
src2 = (v16u8) __msa_sldi_b((v16i8) zero, (v16i8) src1, 3); |
|
src3 = (v16u8) __msa_sldi_b((v16i8) zero, (v16i8) src1, 6); |
|
src4 = (v16u8) __msa_sldi_b((v16i8) zero, (v16i8) src1, 9); |
|
src1 += src0; |
|
src2 += src1; |
|
src3 += src2; |
|
src4 += src3; |
|
src0 = src4; |
|
VSHF_B2_UB(src1, src2, src3, src4, mask0, mask0, dst0, dst1); |
|
dst0 = (v16u8) __msa_vshf_b(mask1, (v16i8) dst1, (v16i8) dst0); |
|
out0 = __msa_copy_s_d((v2i64) dst0, 0); |
|
out1 = __msa_copy_s_w((v4i32) dst0, 2); |
|
|
|
SD(out0, nxt); |
|
nxt += 8; |
|
SW(out1, nxt); |
|
nxt += 4; |
|
} |
|
} |
|
|
|
void png_read_filter_row_avg4_msa(png_row_infop row_info, png_bytep row, |
|
png_const_bytep prev_row) |
|
{ |
|
size_t i; |
|
png_bytep src = row; |
|
png_bytep nxt = row; |
|
png_const_bytep pp = prev_row; |
|
size_t istop = row_info->rowbytes - 4; |
|
int32_t inp0, inp1, out0; |
|
v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, dst0, dst1; |
|
v16u8 zero = { 0 }; |
|
|
|
inp0 = LW(pp); |
|
pp += 4; |
|
inp1 = LW(src); |
|
src += 4; |
|
src0 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp0); |
|
src1 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp1); |
|
src0 = (v16u8) MSA_SRLI_B(src0, 1); |
|
src1 += src0; |
|
out0 = __msa_copy_s_w((v4i32) src1, 0); |
|
SW(out0, nxt); |
|
nxt += 4; |
|
|
|
for (i = 0; i < istop; i += 16) |
|
{ |
|
src2 = LD_UB(pp); |
|
pp += 16; |
|
src6 = LD_UB(src); |
|
src += 16; |
|
|
|
SLDI_B2_0_UB(src2, src6, src3, src7, 4); |
|
SLDI_B2_0_UB(src2, src6, src4, src8, 8); |
|
SLDI_B2_0_UB(src2, src6, src5, src9, 12); |
|
src2 = __msa_ave_u_b(src2, src1); |
|
src6 += src2; |
|
src3 = __msa_ave_u_b(src3, src6); |
|
src7 += src3; |
|
src4 = __msa_ave_u_b(src4, src7); |
|
src8 += src4; |
|
src5 = __msa_ave_u_b(src5, src8); |
|
src9 += src5; |
|
src1 = src9; |
|
ILVEV_W2_UB(src6, src7, src8, src9, dst0, dst1); |
|
dst0 = (v16u8) __msa_pckev_d((v2i64) dst1, (v2i64) dst0); |
|
|
|
ST_UB(dst0, nxt); |
|
nxt += 16; |
|
} |
|
} |
|
|
|
void png_read_filter_row_avg3_msa(png_row_infop row_info, png_bytep row, |
|
png_const_bytep prev_row) |
|
{ |
|
size_t i; |
|
png_bytep src = row; |
|
png_bytep nxt = row; |
|
png_const_bytep pp = prev_row; |
|
size_t istop = row_info->rowbytes - 3; |
|
int64_t out0; |
|
int32_t inp0, inp1, out1; |
|
int16_t out2; |
|
v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, dst0, dst1; |
|
v16u8 zero = { 0 }; |
|
v16i8 mask0 = { 0, 1, 2, 16, 17, 18, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 }; |
|
v16i8 mask1 = { 0, 1, 2, 3, 4, 5, 16, 17, 18, 19, 20, 21, 0, 0, 0, 0 }; |
|
|
|
inp0 = LW(pp); |
|
pp += 3; |
|
inp1 = LW(src); |
|
src += 3; |
|
src0 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp0); |
|
src1 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp1); |
|
src0 = (v16u8) MSA_SRLI_B(src0, 1); |
|
src1 += src0; |
|
out2 = __msa_copy_s_h((v8i16) src1, 0); |
|
SH(out2, nxt); |
|
nxt += 2; |
|
nxt[0] = src1[2]; |
|
nxt++; |
|
|
|
for (i = 0; i < istop; i += 12) |
|
{ |
|
src2 = LD_UB(pp); |
|
pp += 12; |
|
src6 = LD_UB(src); |
|
src += 12; |
|
|
|
SLDI_B2_0_UB(src2, src6, src3, src7, 3); |
|
SLDI_B2_0_UB(src2, src6, src4, src8, 6); |
|
SLDI_B2_0_UB(src2, src6, src5, src9, 9); |
|
src2 = __msa_ave_u_b(src2, src1); |
|
src6 += src2; |
|
src3 = __msa_ave_u_b(src3, src6); |
|
src7 += src3; |
|
src4 = __msa_ave_u_b(src4, src7); |
|
src8 += src4; |
|
src5 = __msa_ave_u_b(src5, src8); |
|
src9 += src5; |
|
src1 = src9; |
|
VSHF_B2_UB(src6, src7, src8, src9, mask0, mask0, dst0, dst1); |
|
dst0 = (v16u8) __msa_vshf_b(mask1, (v16i8) dst1, (v16i8) dst0); |
|
out0 = __msa_copy_s_d((v2i64) dst0, 0); |
|
out1 = __msa_copy_s_w((v4i32) dst0, 2); |
|
|
|
SD(out0, nxt); |
|
nxt += 8; |
|
SW(out1, nxt); |
|
nxt += 4; |
|
} |
|
} |
|
|
|
void png_read_filter_row_paeth4_msa(png_row_infop row_info, |
|
png_bytep row, |
|
png_const_bytep prev_row) |
|
{ |
|
int32_t count, rp_end; |
|
png_bytep nxt; |
|
png_const_bytep prev_nxt; |
|
int32_t inp0, inp1, res0; |
|
v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9; |
|
v16u8 src10, src11, src12, src13, dst0, dst1; |
|
v8i16 vec0, vec1, vec2; |
|
v16u8 zero = { 0 }; |
|
|
|
nxt = row; |
|
prev_nxt = prev_row; |
|
|
|
inp0 = LW(nxt); |
|
inp1 = LW(prev_nxt); |
|
prev_nxt += 4; |
|
src0 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp0); |
|
src1 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp1); |
|
|
|
src1 += src0; |
|
res0 = __msa_copy_s_w((v4i32) src1, 0); |
|
|
|
SW(res0, nxt); |
|
nxt += 4; |
|
|
|
/* Remainder */ |
|
rp_end = row_info->rowbytes - 4; |
|
|
|
for (count = 0; count < rp_end; count += 16) |
|
{ |
|
src2 = LD_UB(prev_nxt); |
|
prev_nxt += 16; |
|
src6 = LD_UB(prev_row); |
|
prev_row += 16; |
|
src10 = LD_UB(nxt); |
|
|
|
SLDI_B3_0_UB(src2, src6, src10, src3, src7, src11, 4); |
|
SLDI_B3_0_UB(src2, src6, src10, src4, src8, src12, 8); |
|
SLDI_B3_0_UB(src2, src6, src10, src5, src9, src13, 12); |
|
ILVR_B2_SH(src2, src6, src1, src6, vec0, vec1); |
|
HSUB_UB2_SH(vec0, vec1, vec0, vec1); |
|
vec2 = vec0 + vec1; |
|
ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); |
|
CMP_AND_SELECT(vec0, vec1, vec2, src1, src2, src6, src10); |
|
ILVR_B2_SH(src3, src7, src10, src7, vec0, vec1); |
|
HSUB_UB2_SH(vec0, vec1, vec0, vec1); |
|
vec2 = vec0 + vec1; |
|
ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); |
|
CMP_AND_SELECT(vec0, vec1, vec2, src10, src3, src7, src11); |
|
ILVR_B2_SH(src4, src8, src11, src8, vec0, vec1); |
|
HSUB_UB2_SH(vec0, vec1, vec0, vec1); |
|
vec2 = vec0 + vec1; |
|
ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); |
|
CMP_AND_SELECT(vec0, vec1, vec2, src11, src4, src8, src12); |
|
ILVR_B2_SH(src5, src9, src12, src9, vec0, vec1); |
|
HSUB_UB2_SH(vec0, vec1, vec0, vec1); |
|
vec2 = vec0 + vec1; |
|
ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); |
|
CMP_AND_SELECT(vec0, vec1, vec2, src12, src5, src9, src13); |
|
src1 = src13; |
|
ILVEV_W2_UB(src10, src11, src12, src1, dst0, dst1); |
|
dst0 = (v16u8) __msa_pckev_d((v2i64) dst1, (v2i64) dst0); |
|
|
|
ST_UB(dst0, nxt); |
|
nxt += 16; |
|
} |
|
} |
|
|
|
void png_read_filter_row_paeth3_msa(png_row_infop row_info, |
|
png_bytep row, |
|
png_const_bytep prev_row) |
|
{ |
|
int32_t count, rp_end; |
|
png_bytep nxt; |
|
png_const_bytep prev_nxt; |
|
int64_t out0; |
|
int32_t inp0, inp1, out1; |
|
int16_t out2; |
|
v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, dst0, dst1; |
|
v16u8 src10, src11, src12, src13; |
|
v8i16 vec0, vec1, vec2; |
|
v16u8 zero = { 0 }; |
|
v16i8 mask0 = { 0, 1, 2, 16, 17, 18, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 }; |
|
v16i8 mask1 = { 0, 1, 2, 3, 4, 5, 16, 17, 18, 19, 20, 21, 0, 0, 0, 0 }; |
|
|
|
nxt = row; |
|
prev_nxt = prev_row; |
|
|
|
inp0 = LW(nxt); |
|
inp1 = LW(prev_nxt); |
|
prev_nxt += 3; |
|
src0 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp0); |
|
src1 = (v16u8) __msa_insert_w((v4i32) zero, 0, inp1); |
|
|
|
src1 += src0; |
|
out2 = __msa_copy_s_h((v8i16) src1, 0); |
|
|
|
SH(out2, nxt); |
|
nxt += 2; |
|
nxt[0] = src1[2]; |
|
nxt++; |
|
|
|
/* Remainder */ |
|
rp_end = row_info->rowbytes - 3; |
|
|
|
for (count = 0; count < rp_end; count += 12) |
|
{ |
|
src2 = LD_UB(prev_nxt); |
|
prev_nxt += 12; |
|
src6 = LD_UB(prev_row); |
|
prev_row += 12; |
|
src10 = LD_UB(nxt); |
|
|
|
SLDI_B3_0_UB(src2, src6, src10, src3, src7, src11, 3); |
|
SLDI_B3_0_UB(src2, src6, src10, src4, src8, src12, 6); |
|
SLDI_B3_0_UB(src2, src6, src10, src5, src9, src13, 9); |
|
ILVR_B2_SH(src2, src6, src1, src6, vec0, vec1); |
|
HSUB_UB2_SH(vec0, vec1, vec0, vec1); |
|
vec2 = vec0 + vec1; |
|
ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); |
|
CMP_AND_SELECT(vec0, vec1, vec2, src1, src2, src6, src10); |
|
ILVR_B2_SH(src3, src7, src10, src7, vec0, vec1); |
|
HSUB_UB2_SH(vec0, vec1, vec0, vec1); |
|
vec2 = vec0 + vec1; |
|
ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); |
|
CMP_AND_SELECT(vec0, vec1, vec2, src10, src3, src7, src11); |
|
ILVR_B2_SH(src4, src8, src11, src8, vec0, vec1); |
|
HSUB_UB2_SH(vec0, vec1, vec0, vec1); |
|
vec2 = vec0 + vec1; |
|
ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); |
|
CMP_AND_SELECT(vec0, vec1, vec2, src11, src4, src8, src12); |
|
ILVR_B2_SH(src5, src9, src12, src9, vec0, vec1); |
|
HSUB_UB2_SH(vec0, vec1, vec0, vec1); |
|
vec2 = vec0 + vec1; |
|
ADD_ABS_H3_SH(vec0, vec1, vec2, vec0, vec1, vec2); |
|
CMP_AND_SELECT(vec0, vec1, vec2, src12, src5, src9, src13); |
|
src1 = src13; |
|
VSHF_B2_UB(src10, src11, src12, src13, mask0, mask0, dst0, dst1); |
|
dst0 = (v16u8) __msa_vshf_b(mask1, (v16i8) dst1, (v16i8) dst0); |
|
out0 = __msa_copy_s_d((v2i64) dst0, 0); |
|
out1 = __msa_copy_s_w((v4i32) dst0, 2); |
|
|
|
SD(out0, nxt); |
|
nxt += 8; |
|
SW(out1, nxt); |
|
nxt += 4; |
|
} |
|
} |
|
|
|
#endif /* PNG_MIPS_MSA_OPT > 0 */ |
|
#endif /* PNG_MIPS_MSA_IMPLEMENTATION == 1 (intrinsics) */ |
|
#endif /* READ */
|
|
|