| /* |
| * Loongson SIMD optimized h264chroma |
| * |
| * Copyright (c) 2015 Loongson Technology Corporation Limited |
| * Copyright (c) 2015 Zhou Xiaoyong <zhouxiaoyong@loongson.cn> |
| * Zhang Shuangshuang <zhangshuangshuang@ict.ac.cn> |
| * |
| * This file is part of FFmpeg. |
| * |
| * FFmpeg is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2.1 of the License, or (at your option) any later version. |
| * |
| * FFmpeg is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with FFmpeg; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| |
| #include "h264chroma_mips.h" |
| #include "constants.h" |
| #include "libavutil/mips/mmiutils.h" |
| |
| void ff_put_h264_chroma_mc8_mmi(uint8_t *dst, const uint8_t *src, ptrdiff_t stride, |
| int h, int x, int y) |
| { |
| double ftmp[12]; |
| union mmi_intfloat64 A, B, C, D, E; |
| DECLARE_VAR_ALL64; |
| |
| A.i = 64; |
| |
| if (!(x || y)) { |
| /* x=0, y=0, A.i=64 */ |
| __asm__ volatile ( |
| "1: \n\t" |
| MMI_ULDC1(%[ftmp0], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp1], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp2], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp3], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| |
| "addi %[h], %[h], -0x04 \n\t" |
| |
| MMI_SDC1(%[ftmp0], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| MMI_SDC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| MMI_SDC1(%[ftmp2], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| MMI_SDC1(%[ftmp3], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : RESTRICT_ASM_ALL64 |
| [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride) |
| : "memory" |
| ); |
| } else if (x && y) { |
| /* x!=0, y!=0 */ |
| D.i = x * y; |
| B.i = (x << 3) - D.i; |
| C.i = (y << 3) - D.i; |
| A.i = 64 - D.i - B.i - C.i; |
| |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[B], %[B], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp9] \n\t" |
| "pshufh %[C], %[C], %[ftmp0] \n\t" |
| "pshufh %[D], %[D], %[ftmp0] \n\t" |
| |
| "1: \n\t" |
| MMI_ULDC1(%[ftmp1], %[src], 0x00) |
| MMI_ULDC1(%[ftmp2], %[src], 0x01) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp3], %[src], 0x00) |
| MMI_ULDC1(%[ftmp4], %[src], 0x01) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp10], %[src], 0x00) |
| MMI_ULDC1(%[ftmp11], %[src], 0x01) |
| "addi %[h], %[h], -0x02 \n\t" |
| |
| "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp7], %[ftmp2], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp8], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[A] \n\t" |
| "pmullh %[ftmp7], %[ftmp7], %[B] \n\t" |
| "paddh %[ftmp1], %[ftmp5], %[ftmp7] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[A] \n\t" |
| "pmullh %[ftmp8], %[ftmp8], %[B] \n\t" |
| "paddh %[ftmp2], %[ftmp6], %[ftmp8] \n\t" |
| "punpcklbh %[ftmp5], %[ftmp3], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp3], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp7], %[ftmp4], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp8], %[ftmp4], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[C] \n\t" |
| "pmullh %[ftmp7], %[ftmp7], %[D] \n\t" |
| "paddh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[C] \n\t" |
| "pmullh %[ftmp8], %[ftmp8], %[D] \n\t" |
| "paddh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" |
| "paddh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" |
| "paddh %[ftmp2], %[ftmp2], %[ftmp6] \n\t" |
| "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" |
| "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp1], %[ftmp1], %[ftmp9] \n\t" |
| "psrlh %[ftmp2], %[ftmp2], %[ftmp9] \n\t" |
| "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| |
| "punpcklbh %[ftmp5], %[ftmp3], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp3], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp7], %[ftmp4], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp8], %[ftmp4], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[A] \n\t" |
| "pmullh %[ftmp7], %[ftmp7], %[B] \n\t" |
| "paddh %[ftmp3], %[ftmp5], %[ftmp7] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[A] \n\t" |
| "pmullh %[ftmp8], %[ftmp8], %[B] \n\t" |
| "paddh %[ftmp4], %[ftmp6], %[ftmp8] \n\t" |
| "punpcklbh %[ftmp5], %[ftmp10], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp10], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp7], %[ftmp11], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp8], %[ftmp11], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[C] \n\t" |
| "pmullh %[ftmp7], %[ftmp7], %[D] \n\t" |
| "paddh %[ftmp5], %[ftmp5], %[ftmp7] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[C] \n\t" |
| "pmullh %[ftmp8], %[ftmp8], %[D] \n\t" |
| "paddh %[ftmp6], %[ftmp6], %[ftmp8] \n\t" |
| "paddh %[ftmp3], %[ftmp3], %[ftmp5] \n\t" |
| "paddh %[ftmp4], %[ftmp4], %[ftmp6] \n\t" |
| "paddh %[ftmp3], %[ftmp3], %[ff_pw_32] \n\t" |
| "paddh %[ftmp4], %[ftmp4], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp3], %[ftmp3], %[ftmp9] \n\t" |
| "psrlh %[ftmp4], %[ftmp4], %[ftmp9] \n\t" |
| "packushb %[ftmp3], %[ftmp3], %[ftmp4] \n\t" |
| |
| MMI_SDC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| MMI_SDC1(%[ftmp3], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : RESTRICT_ASM_ALL64 |
| [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), |
| [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), |
| [ftmp10]"=&f"(ftmp[10]), [ftmp11]"=&f"(ftmp[11]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32.f), |
| [A]"f"(A.f), [B]"f"(B.f), |
| [C]"f"(C.f), [D]"f"(D.f), |
| [tmp0]"r"(0x06) |
| : "memory" |
| ); |
| } else if (x) { |
| /* x!=0, y==0 */ |
| E.i = x << 3; |
| A.i = 64 - E.i; |
| |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[E], %[E], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp7] \n\t" |
| |
| "1: \n\t" |
| MMI_ULDC1(%[ftmp1], %[src], 0x00) |
| MMI_ULDC1(%[ftmp2], %[src], 0x01) |
| "addi %[h], %[h], -0x01 \n\t" |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| |
| "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp3], %[ftmp3], %[A] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[E] \n\t" |
| "paddh %[ftmp1], %[ftmp3], %[ftmp5] \n\t" |
| "pmullh %[ftmp4], %[ftmp4], %[A] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[E] \n\t" |
| "paddh %[ftmp2], %[ftmp4], %[ftmp6] \n\t" |
| |
| "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" |
| "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" |
| "psrlh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" |
| "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| MMI_SDC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : RESTRICT_ASM_ALL64 |
| [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride), |
| [ff_pw_32]"f"(ff_pw_32.f), [tmp0]"r"(0x06), |
| [A]"f"(A.f), [E]"f"(E.f) |
| : "memory" |
| ); |
| } else { |
| /* x==0, y!=0 */ |
| E.i = y << 3; |
| A.i = 64 - E.i; |
| |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[E], %[E], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp7] \n\t" |
| |
| "1: \n\t" |
| MMI_ULDC1(%[ftmp1], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp2], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp8], %[src], 0x00) |
| "addi %[h], %[h], -0x02 \n\t" |
| |
| "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp3], %[ftmp3], %[A] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[E] \n\t" |
| "paddh %[ftmp3], %[ftmp3], %[ftmp5] \n\t" |
| "pmullh %[ftmp4], %[ftmp4], %[A] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[E] \n\t" |
| "paddh %[ftmp4], %[ftmp4], %[ftmp6] \n\t" |
| "paddh %[ftmp3], %[ftmp3], %[ff_pw_32] \n\t" |
| "paddh %[ftmp4], %[ftmp4], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" |
| "psrlh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" |
| "packushb %[ftmp1], %[ftmp3], %[ftmp4] \n\t" |
| |
| "punpcklbh %[ftmp3], %[ftmp2], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp4], %[ftmp2], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp5], %[ftmp8], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp8], %[ftmp0] \n\t" |
| "pmullh %[ftmp3], %[ftmp3], %[A] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[E] \n\t" |
| "paddh %[ftmp3], %[ftmp3], %[ftmp5] \n\t" |
| "pmullh %[ftmp4], %[ftmp4], %[A] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[E] \n\t" |
| "paddh %[ftmp4], %[ftmp4], %[ftmp6] \n\t" |
| "paddh %[ftmp3], %[ftmp3], %[ff_pw_32] \n\t" |
| "paddh %[ftmp4], %[ftmp4], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp3], %[ftmp3], %[ftmp7] \n\t" |
| "psrlh %[ftmp4], %[ftmp4], %[ftmp7] \n\t" |
| "packushb %[ftmp2], %[ftmp3], %[ftmp4] \n\t" |
| |
| MMI_SDC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| MMI_SDC1(%[ftmp2], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : RESTRICT_ASM_ALL64 |
| [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), |
| [ftmp8]"=&f"(ftmp[8]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride), |
| [ff_pw_32]"f"(ff_pw_32.f), [A]"f"(A.f), |
| [E]"f"(E.f), [tmp0]"r"(0x06) |
| : "memory" |
| ); |
| } |
| } |
| |
| void ff_avg_h264_chroma_mc8_mmi(uint8_t *dst, const uint8_t *src, ptrdiff_t stride, |
| int h, int x, int y) |
| { |
| double ftmp[10]; |
| union mmi_intfloat64 A, B, C, D, E; |
| DECLARE_VAR_ALL64; |
| |
| A.i = 64; |
| |
| if(!(x || y)){ |
| /* x=0, y=0, A.i=64 */ |
| __asm__ volatile ( |
| "1: \n\t" |
| MMI_ULDC1(%[ftmp0], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp1], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_LDC1(%[ftmp2], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| MMI_LDC1(%[ftmp3], %[dst], 0x00) |
| PTR_SUBU "%[dst], %[dst], %[stride] \n\t" |
| "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" |
| "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" |
| MMI_SDC1(%[ftmp0], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| MMI_SDC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "addi %[h], %[h], -0x02 \n\t" |
| "bnez %[h], 1b \n\t" |
| : RESTRICT_ASM_ALL64 |
| [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride) |
| : "memory" |
| ); |
| } else if (x && y) { |
| /* x!=0, y!=0 */ |
| D.i = x * y; |
| B.i = (x << 3) - D.i; |
| C.i = (y << 3) - D.i; |
| A.i = 64 - D.i - B.i - C.i; |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[B], %[B], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp9] \n\t" |
| "pshufh %[C], %[C], %[ftmp0] \n\t" |
| "pshufh %[D], %[D], %[ftmp0] \n\t" |
| |
| "1: \n\t" |
| MMI_ULDC1(%[ftmp1], %[src], 0x00) |
| MMI_ULDC1(%[ftmp2], %[src], 0x01) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp3], %[src], 0x00) |
| MMI_ULDC1(%[ftmp4], %[src], 0x01) |
| "addi %[h], %[h], -0x01 \n\t" |
| |
| "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp7], %[ftmp2], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp8], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[A] \n\t" |
| "pmullh %[ftmp7], %[ftmp7], %[B] \n\t" |
| "paddh %[ftmp1], %[ftmp5], %[ftmp7] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[A] \n\t" |
| "pmullh %[ftmp8], %[ftmp8], %[B] \n\t" |
| "paddh %[ftmp2], %[ftmp6], %[ftmp8] \n\t" |
| |
| "punpcklbh %[ftmp5], %[ftmp3], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp3], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp7], %[ftmp4], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp8], %[ftmp4], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[C] \n\t" |
| "pmullh %[ftmp7], %[ftmp7], %[D] \n\t" |
| "paddh %[ftmp3], %[ftmp5], %[ftmp7] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[C] \n\t" |
| "pmullh %[ftmp8], %[ftmp8], %[D] \n\t" |
| "paddh %[ftmp4], %[ftmp6], %[ftmp8] \n\t" |
| |
| "paddh %[ftmp1], %[ftmp1], %[ftmp3] \n\t" |
| "paddh %[ftmp2], %[ftmp2], %[ftmp4] \n\t" |
| "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" |
| "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp1], %[ftmp1], %[ftmp9] \n\t" |
| "psrlh %[ftmp2], %[ftmp2], %[ftmp9] \n\t" |
| "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| MMI_LDC1(%[ftmp2], %[dst], 0x00) |
| "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| MMI_SDC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : RESTRICT_ASM_ALL64 |
| [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), |
| [ftmp8]"=&f"(ftmp[8]), [ftmp9]"=&f"(ftmp[9]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32.f), |
| [A]"f"(A.f), [B]"f"(B.f), |
| [C]"f"(C.f), [D]"f"(D.f), |
| [tmp0]"r"(0x06) |
| : "memory" |
| ); |
| } else if (x) { |
| /* x!=0, y==0 */ |
| E.i = x << 3; |
| A.i = 64 - E.i; |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[E], %[E], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp7] \n\t" |
| |
| "1: \n\t" |
| MMI_ULDC1(%[ftmp1], %[src], 0x00) |
| MMI_ULDC1(%[ftmp2], %[src], 0x01) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| "addi %[h], %[h], -0x01 \n\t" |
| |
| "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp3], %[ftmp3], %[A] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[E] \n\t" |
| "paddh %[ftmp1], %[ftmp3], %[ftmp5] \n\t" |
| "pmullh %[ftmp4], %[ftmp4], %[A] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[E] \n\t" |
| "paddh %[ftmp2], %[ftmp4], %[ftmp6] \n\t" |
| |
| "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" |
| "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" |
| "psrlh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" |
| "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| MMI_LDC1(%[ftmp2], %[dst], 0x00) |
| "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| MMI_SDC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : RESTRICT_ASM_ALL64 |
| [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride), |
| [ff_pw_32]"f"(ff_pw_32.f), [tmp0]"r"(0x06), |
| [A]"f"(A.f), [E]"f"(E.f) |
| : "memory" |
| ); |
| } else { |
| /* x==0, y!=0 */ |
| E.i = y << 3; |
| A.i = 64 - E.i; |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[E], %[E], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp7] \n\t" |
| |
| "1: \n\t" |
| MMI_ULDC1(%[ftmp1], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULDC1(%[ftmp2], %[src], 0x00) |
| "addi %[h], %[h], -0x01 \n\t" |
| |
| "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp4], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp5], %[ftmp2], %[ftmp0] \n\t" |
| "punpckhbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp3], %[ftmp3], %[A] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[E] \n\t" |
| "paddh %[ftmp1], %[ftmp3], %[ftmp5] \n\t" |
| "pmullh %[ftmp4], %[ftmp4], %[A] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[E] \n\t" |
| "paddh %[ftmp2], %[ftmp4], %[ftmp6] \n\t" |
| |
| "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" |
| "paddh %[ftmp2], %[ftmp2], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" |
| "psrlh %[ftmp2], %[ftmp2], %[ftmp7] \n\t" |
| "packushb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| MMI_LDC1(%[ftmp2], %[dst], 0x00) |
| "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| MMI_SDC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : RESTRICT_ASM_ALL64 |
| [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride), |
| [ff_pw_32]"f"(ff_pw_32.f), [tmp0]"r"(0x06), |
| [A]"f"(A.f), [E]"f"(E.f) |
| : "memory" |
| ); |
| } |
| } |
| |
| void ff_put_h264_chroma_mc4_mmi(uint8_t *dst, const uint8_t *src, ptrdiff_t stride, |
| int h, int x, int y) |
| { |
| double ftmp[8]; |
| mips_reg addr[1]; |
| union mmi_intfloat64 A, B, C, D, E; |
| DECLARE_VAR_LOW32; |
| A.i = (8 - x) * (8 - y); |
| B.i = x * (8 - y); |
| C.i = (8 - x) * y; |
| D.i = x * y; |
| E.i = B.i + C.i; |
| |
| if (D.i) { |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[B], %[B], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp7] \n\t" |
| "pshufh %[C], %[C], %[ftmp0] \n\t" |
| "pshufh %[D], %[D], %[ftmp0] \n\t" |
| |
| "1: \n\t" |
| MMI_ULWC1(%[ftmp1], %[src], 0x00) |
| MMI_ULWC1(%[ftmp2], %[src], 0x01) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULWC1(%[ftmp3], %[src], 0x00) |
| MMI_ULWC1(%[ftmp4], %[src], 0x01) |
| |
| "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[A] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[B] \n\t" |
| "paddh %[ftmp1], %[ftmp5], %[ftmp6] \n\t" |
| "punpcklbh %[ftmp5], %[ftmp3], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp6], %[ftmp4], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[C] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[D] \n\t" |
| "paddh %[ftmp2], %[ftmp5], %[ftmp6] \n\t" |
| "paddh %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" |
| "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" |
| |
| "addi %[h], %[h], -0x01 \n\t" |
| MMI_SWC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), |
| RESTRICT_ASM_LOW32 |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32.f), |
| [A]"f"(A.f), [B]"f"(B.f), |
| [C]"f"(C.f), [D]"f"(D.f), |
| [tmp0]"r"(0x06) |
| : "memory" |
| ); |
| } else if (E.i) { |
| const int step = C.i ? stride : 1; |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[E], %[E], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp5] \n\t" |
| |
| "1: \n\t" |
| MMI_ULWC1(%[ftmp1], %[src], 0x00) |
| PTR_ADDU "%[addr0], %[src], %[step] \n\t" |
| MMI_ULWC1(%[ftmp2], %[addr0], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| "addi %[h], %[h], -0x01 \n\t" |
| "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp4], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp3], %[ftmp3], %[A] \n\t" |
| "pmullh %[ftmp4], %[ftmp4], %[E] \n\t" |
| "paddh %[ftmp1], %[ftmp3], %[ftmp4] \n\t" |
| "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" |
| "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" |
| MMI_SWC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| RESTRICT_ASM_LOW32 |
| [addr0]"=&r"(addr[0]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride),[step]"r"((mips_reg)step), |
| [ff_pw_32]"f"(ff_pw_32.f), [tmp0]"r"(0x06), |
| [A]"f"(A.f), [E]"f"(E.f) |
| : "memory" |
| ); |
| } else { |
| __asm__ volatile ( |
| "1: \n\t" |
| MMI_ULWC1(%[ftmp0], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULWC1(%[ftmp1], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| "addi %[h], %[h], -0x02 \n\t" |
| MMI_SWC1(%[ftmp0], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| MMI_SWC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| RESTRICT_ASM_LOW32 |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride) |
| : "memory" |
| ); |
| } |
| } |
| |
| void ff_avg_h264_chroma_mc4_mmi(uint8_t *dst, const uint8_t *src, ptrdiff_t stride, |
| int h, int x, int y) |
| { |
| double ftmp[8]; |
| mips_reg addr[1]; |
| union mmi_intfloat64 A, B, C, D, E; |
| DECLARE_VAR_LOW32; |
| A.i = (8 - x) *(8 - y); |
| B.i = x * (8 - y); |
| C.i = (8 - x) * y; |
| D.i = x * y; |
| E.i = B.i + C.i; |
| |
| if (D.i) { |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[B], %[B], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp7] \n\t" |
| "pshufh %[C], %[C], %[ftmp0] \n\t" |
| "pshufh %[D], %[D], %[ftmp0] \n\t" |
| |
| "1: \n\t" |
| MMI_ULWC1(%[ftmp1], %[src], 0x00) |
| MMI_ULWC1(%[ftmp2], %[src], 0x01) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULWC1(%[ftmp3], %[src], 0x00) |
| MMI_ULWC1(%[ftmp4], %[src], 0x01) |
| |
| "punpcklbh %[ftmp5], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp6], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[A] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[B] \n\t" |
| "paddh %[ftmp1], %[ftmp5], %[ftmp6] \n\t" |
| "punpcklbh %[ftmp5], %[ftmp3], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp6], %[ftmp4], %[ftmp0] \n\t" |
| "pmullh %[ftmp5], %[ftmp5], %[C] \n\t" |
| "pmullh %[ftmp6], %[ftmp6], %[D] \n\t" |
| "paddh %[ftmp2], %[ftmp5], %[ftmp6] \n\t" |
| "paddh %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp1], %[ftmp1], %[ftmp7] \n\t" |
| "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" |
| MMI_LWC1(%[ftmp2], %[dst], 0x00) |
| "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| |
| "addi %[h], %[h], -0x01 \n\t" |
| MMI_SWC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| [ftmp6]"=&f"(ftmp[6]), [ftmp7]"=&f"(ftmp[7]), |
| RESTRICT_ASM_LOW32 |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride),[ff_pw_32]"f"(ff_pw_32.f), |
| [A]"f"(A.f), [B]"f"(B.f), |
| [C]"f"(C.f), [D]"f"(D.f), |
| [tmp0]"r"(0x06) |
| : "memory" |
| ); |
| } else if (E.i) { |
| const int step = C.i ? stride : 1; |
| __asm__ volatile ( |
| "pxor %[ftmp0], %[ftmp0], %[ftmp0] \n\t" |
| "pshufh %[A], %[A], %[ftmp0] \n\t" |
| "pshufh %[E], %[E], %[ftmp0] \n\t" |
| "mtc1 %[tmp0], %[ftmp5] \n\t" |
| |
| "1: \n\t" |
| MMI_ULWC1(%[ftmp1], %[src], 0x00) |
| PTR_ADDU "%[addr0], %[src], %[step] \n\t" |
| MMI_ULWC1(%[ftmp2], %[addr0], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| "addi %[h], %[h], -0x01 \n\t" |
| "punpcklbh %[ftmp3], %[ftmp1], %[ftmp0] \n\t" |
| "punpcklbh %[ftmp4], %[ftmp2], %[ftmp0] \n\t" |
| "pmullh %[ftmp3], %[ftmp3], %[A] \n\t" |
| "pmullh %[ftmp4], %[ftmp4], %[E] \n\t" |
| "paddh %[ftmp1], %[ftmp3], %[ftmp4] \n\t" |
| "paddh %[ftmp1], %[ftmp1], %[ff_pw_32] \n\t" |
| "psrlh %[ftmp1], %[ftmp1], %[ftmp5] \n\t" |
| "packushb %[ftmp1], %[ftmp1], %[ftmp0] \n\t" |
| MMI_LWC1(%[ftmp2], %[dst], 0x00) |
| "pavgb %[ftmp1], %[ftmp1], %[ftmp2] \n\t" |
| MMI_SWC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [ftmp4]"=&f"(ftmp[4]), [ftmp5]"=&f"(ftmp[5]), |
| RESTRICT_ASM_LOW32 |
| [addr0]"=&r"(addr[0]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride),[step]"r"((mips_reg)step), |
| [ff_pw_32]"f"(ff_pw_32.f), [tmp0]"r"(0x06), |
| [A]"f"(A.f), [E]"f"(E.f) |
| : "memory" |
| ); |
| } else { |
| __asm__ volatile ( |
| "1: \n\t" |
| MMI_ULWC1(%[ftmp0], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| MMI_ULWC1(%[ftmp1], %[src], 0x00) |
| PTR_ADDU "%[src], %[src], %[stride] \n\t" |
| "addi %[h], %[h], -0x02 \n\t" |
| MMI_LWC1(%[ftmp2], %[dst], 0x00) |
| "pavgb %[ftmp0], %[ftmp0], %[ftmp2] \n\t" |
| MMI_SWC1(%[ftmp0], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| MMI_LWC1(%[ftmp3], %[dst], 0x00) |
| "pavgb %[ftmp1], %[ftmp1], %[ftmp3] \n\t" |
| MMI_SWC1(%[ftmp1], %[dst], 0x00) |
| PTR_ADDU "%[dst], %[dst], %[stride] \n\t" |
| "bnez %[h], 1b \n\t" |
| : [ftmp0]"=&f"(ftmp[0]), [ftmp1]"=&f"(ftmp[1]), |
| [ftmp2]"=&f"(ftmp[2]), [ftmp3]"=&f"(ftmp[3]), |
| [dst]"+&r"(dst), [src]"+&r"(src), |
| RESTRICT_ASM_LOW32 |
| [h]"+&r"(h) |
| : [stride]"r"((mips_reg)stride) |
| : "memory" |
| ); |
| } |
| } |