| ;****************************************************************************** |
| ;* SIMD-optimized IDCT-related routines |
| ;* Copyright (c) 2008 Loren Merritt |
| ;* Copyright (c) 2003-2013 Michael Niedermayer |
| ;* Copyright (c) 2013 Daniel Kang |
| ;* |
| ;* This file is part of FFmpeg. |
| ;* |
| ;* FFmpeg is free software; you can redistribute it and/or |
| ;* modify it under the terms of the GNU Lesser General Public |
| ;* License as published by the Free Software Foundation; either |
| ;* version 2.1 of the License, or (at your option) any later version. |
| ;* |
| ;* FFmpeg is distributed in the hope that it will be useful, |
| ;* but WITHOUT ANY WARRANTY; without even the implied warranty of |
| ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| ;* Lesser General Public License for more details. |
| ;* |
| ;* You should have received a copy of the GNU Lesser General Public |
| ;* License along with FFmpeg; if not, write to the Free Software |
| ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| ;****************************************************************************** |
| |
| %include "libavutil/x86/x86util.asm" |
| |
| SECTION_RODATA |
| |
| cextern pb_80 |
| |
| SECTION .text |
| |
| ;-------------------------------------------------------------------------- |
| ;void ff_put_signed_pixels_clamped(const int16_t *block, uint8_t *pixels, |
| ; ptrdiff_t line_size) |
| ;-------------------------------------------------------------------------- |
| |
| %macro PUT_SIGNED_PIXELS_CLAMPED_HALF 1 |
| mova m1, [blockq+mmsize*0+%1] |
| mova m2, [blockq+mmsize*2+%1] |
| %if mmsize == 8 |
| mova m3, [blockq+mmsize*4+%1] |
| mova m4, [blockq+mmsize*6+%1] |
| %endif |
| packsswb m1, [blockq+mmsize*1+%1] |
| packsswb m2, [blockq+mmsize*3+%1] |
| %if mmsize == 8 |
| packsswb m3, [blockq+mmsize*5+%1] |
| packsswb m4, [blockq+mmsize*7+%1] |
| %endif |
| paddb m1, m0 |
| paddb m2, m0 |
| %if mmsize == 8 |
| paddb m3, m0 |
| paddb m4, m0 |
| movq [pixelsq+lsizeq*0], m1 |
| movq [pixelsq+lsizeq*1], m2 |
| movq [pixelsq+lsizeq*2], m3 |
| movq [pixelsq+lsize3q ], m4 |
| %else |
| movq [pixelsq+lsizeq*0], m1 |
| movhps [pixelsq+lsizeq*1], m1 |
| movq [pixelsq+lsizeq*2], m2 |
| movhps [pixelsq+lsize3q ], m2 |
| %endif |
| %endmacro |
| |
| %macro PUT_SIGNED_PIXELS_CLAMPED 1 |
| cglobal put_signed_pixels_clamped, 3, 4, %1, block, pixels, lsize, lsize3 |
| mova m0, [pb_80] |
| lea lsize3q, [lsizeq*3] |
| PUT_SIGNED_PIXELS_CLAMPED_HALF 0 |
| lea pixelsq, [pixelsq+lsizeq*4] |
| PUT_SIGNED_PIXELS_CLAMPED_HALF 64 |
| RET |
| %endmacro |
| |
| INIT_MMX mmx |
| PUT_SIGNED_PIXELS_CLAMPED 0 |
| INIT_XMM sse2 |
| PUT_SIGNED_PIXELS_CLAMPED 3 |
| |
| ;-------------------------------------------------------------------------- |
| ; void ff_put_pixels_clamped(const int16_t *block, uint8_t *pixels, |
| ; ptrdiff_t line_size); |
| ;-------------------------------------------------------------------------- |
| ; %1 = block offset |
| %macro PUT_PIXELS_CLAMPED_HALF 1 |
| mova m0, [blockq+mmsize*0+%1] |
| mova m1, [blockq+mmsize*2+%1] |
| %if mmsize == 8 |
| mova m2, [blockq+mmsize*4+%1] |
| mova m3, [blockq+mmsize*6+%1] |
| %endif |
| packuswb m0, [blockq+mmsize*1+%1] |
| packuswb m1, [blockq+mmsize*3+%1] |
| %if mmsize == 8 |
| packuswb m2, [blockq+mmsize*5+%1] |
| packuswb m3, [blockq+mmsize*7+%1] |
| movq [pixelsq], m0 |
| movq [lsizeq+pixelsq], m1 |
| movq [2*lsizeq+pixelsq], m2 |
| movq [lsize3q+pixelsq], m3 |
| %else |
| movq [pixelsq], m0 |
| movhps [lsizeq+pixelsq], m0 |
| movq [2*lsizeq+pixelsq], m1 |
| movhps [lsize3q+pixelsq], m1 |
| %endif |
| %endmacro |
| |
| %macro PUT_PIXELS_CLAMPED 0 |
| cglobal put_pixels_clamped, 3, 4, 2, block, pixels, lsize, lsize3 |
| lea lsize3q, [lsizeq*3] |
| PUT_PIXELS_CLAMPED_HALF 0 |
| lea pixelsq, [pixelsq+lsizeq*4] |
| PUT_PIXELS_CLAMPED_HALF 64 |
| RET |
| %endmacro |
| |
| INIT_MMX mmx |
| PUT_PIXELS_CLAMPED |
| INIT_XMM sse2 |
| PUT_PIXELS_CLAMPED |
| |
| ;-------------------------------------------------------------------------- |
| ; void ff_add_pixels_clamped(const int16_t *block, uint8_t *pixels, |
| ; ptrdiff_t line_size); |
| ;-------------------------------------------------------------------------- |
| ; %1 = block offset |
| %macro ADD_PIXELS_CLAMPED 1 |
| mova m0, [blockq+mmsize*0+%1] |
| mova m1, [blockq+mmsize*1+%1] |
| %if mmsize == 8 |
| mova m5, [blockq+mmsize*2+%1] |
| mova m6, [blockq+mmsize*3+%1] |
| %endif |
| movq m2, [pixelsq] |
| movq m3, [pixelsq+lsizeq] |
| %if mmsize == 8 |
| mova m7, m2 |
| punpcklbw m2, m4 |
| punpckhbw m7, m4 |
| paddsw m0, m2 |
| paddsw m1, m7 |
| mova m7, m3 |
| punpcklbw m3, m4 |
| punpckhbw m7, m4 |
| paddsw m5, m3 |
| paddsw m6, m7 |
| %else |
| punpcklbw m2, m4 |
| punpcklbw m3, m4 |
| paddsw m0, m2 |
| paddsw m1, m3 |
| %endif |
| packuswb m0, m1 |
| %if mmsize == 8 |
| packuswb m5, m6 |
| movq [pixelsq], m0 |
| movq [pixelsq+lsizeq], m5 |
| %else |
| movq [pixelsq], m0 |
| movhps [pixelsq+lsizeq], m0 |
| %endif |
| %endmacro |
| |
| %macro ADD_PIXELS_CLAMPED 0 |
| cglobal add_pixels_clamped, 3, 3, 5, block, pixels, lsize |
| pxor m4, m4 |
| ADD_PIXELS_CLAMPED 0 |
| lea pixelsq, [pixelsq+lsizeq*2] |
| ADD_PIXELS_CLAMPED 32 |
| lea pixelsq, [pixelsq+lsizeq*2] |
| ADD_PIXELS_CLAMPED 64 |
| lea pixelsq, [pixelsq+lsizeq*2] |
| ADD_PIXELS_CLAMPED 96 |
| RET |
| %endmacro |
| |
| INIT_MMX mmx |
| ADD_PIXELS_CLAMPED |
| INIT_XMM sse2 |
| ADD_PIXELS_CLAMPED |