diff options
Diffstat (limited to 'libavcodec/x86/hpeldsp.asm')
-rw-r--r-- | libavcodec/x86/hpeldsp.asm | 465 |
1 files changed, 465 insertions, 0 deletions
diff --git a/libavcodec/x86/hpeldsp.asm b/libavcodec/x86/hpeldsp.asm new file mode 100644 index 0000000000..8afd955bf0 --- /dev/null +++ b/libavcodec/x86/hpeldsp.asm @@ -0,0 +1,465 @@ +;****************************************************************************** +;* MMX optimized hpel functions +;* +;* This file is part of Libav. +;* +;* Libav is free software; you can redistribute it and/or +;* modify it under the terms of the GNU Lesser General Public +;* License as published by the Free Software Foundation; either +;* version 2.1 of the License, or (at your option) any later version. +;* +;* Libav is distributed in the hope that it will be useful, +;* but WITHOUT ANY WARRANTY; without even the implied warranty of +;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU +;* Lesser General Public License for more details. +;* +;* You should have received a copy of the GNU Lesser General Public +;* License along with Libav; if not, write to the Free Software +;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA +;****************************************************************************** + +%include "libavutil/x86/x86util.asm" + +SECTION_RODATA +cextern pb_1 + +SECTION_TEXT + +; put_pixels8_x2(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro PUT_PIXELS8_X2 0 +cglobal put_pixels8_x2, 4,5 + movsxdifnidn r2, r2d + lea r4, [r2*2] +.loop: + mova m0, [r1] + mova m1, [r1+r2] + PAVGB m0, [r1+1] + PAVGB m1, [r1+r2+1] + mova [r0], m0 + mova [r0+r2], m1 + add r1, r4 + add r0, r4 + mova m0, [r1] + mova m1, [r1+r2] + PAVGB m0, [r1+1] + PAVGB m1, [r1+r2+1] + add r1, r4 + mova [r0], m0 + mova [r0+r2], m1 + add r0, r4 + sub r3d, 4 + jne .loop + REP_RET +%endmacro + +INIT_MMX mmxext +PUT_PIXELS8_X2 +INIT_MMX 3dnow +PUT_PIXELS8_X2 + + +; put_pixels16_x2(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro PUT_PIXELS_16 0 +cglobal put_pixels16_x2, 4,5 + movsxdifnidn r2, r2d + lea r4, [r2*2] +.loop: + mova m0, [r1] + mova m1, [r1+r2] + mova m2, [r1+8] + mova m3, [r1+r2+8] + PAVGB m0, [r1+1] + PAVGB m1, [r1+r2+1] + PAVGB m2, [r1+9] + PAVGB m3, [r1+r2+9] + mova [r0], m0 + mova [r0+r2], m1 + mova [r0+8], m2 + mova [r0+r2+8], m3 + add r1, r4 + add r0, r4 + mova m0, [r1] + mova m1, [r1+r2] + mova m2, [r1+8] + mova m3, [r1+r2+8] + PAVGB m0, [r1+1] + PAVGB m1, [r1+r2+1] + PAVGB m2, [r1+9] + PAVGB m3, [r1+r2+9] + add r1, r4 + mova [r0], m0 + mova [r0+r2], m1 + mova [r0+8], m2 + mova [r0+r2+8], m3 + add r0, r4 + sub r3d, 4 + jne .loop + REP_RET +%endmacro + +INIT_MMX mmxext +PUT_PIXELS_16 +INIT_MMX 3dnow +PUT_PIXELS_16 + + +; put_no_rnd_pixels8_x2(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro PUT_NO_RND_PIXELS8_X2 0 +cglobal put_no_rnd_pixels8_x2, 4,5 + mova m6, [pb_1] + movsxdifnidn r2, r2d + lea r4, [r2*2] +.loop: + mova m0, [r1] + mova m2, [r1+r2] + mova m1, [r1+1] + mova m3, [r1+r2+1] + add r1, r4 + psubusb m0, m6 + psubusb m2, m6 + PAVGB m0, m1 + PAVGB m2, m3 + mova [r0], m0 + mova [r0+r2], m2 + mova m0, [r1] + mova m1, [r1+1] + mova m2, [r1+r2] + mova m3, [r1+r2+1] + add r0, r4 + add r1, r4 + psubusb m0, m6 + psubusb m2, m6 + PAVGB m0, m1 + PAVGB m2, m3 + mova [r0], m0 + mova [r0+r2], m2 + add r0, r4 + sub r3d, 4 + jne .loop + REP_RET +%endmacro + +INIT_MMX mmxext +PUT_NO_RND_PIXELS8_X2 +INIT_MMX 3dnow +PUT_NO_RND_PIXELS8_X2 + + +; put_no_rnd_pixels8_x2_exact(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro PUT_NO_RND_PIXELS8_X2_EXACT 0 +cglobal put_no_rnd_pixels8_x2_exact, 4,5 + movsxdifnidn r2, r2d + lea r4, [r2*3] + pcmpeqb m6, m6 +.loop: + mova m0, [r1] + mova m2, [r1+r2] + mova m1, [r1+1] + mova m3, [r1+r2+1] + pxor m0, m6 + pxor m2, m6 + pxor m1, m6 + pxor m3, m6 + PAVGB m0, m1 + PAVGB m2, m3 + pxor m0, m6 + pxor m2, m6 + mova [r0], m0 + mova [r0+r2], m2 + mova m0, [r1+r2*2] + mova m1, [r1+r2*2+1] + mova m2, [r1+r4] + mova m3, [r1+r4+1] + pxor m0, m6 + pxor m1, m6 + pxor m2, m6 + pxor m3, m6 + PAVGB m0, m1 + PAVGB m2, m3 + pxor m0, m6 + pxor m2, m6 + mova [r0+r2*2], m0 + mova [r0+r4], m2 + lea r1, [r1+r2*4] + lea r0, [r0+r2*4] + sub r3d, 4 + jg .loop + REP_RET +%endmacro + +INIT_MMX mmxext +PUT_NO_RND_PIXELS8_X2_EXACT +INIT_MMX 3dnow +PUT_NO_RND_PIXELS8_X2_EXACT + + +; put_pixels8_y2(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro PUT_PIXELS8_Y2 0 +cglobal put_pixels8_y2, 4,5 + movsxdifnidn r2, r2d + lea r4, [r2*2] + mova m0, [r1] + sub r0, r2 +.loop: + mova m1, [r1+r2] + mova m2, [r1+r4] + add r1, r4 + PAVGB m0, m1 + PAVGB m1, m2 + mova [r0+r2], m0 + mova [r0+r4], m1 + mova m1, [r1+r2] + mova m0, [r1+r4] + add r0, r4 + add r1, r4 + PAVGB m2, m1 + PAVGB m1, m0 + mova [r0+r2], m2 + mova [r0+r4], m1 + add r0, r4 + sub r3d, 4 + jne .loop + REP_RET +%endmacro + +INIT_MMX mmxext +PUT_PIXELS8_Y2 +INIT_MMX 3dnow +PUT_PIXELS8_Y2 + + +; put_no_rnd_pixels8_y2(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro PUT_NO_RND_PIXELS8_Y2 0 +cglobal put_no_rnd_pixels8_y2, 4,5 + mova m6, [pb_1] + movsxdifnidn r2, r2d + lea r4, [r2+r2] + mova m0, [r1] + sub r0, r2 +.loop: + mova m1, [r1+r2] + mova m2, [r1+r4] + add r1, r4 + psubusb m1, m6 + PAVGB m0, m1 + PAVGB m1, m2 + mova [r0+r2], m0 + mova [r0+r4], m1 + mova m1, [r1+r2] + mova m0, [r1+r4] + add r0, r4 + add r1, r4 + psubusb m1, m6 + PAVGB m2, m1 + PAVGB m1, m0 + mova [r0+r2], m2 + mova [r0+r4], m1 + add r0, r4 + sub r3d, 4 + jne .loop + REP_RET +%endmacro + +INIT_MMX mmxext +PUT_NO_RND_PIXELS8_Y2 +INIT_MMX 3dnow +PUT_NO_RND_PIXELS8_Y2 + + +; put_no_rnd_pixels8_y2_exact(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro PUT_NO_RND_PIXELS8_Y2_EXACT 0 +cglobal put_no_rnd_pixels8_y2_exact, 4,5 + movsxdifnidn r2, r2d + lea r4, [r2*3] + mova m0, [r1] + pcmpeqb m6, m6 + add r1, r2 + pxor m0, m6 +.loop: + mova m1, [r1] + mova m2, [r1+r2] + pxor m1, m6 + pxor m2, m6 + PAVGB m0, m1 + PAVGB m1, m2 + pxor m0, m6 + pxor m1, m6 + mova [r0], m0 + mova [r0+r2], m1 + mova m1, [r1+r2*2] + mova m0, [r1+r4] + pxor m1, m6 + pxor m0, m6 + PAVGB m2, m1 + PAVGB m1, m0 + pxor m2, m6 + pxor m1, m6 + mova [r0+r2*2], m2 + mova [r0+r4], m1 + lea r1, [r1+r2*4] + lea r0, [r0+r2*4] + sub r3d, 4 + jg .loop + REP_RET +%endmacro + +INIT_MMX mmxext +PUT_NO_RND_PIXELS8_Y2_EXACT +INIT_MMX 3dnow +PUT_NO_RND_PIXELS8_Y2_EXACT + + +; avg_pixels8(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro AVG_PIXELS8 0 +cglobal avg_pixels8, 4,5 + movsxdifnidn r2, edx + lea r4, [r2+r2] +.loop: + mova m0, [r0] + mova m1, [r0+r2] + PAVGB m0, [r1] + PAVGB m1, [r1+r2] + mova [r0], m0 + mova [r0+r2], m1 + add r1, r4 + add r0, r4 + mova m0, [r0] + mova m1, [r0+r2] + PAVGB m0, [r1] + PAVGB m1, [r1+r2] + add r1, r4 + mova [r0], m0 + mova [r0+r2], m1 + add r0, r4 + sub r3d, 4 + jne .loop + REP_RET +%endmacro + +INIT_MMX 3dnow +AVG_PIXELS8 + + +; avg_pixels8_x2(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro AVG_PIXELS8_X2 0 +cglobal avg_pixels8_x2, 4,5 + movsxdifnidn r2, edx + lea r4, [r2*2] +.loop: + mova m0, [r1] + mova m2, [r1+r2] + PAVGB m0, [r1+1] + PAVGB m2, [r1+r2+1] + PAVGB m0, [r0] + PAVGB m2, [r0+r2] + add r1, r4 + mova [r0], m0 + mova [r0+r2], m2 + mova m0, [r1] + mova m2, [r1+r2] + PAVGB m0, [r1+1] + PAVGB m2, [r1+r2+1] + add r0, r4 + add r1, r4 + PAVGB m0, [r0] + PAVGB m2, [r0+r2] + mova [r0], m0 + mova [r0+r2], m2 + add r0, r4 + sub r3d, 4 + jne .loop + REP_RET +%endmacro + +INIT_MMX mmxext +AVG_PIXELS8_X2 +INIT_MMX 3dnow +AVG_PIXELS8_X2 + + +; avg_pixels8_y2(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro AVG_PIXELS8_Y2 0 +cglobal avg_pixels8_y2, 4,5 + movsxdifnidn r2, r2d + lea r4, [r2*2] + mova m0, [r1] + sub r0, r2 +.loop: + mova m1, [r1+r2] + mova m2, [r1+r4] + add r1, r4 + PAVGB m0, m1 + PAVGB m1, m2 + mova m3, [r0+r2] + mova m4, [r0+r4] + PAVGB m0, m3 + PAVGB m1, m4 + mova [r0+r2], m0 + mova [r0+r4], m1 + mova m1, [r1+r2] + mova m0, [r1+r4] + PAVGB m2, m1 + PAVGB m1, m0 + add r0, r4 + add r1, r4 + mova m3, [r0+r2] + mova m4, [r0+r4] + PAVGB m2, m3 + PAVGB m1, m4 + mova [r0+r2], m2 + mova [r0+r4], m1 + add r0, r4 + sub r3d, 4 + jne .loop + REP_RET +%endmacro + +INIT_MMX mmxext +AVG_PIXELS8_Y2 +INIT_MMX 3dnow +AVG_PIXELS8_Y2 + + +; avg_pixels8_xy2(uint8_t *block, const uint8_t *pixels, int line_size, int h) +%macro AVG_PIXELS8_XY2 0 +cglobal avg_pixels8_xy2, 4,5 + mova m6, [pb_1] + movsxdifnidn r2, r2d + lea r4, [r2*2] + mova m0, [r1] + pavgb m0, [r1+1] +.loop: + mova m2, [r1+r4] + mova m1, [r1+r2] + psubusb m2, m6 + pavgb m1, [r1+r2+1] + pavgb m2, [r1+r4+1] + add r1, r4 + pavgb m0, m1 + pavgb m1, m2 + pavgb m0, [r0] + pavgb m1, [r0+r2] + mova [r0], m0 + mova [r0+r2], m1 + mova m1, [r1+r2] + mova m0, [r1+r4] + pavgb m1, [r1+r2+1] + pavgb m0, [r1+r4+1] + add r0, r4 + add r1, r4 + pavgb m2, m1 + pavgb m1, m0 + pavgb m2, [r0] + pavgb m1, [r0+r2] + mova [r0], m2 + mova [r0+r2], m2 + add r0, r4 + sub r3d, 4 + jne .loop + REP_RET +%endmacro + +INIT_MMX mmxext +AVG_PIXELS8_XY2 +INIT_MMX 3dnow +AVG_PIXELS8_XY2 |