;****************************************************************************** ;* Copyright Nick Kurshev ;* Copyright Michael (michaelni@gmx.at) ;* Copyright 2018 Jokyo Images ;* Copyright Ivo van Poorten ;* ;* This file is part of FFmpeg. ;* ;* FFmpeg is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* FFmpeg is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with FFmpeg; if not, write to the Free Software ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "libavutil/x86/x86util.asm" SECTION_RODATA pb_mask_shuffle2103_mmx times 8 dw 255 pb_shuffle2103: db 2, 1, 0, 3, 6, 5, 4, 7, 10, 9, 8, 11, 14, 13, 12, 15 pb_shuffle0321: db 0, 3, 2, 1, 4, 7, 6, 5, 8, 11, 10, 9, 12, 15, 14, 13 pb_shuffle1230: db 1, 2, 3, 0, 5, 6, 7, 4, 9, 10, 11, 8, 13, 14, 15, 12 pb_shuffle3012: db 3, 0, 1, 2, 7, 4, 5, 6, 11, 8, 9, 10, 15, 12, 13, 14 pb_shuffle3210: db 3, 2, 1, 0, 7, 6, 5, 4, 11, 10, 9, 8, 15, 14, 13, 12 SECTION .text %macro RSHIFT_COPY 3 ; %1 dst ; %2 src ; %3 shift %if cpuflag(avx) psrldq %1, %2, %3 %else mova %1, %2 RSHIFT %1, %3 %endif %endmacro ;------------------------------------------------------------------------------ ; shuffle_bytes_2103_mmext (const uint8_t *src, uint8_t *dst, int src_size) ;------------------------------------------------------------------------------ INIT_MMX mmxext cglobal shuffle_bytes_2103, 3, 5, 8, src, dst, w, tmp, x mova m6, [pb_mask_shuffle2103_mmx] mova m7, m6 psllq m7, 8 movsxdifnidn wq, wd mov xq, wq add srcq, wq add dstq, wq neg wq ;calc scalar loop and xq, mmsize*2 -4 je .loop_simd .loop_scalar: mov tmpb, [srcq + wq + 2] mov [dstq+wq + 0], tmpb mov tmpb, [srcq + wq + 1] mov [dstq+wq + 1], tmpb mov tmpb, [srcq + wq + 0] mov [dstq+wq + 2], tmpb mov tmpb, [srcq + wq + 3] mov [dstq+wq + 3], tmpb add wq, 4 sub xq, 4 jg .loop_scalar ;check if src_size < mmsize * 2 cmp wq, 0 jge .end .loop_simd: movu m0, [srcq+wq] movu m1, [srcq+wq+8] pshufw m3, m0, 177 pshufw m5, m1, 177 pand m0, m7 pand m3, m6 pand m1, m7 pand m5, m6 por m0, m3 por m1, m5 movu [dstq+wq], m0 movu [dstq+wq + 8], m1 add wq, mmsize*2 jl .loop_simd .end: RET ;------------------------------------------------------------------------------ ; shuffle_bytes_## (const uint8_t *src, uint8_t *dst, int src_size) ;------------------------------------------------------------------------------ ; %1-4 index shuffle %macro SHUFFLE_BYTES 4 cglobal shuffle_bytes_%1%2%3%4, 3, 5, 2, src, dst, w, tmp, x VBROADCASTI128 m0, [pb_shuffle%1%2%3%4] movsxdifnidn wq, wd mov xq, wq add srcq, wq add dstq, wq neg wq ;calc scalar loop and xq, mmsize-4 je .loop_simd .loop_scalar: mov tmpb, [srcq + wq + %1] mov [dstq+wq + 0], tmpb mov tmpb, [srcq + wq + %2] mov [dstq+wq + 1], tmpb mov tmpb, [srcq + wq + %3] mov [dstq+wq + 2], tmpb mov tmpb, [srcq + wq + %4] mov [dstq+wq + 3], tmpb add wq, 4 sub xq, 4 jg .loop_scalar ;check if src_size < mmsize cmp wq, 0 jge .end .loop_simd: movu m1, [srcq+wq] pshufb m1, m0 movu [dstq+wq], m1 add wq, mmsize jl .loop_simd .end: RET %endmacro INIT_XMM ssse3 SHUFFLE_BYTES 2, 1, 0, 3 SHUFFLE_BYTES 0, 3, 2, 1 SHUFFLE_BYTES 1, 2, 3, 0 SHUFFLE_BYTES 3, 0, 1, 2 SHUFFLE_BYTES 3, 2, 1, 0 ;----------------------------------------------------------------------------------------------- ; uyvytoyuv422(uint8_t *ydst, uint8_t *udst, uint8_t *vdst, ; const uint8_t *src, int width, int height, ; int lumStride, int chromStride, int srcStride) ;----------------------------------------------------------------------------------------------- %macro UYVY_TO_YUV422 0 cglobal uyvytoyuv422, 9, 14, 8, ydst, udst, vdst, src, w, h, lum_stride, chrom_stride, src_stride, wtwo, whalf, tmp, x, back_w pxor m0, m0 pcmpeqw m1, m1 psrlw m1, 8 movsxdifnidn wq, wd movsxdifnidn lum_strideq, lum_strided movsxdifnidn chrom_strideq, chrom_strided movsxdifnidn src_strideq, src_strided mov back_wq, wq mov whalfq, wq shr whalfq, 1 ; whalf = width / 2 lea srcq, [srcq + wq * 2] add ydstq, wq add udstq, whalfq add vdstq, whalfq .loop_line: mov xq, wq mov wtwoq, wq add wtwoq, wtwoq ; wtwo = width * 2 neg wq neg wtwoq neg whalfq ;calc scalar loop count and xq, mmsize * 2 - 1 je .loop_simd .loop_scalar: mov tmpb, [srcq + wtwoq + 0] mov [udstq + whalfq], tmpb mov tmpb, [srcq + wtwoq + 1] mov [ydstq + wq], tmpb mov tmpb, [srcq + wtwoq + 2] mov [vdstq + whalfq], tmpb mov tmpb, [srcq + wtwoq + 3] mov [ydstq + wq + 1], tmpb add wq, 2 add wtwoq, 4 add whalfq, 1 sub xq, 2 jg .loop_scalar ; check if simd loop is need cmp wq, 0 jge .end_line .loop_simd: movu m2, [srcq + wtwoq ] movu m3, [srcq + wtwoq + mmsize ] movu m4, [srcq + wtwoq + mmsize * 2] movu m5, [srcq + wtwoq + mmsize * 3] ; extract y part 1 RSHIFT_COPY m6, m2, 1 ; UYVY UYVY -> YVYU YVY... pand m6, m1; YxYx YxYx... RSHIFT_COPY m7, m3, 1 ; UYVY UYVY -> YVYU YVY... pand m7, m1 ; YxYx YxYx... packuswb m6, m7 ; YYYY YYYY... movu [ydstq + wq], m6 ; extract y part 2 RSHIFT_COPY m6, m4, 1 ; UYVY UYVY -> YVYU YVY... pand m6, m1; YxYx YxYx... RSHIFT_COPY m7, m5, 1 ; UYVY UYVY -> YVYU YVY... pand m7, m1 ; YxYx YxYx... packuswb m6, m7 ; YYYY YYYY... movu [ydstq + wq + mmsize], m6 ; extract uv pand m2, m1 ; UxVx... pand m3, m1 ; UxVx... pand m4, m1 ; UxVx... pand m5, m1 ; UxVx... packuswb m2, m3 ; UVUV... packuswb m4, m5 ; UVUV... ; U pand m6, m2, m1 ; UxUx... pand m7, m4, m1 ; UxUx... packuswb m6, m7 ; UUUU movu [udstq + whalfq], m6 ; V psrlw m2, 8 ; VxVx... psrlw m4, 8 ; VxVx... packuswb m2, m4 ; VVVV movu [vdstq + whalfq], m2 add whalfq, mmsize add wtwoq, mmsize * 4 add wq, mmsize * 2 jl .loop_simd .end_line: add srcq, src_strideq add ydstq, lum_strideq add udstq, chrom_strideq add vdstq, chrom_strideq ;restore initial state of line variable mov wq, back_wq mov xq, wq mov whalfq, wq shr whalfq, 1 ; whalf = width / 2 sub hd, 1 jg .loop_line RET %endmacro %if ARCH_X86_64 INIT_XMM sse2 UYVY_TO_YUV422 INIT_XMM avx UYVY_TO_YUV422 %endif