;****************************************************************************** ;* SIMD-optimized HuffYUV functions ;* Copyright (c) 2008 Loren Merritt ;* Copyright (c) 2014 Christophe Gisquet ;* ;* This file is part of FFmpeg. ;* ;* FFmpeg is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* FFmpeg is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with FFmpeg; if not, write to the Free Software ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %macro INT16_LOOP 2 ; %1 = a/u (aligned/unaligned), %2 = add/sub movd xm4, maskd SPLATW m4, xm4 add wd, wd test wq, 2*mmsize - 1 jz %%.tomainloop push tmpq %%.wordloop: sub wq, 2 %ifidn %2, add mov tmpw, [srcq+wq] add tmpw, [dstq+wq] %else mov tmpw, [src1q+wq] sub tmpw, [src2q+wq] %endif and tmpw, maskw mov [dstq+wq], tmpw test wq, 2*mmsize - 1 jnz %%.wordloop pop tmpq %%.tomainloop: %ifidn %2, add add srcq, wq %else add src1q, wq add src2q, wq %endif add dstq, wq neg wq jz %%.end %%.loop: %ifidn %2, add mov%1 m0, [srcq+wq] mov%1 m1, [dstq+wq] mov%1 m2, [srcq+wq+mmsize] mov%1 m3, [dstq+wq+mmsize] %else mov%1 m0, [src1q+wq] mov%1 m1, [src2q+wq] mov%1 m2, [src1q+wq+mmsize] mov%1 m3, [src2q+wq+mmsize] %endif p%2w m0, m1 p%2w m2, m3 pand m0, m4 pand m2, m4 mov%1 [dstq+wq] , m0 mov%1 [dstq+wq+mmsize], m2 add wq, 2*mmsize jl %%.loop %%.end: RET %endmacro