*
* based on code by Copyright (C) 2001-2003 Michael Niedermayer (michaelni@gmx.at)
*
- * This file is part of FFmpeg.
+ * This file is part of Libav.
*
- * FFmpeg is free software; you can redistribute it and/or modify
+ * Libav is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
- * FFmpeg is distributed in the hope that it will be useful,
+ * Libav is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
- * along with FFmpeg; if not, write to the Free Software
+ * along with Libav; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
vector by assuming (stride % 16) == 0, unfortunately
this is not always true.
*/
- DECLARE_ALIGNED(16, short, data[8]) =
+ short data_0 = ((c->nonBQP*c->ppMode.baseDcDiff)>>8) + 1;
+ DECLARE_ALIGNED(16, short, data)[8] =
{
- ((c->nonBQP*c->ppMode.baseDcDiff)>>8) + 1,
- data[0] * 2 + 1,
+ data_0,
+ data_0 * 2 + 1,
c->QP * 2,
c->QP * 4
};
const vector signed int zero = vec_splat_s32(0);
const int properStride = (stride % 16);
const int srcAlign = ((unsigned long)src2 % 16);
- DECLARE_ALIGNED(16, short, qp[8]) = {c->QP};
+ DECLARE_ALIGNED(16, short, qp)[8] = {c->QP};
vector signed short vqp = vec_ld(0, qp);
vector signed short vb0, vb1, vb2, vb3, vb4, vb5, vb6, vb7, vb8, vb9;
vector unsigned char vbA0, av_uninit(vbA1), av_uninit(vbA2), av_uninit(vbA3), av_uninit(vbA4), av_uninit(vbA5), av_uninit(vbA6), av_uninit(vbA7), av_uninit(vbA8), vbA9;
*/
uint8_t *src2 = src + stride*3;
const vector signed int zero = vec_splat_s32(0);
- DECLARE_ALIGNED(16, short, qp[8]) = {8*c->QP};
+ DECLARE_ALIGNED(16, short, qp)[8] = {8*c->QP};
vector signed short vqp = vec_splat(
(vector signed short)vec_ld(0, qp), 0);
src & stride :-(
*/
uint8_t *srcCopy = src;
- DECLARE_ALIGNED(16, uint8_t, dt[16]);
+ DECLARE_ALIGNED(16, uint8_t, dt)[16];
const vector signed int zero = vec_splat_s32(0);
vector unsigned char v_dt;
dt[0] = deringThreshold;
v_avg = vec_avg(v_min, v_max);
}
- DECLARE_ALIGNED(16, signed int, S[8]);
+ DECLARE_ALIGNED(16, signed int, S)[8];
{
const vector unsigned short mask1 = (vector unsigned short)
{0x0001, 0x0002, 0x0004, 0x0008,
/* I'm not sure the following is actually faster
than straight, unvectorized C code :-( */
- DECLARE_ALIGNED(16, int, tQP2[4]);
+ DECLARE_ALIGNED(16, int, tQP2)[4];
tQP2[0]= c->QP/2 + 1;
vector signed int vQP2 = vec_ld(0, tQP2);
vQP2 = vec_splat(vQP2, 0);