add r0, r0, #8
add r1, r1, #8
b ff_\type\()_pixels8\subp\()_armv6
-.endfunc
+endfunc
.endm
call_2x_pixels avg
pop {r4-r11}
bx lr
-.endfunc
+endfunc
function ff_put_pixels8_armv6, export=1
push {r4-r7}
pop {r4-r7}
bx lr
-.endfunc
+endfunc
function ff_put_pixels8_x2_armv6, export=1
push {r4-r11, lr}
bne 1b
pop {r4-r11, pc}
-.endfunc
+endfunc
function ff_put_pixels8_y2_armv6, export=1
push {r4-r11}
pop {r4-r11}
bx lr
-.endfunc
+endfunc
function ff_put_pixels8_x2_no_rnd_armv6, export=1
push {r4-r9, lr}
bne 1b
pop {r4-r9, pc}
-.endfunc
+endfunc
function ff_put_pixels8_y2_no_rnd_armv6, export=1
push {r4-r9, lr}
bne 1b
pop {r4-r9, pc}
-.endfunc
+endfunc
function ff_avg_pixels8_armv6, export=1
pld [r1, r2]
strd r6, r7, [r0], r2
pop {r4-r10, pc}
-.endfunc
+endfunc
function ff_add_pixels_clamped_armv6, export=1
push {r4-r8,lr}
strd r6, r7, [r1], r2
bgt 1b
pop {r4-r8,pc}
-.endfunc
+endfunc
function ff_get_pixels_armv6, export=1
pld [r1, r2]
bgt 1b
pop {r4-r8, pc}
-.endfunc
+endfunc
function ff_diff_pixels_armv6, export=1
pld [r1, r3]
bgt 1b
pop {r4-r9, pc}
-.endfunc
+endfunc
function ff_pix_abs16_armv6, export=1
ldr r0, [sp]
2:
add r0, r12, lr
pop {r4-r9, pc}
-.endfunc
+endfunc
function ff_pix_abs16_x2_armv6, export=1
ldr r12, [sp]
bgt 1b
pop {r4-r11, pc}
-.endfunc
+endfunc
.macro usad_y2 p0, p1, p2, p3, n0, n1, n2, n3
ldr \n0, [r2]
bgt 1b
pop {r4-r11, pc}
-.endfunc
+endfunc
function ff_pix_abs8_armv6, export=1
pld [r2, r3]
usada8 lr, r9, r7, lr
add r0, r0, lr
pop {r4-r9, pc}
-.endfunc
+endfunc
function ff_sse16_armv6, export=1
ldr r12, [sp]
bgt 1b
pop {r4-r9, pc}
-.endfunc
+endfunc
function ff_pix_norm1_armv6, export=1
push {r4-r6, lr}
mov r0, lr
pop {r4-r6, pc}
-.endfunc
+endfunc
function ff_pix_sum_armv6, export=1
push {r4-r7, lr}
usada8 r3, r7, lr, r3
add r0, r2, r3
pop {r4-r7, pc}
-.endfunc
+endfunc