Mercurial > libavcodec.hg
changeset 8625:6f1b210e58d1 libavcodec
Alpha: fix pix_abs16
author | mru |
---|---|
date | Sun, 18 Jan 2009 18:31:52 +0000 |
parents | b1663f732e67 |
children | 8d425ee85ddb |
files | alpha/dsputil_alpha.c alpha/motion_est_mvi_asm.S |
diffstat | 2 files changed, 55 insertions(+), 62 deletions(-) [+] |
line wrap: on
line diff
--- a/alpha/dsputil_alpha.c Sun Jan 18 07:20:12 2009 +0000 +++ b/alpha/dsputil_alpha.c Sun Jan 18 18:31:52 2009 +0000 @@ -42,7 +42,7 @@ void diff_pixels_mvi(DCTELEM *block, const uint8_t *s1, const uint8_t *s2, int stride); int pix_abs8x8_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h); -int pix_abs16x16_mvi_asm(uint8_t *pix1, uint8_t *pix2, int line_size); +int pix_abs16x16_mvi_asm(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h); int pix_abs16x16_x2_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h); int pix_abs16x16_y2_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h); int pix_abs16x16_xy2_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h); @@ -287,11 +287,6 @@ put_pixels_axp_asm(block + 8, pixels + 8, line_size, h); } -static int sad16x16_mvi(void *s, uint8_t *a, uint8_t *b, int stride) -{ - return pix_abs16x16_mvi_asm(a, b, stride); -} - void dsputil_init_alpha(DSPContext* c, AVCodecContext *avctx) { c->put_pixels_tab[0][0] = put_pixels16_axp_asm; @@ -343,10 +338,9 @@ c->get_pixels = get_pixels_mvi; c->diff_pixels = diff_pixels_mvi; - c->sad[0] = sad16x16_mvi; + c->sad[0] = pix_abs16x16_mvi_asm; c->sad[1] = pix_abs8x8_mvi; -// c->pix_abs[0][0] = pix_abs16x16_mvi_asm; //FIXME function arguments for the asm must be fixed - c->pix_abs[0][0] = sad16x16_mvi; + c->pix_abs[0][0] = pix_abs16x16_mvi_asm; c->pix_abs[1][0] = pix_abs8x8_mvi; c->pix_abs[0][1] = pix_abs16x16_x2_mvi; c->pix_abs[0][2] = pix_abs16x16_y2_mvi;
--- a/alpha/motion_est_mvi_asm.S Sun Jan 18 07:20:12 2009 +0000 +++ b/alpha/motion_est_mvi_asm.S Sun Jan 18 18:31:52 2009 +0000 @@ -60,9 +60,8 @@ jsr AT, (AT), _mcount #endif - and a1, 7, t0 + and a2, 7, t0 clr v0 - lda a3, 16 beq t0, $aligned .align 4 $unaligned: @@ -86,80 +85,80 @@ td: error right */ /* load line 0 */ - ldq_u t0, 0(a1) # left_u - ldq_u t1, 8(a1) # mid - ldq_u t2, 16(a1) # right_u - ldq t3, 0(a0) # ref left - ldq t4, 8(a0) # ref right - addq a0, a2, a0 # pix1 - addq a1, a2, a1 # pix2 + ldq_u t0, 0(a2) # left_u + ldq_u t1, 8(a2) # mid + ldq_u t2, 16(a2) # right_u + ldq t3, 0(a1) # ref left + ldq t4, 8(a1) # ref right + addq a1, a3, a1 # pix1 + addq a2, a3, a2 # pix2 /* load line 1 */ - ldq_u t5, 0(a1) # left_u - ldq_u t6, 8(a1) # mid - ldq_u t7, 16(a1) # right_u - ldq t8, 0(a0) # ref left - ldq t9, 8(a0) # ref right - addq a0, a2, a0 # pix1 - addq a1, a2, a1 # pix2 + ldq_u t5, 0(a2) # left_u + ldq_u t6, 8(a2) # mid + ldq_u t7, 16(a2) # right_u + ldq t8, 0(a1) # ref left + ldq t9, 8(a1) # ref right + addq a1, a3, a1 # pix1 + addq a2, a3, a2 # pix2 /* calc line 0 */ - extql t0, a1, t0 # left lo - extqh t1, a1, ta # left hi - extql t1, a1, tb # right lo + extql t0, a2, t0 # left lo + extqh t1, a2, ta # left hi + extql t1, a2, tb # right lo or t0, ta, t0 # left - extqh t2, a1, t2 # right hi + extqh t2, a2, t2 # right hi perr t3, t0, tc # error left or t2, tb, t2 # right perr t4, t2, td # error right addq v0, tc, v0 # add error left addq v0, td, v0 # add error left /* calc line 1 */ - extql t5, a1, t5 # left lo - extqh t6, a1, ta # left hi - extql t6, a1, tb # right lo + extql t5, a2, t5 # left lo + extqh t6, a2, ta # left hi + extql t6, a2, tb # right lo or t5, ta, t5 # left - extqh t7, a1, t7 # right hi + extqh t7, a2, t7 # right hi perr t8, t5, tc # error left or t7, tb, t7 # right perr t9, t7, td # error right addq v0, tc, v0 # add error left addq v0, td, v0 # add error left /* loop */ - subq a3, 2, a3 # h -= 2 - bne a3, $unaligned + subq a4, 2, a4 # h -= 2 + bne a4, $unaligned ret .align 4 $aligned: /* load line 0 */ - ldq t0, 0(a1) # left - ldq t1, 8(a1) # right - addq a1, a2, a1 # pix2 - ldq t2, 0(a0) # ref left - ldq t3, 8(a0) # ref right - addq a0, a2, a0 # pix1 + ldq t0, 0(a2) # left + ldq t1, 8(a2) # right + addq a2, a3, a2 # pix2 + ldq t2, 0(a1) # ref left + ldq t3, 8(a1) # ref right + addq a1, a3, a1 # pix1 /* load line 1 */ - ldq t4, 0(a1) # left - ldq t5, 8(a1) # right - addq a1, a2, a1 # pix2 - ldq t6, 0(a0) # ref left - ldq t7, 8(a0) # ref right - addq a0, a2, a0 # pix1 + ldq t4, 0(a2) # left + ldq t5, 8(a2) # right + addq a2, a3, a2 # pix2 + ldq t6, 0(a1) # ref left + ldq t7, 8(a1) # ref right + addq a1, a3, a1 # pix1 /* load line 2 */ - ldq t8, 0(a1) # left - ldq t9, 8(a1) # right - addq a1, a2, a1 # pix2 - ldq ta, 0(a0) # ref left - ldq tb, 8(a0) # ref right - addq a0, a2, a0 # pix1 + ldq t8, 0(a2) # left + ldq t9, 8(a2) # right + addq a2, a3, a2 # pix2 + ldq ta, 0(a1) # ref left + ldq tb, 8(a1) # ref right + addq a1, a3, a1 # pix1 /* load line 3 */ - ldq tc, 0(a1) # left - ldq td, 8(a1) # right - addq a1, a2, a1 # pix2 - ldq te, 0(a0) # ref left - ldq tf, 8(a0) # ref right + ldq tc, 0(a2) # left + ldq td, 8(a2) # right + addq a2, a3, a2 # pix2 + ldq te, 0(a1) # ref left + ldq a0, 8(a1) # ref right /* calc line 0 */ perr t0, t2, t0 # error left - addq a0, a2, a0 # pix1 + addq a1, a3, a1 # pix1 perr t1, t3, t1 # error right addq v0, t0, v0 # add error left /* calc line 1 */ @@ -175,11 +174,11 @@ /* calc line 3 */ perr tc, te, t0 # error left addq v0, t1, v0 # add error right - perr td, tf, t1 # error right + perr td, a0, t1 # error right addq v0, t0, v0 # add error left addq v0, t1, v0 # add error right /* loop */ - subq a3, 4, a3 # h -= 4 - bne a3, $aligned + subq a4, 4, a4 # h -= 4 + bne a4, $aligned ret .end pix_abs16x16_mvi_asm