Mercurial > libavcodec.hg
view alpha/motion_est_alpha.c @ 2601:c31a28f27d9a libavcodec
increasing precission of the quantization parameter
this is needed as the quantization stepsize for each subband is also in this precission and insignificant changes to the wavelet like scaling its coefficients slightly differently would lead to wildly variing PSNR and bitrate
note, a encoder could also simply choose to leave the least significant bits of the quantization parameters zero which would give the exact previous behaviour except a y very tiny number of bits in the header
author | michael |
---|---|
date | Sat, 09 Apr 2005 22:15:48 +0000 |
parents | dea5b2946999 |
children | ef2149182f1c |
line wrap: on
line source
/* * Alpha optimized DSP utils * Copyright (c) 2002 Falk Hueffner <falk@debian.org> * * This library is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2 of the License, or (at your option) any later version. * * This library is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with this library; if not, write to the Free Software * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ #include "asm.h" #include "../dsputil.h" void get_pixels_mvi(DCTELEM *restrict block, const uint8_t *restrict pixels, int line_size) { int h = 8; do { uint64_t p; p = ldq(pixels); stq(unpkbw(p), block); stq(unpkbw(p >> 32), block + 4); pixels += line_size; block += 8; } while (--h); } void diff_pixels_mvi(DCTELEM *block, const uint8_t *s1, const uint8_t *s2, int stride) { int h = 8; uint64_t mask = 0x4040; mask |= mask << 16; mask |= mask << 32; do { uint64_t x, y, c, d, a; uint64_t signs; x = ldq(s1); y = ldq(s2); c = cmpbge(x, y); d = x - y; a = zap(mask, c); /* We use 0x4040404040404040 here... */ d += 4 * a; /* ...so we can use s4addq here. */ signs = zap(-1, c); stq(unpkbw(d) | (unpkbw(signs) << 8), block); stq(unpkbw(d >> 32) | (unpkbw(signs >> 32) << 8), block + 4); s1 += stride; s2 += stride; block += 8; } while (--h); } static inline uint64_t avg2(uint64_t a, uint64_t b) { return (a | b) - (((a ^ b) & BYTE_VEC(0xfe)) >> 1); } static inline uint64_t avg4(uint64_t l1, uint64_t l2, uint64_t l3, uint64_t l4) { uint64_t r1 = ((l1 & ~BYTE_VEC(0x03)) >> 2) + ((l2 & ~BYTE_VEC(0x03)) >> 2) + ((l3 & ~BYTE_VEC(0x03)) >> 2) + ((l4 & ~BYTE_VEC(0x03)) >> 2); uint64_t r2 = (( (l1 & BYTE_VEC(0x03)) + (l2 & BYTE_VEC(0x03)) + (l3 & BYTE_VEC(0x03)) + (l4 & BYTE_VEC(0x03)) + BYTE_VEC(0x02)) >> 2) & BYTE_VEC(0x03); return r1 + r2; } int pix_abs8x8_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h) { int result = 0; if ((size_t) pix2 & 0x7) { /* works only when pix2 is actually unaligned */ do { /* do 8 pixel a time */ uint64_t p1, p2; p1 = ldq(pix1); p2 = uldq(pix2); result += perr(p1, p2); pix1 += line_size; pix2 += line_size; } while (--h); } else { do { uint64_t p1, p2; p1 = ldq(pix1); p2 = ldq(pix2); result += perr(p1, p2); pix1 += line_size; pix2 += line_size; } while (--h); } return result; } #if 0 /* now done in assembly */ int pix_abs16x16_mvi(uint8_t *pix1, uint8_t *pix2, int line_size) { int result = 0; int h = 16; if ((size_t) pix2 & 0x7) { /* works only when pix2 is actually unaligned */ do { /* do 16 pixel a time */ uint64_t p1_l, p1_r, p2_l, p2_r; uint64_t t; p1_l = ldq(pix1); p1_r = ldq(pix1 + 8); t = ldq_u(pix2 + 8); p2_l = extql(ldq_u(pix2), pix2) | extqh(t, pix2); p2_r = extql(t, pix2) | extqh(ldq_u(pix2 + 16), pix2); pix1 += line_size; pix2 += line_size; result += perr(p1_l, p2_l) + perr(p1_r, p2_r); } while (--h); } else { do { uint64_t p1_l, p1_r, p2_l, p2_r; p1_l = ldq(pix1); p1_r = ldq(pix1 + 8); p2_l = ldq(pix2); p2_r = ldq(pix2 + 8); pix1 += line_size; pix2 += line_size; result += perr(p1_l, p2_l) + perr(p1_r, p2_r); } while (--h); } return result; } #endif int pix_abs16x16_x2_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h) { int result = 0; uint64_t disalign = (size_t) pix2 & 0x7; switch (disalign) { case 0: do { uint64_t p1_l, p1_r, p2_l, p2_r; uint64_t l, r; p1_l = ldq(pix1); p1_r = ldq(pix1 + 8); l = ldq(pix2); r = ldq(pix2 + 8); p2_l = avg2(l, (l >> 8) | ((uint64_t) r << 56)); p2_r = avg2(r, (r >> 8) | ((uint64_t) pix2[16] << 56)); pix1 += line_size; pix2 += line_size; result += perr(p1_l, p2_l) + perr(p1_r, p2_r); } while (--h); break; case 7: /* |.......l|lllllllr|rrrrrrr*| This case is special because disalign1 would be 8, which gets treated as 0 by extqh. At least it is a bit faster that way :) */ do { uint64_t p1_l, p1_r, p2_l, p2_r; uint64_t l, m, r; p1_l = ldq(pix1); p1_r = ldq(pix1 + 8); l = ldq_u(pix2); m = ldq_u(pix2 + 8); r = ldq_u(pix2 + 16); p2_l = avg2(extql(l, disalign) | extqh(m, disalign), m); p2_r = avg2(extql(m, disalign) | extqh(r, disalign), r); pix1 += line_size; pix2 += line_size; result += perr(p1_l, p2_l) + perr(p1_r, p2_r); } while (--h); break; default: do { uint64_t disalign1 = disalign + 1; uint64_t p1_l, p1_r, p2_l, p2_r; uint64_t l, m, r; p1_l = ldq(pix1); p1_r = ldq(pix1 + 8); l = ldq_u(pix2); m = ldq_u(pix2 + 8); r = ldq_u(pix2 + 16); p2_l = avg2(extql(l, disalign) | extqh(m, disalign), extql(l, disalign1) | extqh(m, disalign1)); p2_r = avg2(extql(m, disalign) | extqh(r, disalign), extql(m, disalign1) | extqh(r, disalign1)); pix1 += line_size; pix2 += line_size; result += perr(p1_l, p2_l) + perr(p1_r, p2_r); } while (--h); break; } return result; } int pix_abs16x16_y2_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h) { int result = 0; if ((size_t) pix2 & 0x7) { uint64_t t, p2_l, p2_r; t = ldq_u(pix2 + 8); p2_l = extql(ldq_u(pix2), pix2) | extqh(t, pix2); p2_r = extql(t, pix2) | extqh(ldq_u(pix2 + 16), pix2); do { uint64_t p1_l, p1_r, np2_l, np2_r; uint64_t t; p1_l = ldq(pix1); p1_r = ldq(pix1 + 8); pix2 += line_size; t = ldq_u(pix2 + 8); np2_l = extql(ldq_u(pix2), pix2) | extqh(t, pix2); np2_r = extql(t, pix2) | extqh(ldq_u(pix2 + 16), pix2); result += perr(p1_l, avg2(p2_l, np2_l)) + perr(p1_r, avg2(p2_r, np2_r)); pix1 += line_size; p2_l = np2_l; p2_r = np2_r; } while (--h); } else { uint64_t p2_l, p2_r; p2_l = ldq(pix2); p2_r = ldq(pix2 + 8); do { uint64_t p1_l, p1_r, np2_l, np2_r; p1_l = ldq(pix1); p1_r = ldq(pix1 + 8); pix2 += line_size; np2_l = ldq(pix2); np2_r = ldq(pix2 + 8); result += perr(p1_l, avg2(p2_l, np2_l)) + perr(p1_r, avg2(p2_r, np2_r)); pix1 += line_size; p2_l = np2_l; p2_r = np2_r; } while (--h); } return result; } int pix_abs16x16_xy2_mvi(void *v, uint8_t *pix1, uint8_t *pix2, int line_size, int h) { int result = 0; uint64_t p1_l, p1_r; uint64_t p2_l, p2_r, p2_x; p1_l = ldq(pix1); p1_r = ldq(pix1 + 8); if ((size_t) pix2 & 0x7) { /* could be optimized a lot */ p2_l = uldq(pix2); p2_r = uldq(pix2 + 8); p2_x = (uint64_t) pix2[16] << 56; } else { p2_l = ldq(pix2); p2_r = ldq(pix2 + 8); p2_x = ldq(pix2 + 16) << 56; } do { uint64_t np1_l, np1_r; uint64_t np2_l, np2_r, np2_x; pix1 += line_size; pix2 += line_size; np1_l = ldq(pix1); np1_r = ldq(pix1 + 8); if ((size_t) pix2 & 0x7) { /* could be optimized a lot */ np2_l = uldq(pix2); np2_r = uldq(pix2 + 8); np2_x = (uint64_t) pix2[16] << 56; } else { np2_l = ldq(pix2); np2_r = ldq(pix2 + 8); np2_x = ldq(pix2 + 16) << 56; } result += perr(p1_l, avg4( p2_l, ( p2_l >> 8) | ((uint64_t) p2_r << 56), np2_l, (np2_l >> 8) | ((uint64_t) np2_r << 56))) + perr(p1_r, avg4( p2_r, ( p2_r >> 8) | ((uint64_t) p2_x), np2_r, (np2_r >> 8) | ((uint64_t) np2_x))); p1_l = np1_l; p1_r = np1_r; p2_l = np2_l; p2_r = np2_r; p2_x = np2_x; } while (--h); return result; }