Mercurial > libavcodec.hg
view avr32/mathops.h @ 12197:fbf4d5b1b664 libavcodec
Remove FF_MM_SSE2/3 flags for CPUs where this is generally not faster than
regular MMX code. Examples of this are the Core1 CPU. Instead, set a new flag,
FF_MM_SSE2/3SLOW, which can be checked for particular SSE2/3 functions that
have been checked specifically on such CPUs and are actually faster than
their MMX counterparts.
In addition, use this flag to enable particular VP8 and LPC SSE2 functions
that are faster than their MMX counterparts.
Based on a patch by Loren Merritt <lorenm AT u washington edu>.
author | rbultje |
---|---|
date | Mon, 19 Jul 2010 22:38:23 +0000 |
parents | 25136467a218 |
children |
line wrap: on
line source
/* * Simple math operations * Copyright (c) 2009 Mans Rullgard <mans@mansr.com> * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #ifndef AVCODEC_AVR32_MATHOPS_H #define AVCODEC_AVR32_MATHOPS_H #include <stdint.h> #include "config.h" #include "libavutil/common.h" #if HAVE_INLINE_ASM #define MULL MULL static inline av_const int MULL(int a, int b, unsigned shift) { union { int64_t x; int hl[2]; } x; __asm__ ("muls.d %0, %1, %2 \n\t" "lsr %0, %3 \n\t" "or %0, %0, %m0<<%4 \n\t" : "=r"(x) : "r"(b), "r"(a), "i"(shift), "i"(32-shift)); return x.hl[1]; } #define MULH MULH static inline av_const int MULH(int a, int b) { union { int64_t x; int hl[2]; } x; __asm__ ("muls.d %0, %1, %2" : "=r"(x.x) : "r"(a), "r"(b)); return x.hl[0]; } #define MUL64 MUL64 static inline av_const int64_t MUL64(int a, int b) { int64_t x; __asm__ ("muls.d %0, %1, %2" : "=r"(x) : "r"(a), "r"(b)); return x; } static inline av_const int64_t MAC64(int64_t d, int a, int b) { __asm__ ("macs.d %0, %1, %2" : "+r"(d) : "r"(a), "r"(b)); return d; } #define MAC64(d, a, b) ((d) = MAC64(d, a, b)) #define MLS64(d, a, b) MAC64(d, -(a), b) static inline av_const int MAC16(int d, int a, int b) { __asm__ ("machh.w %0, %1:b, %2:b" : "+r"(d) : "r"(a), "r"(b)); return d; } #define MAC16(d, a, b) ((d) = MAC16(d, a, b)) #define MLS16(d, a, b) MAC16(d, -(a), b) #define MUL16 MUL16 static inline av_const int MUL16(int a, int b) { int d; __asm__ ("mulhh.w %0, %1:b, %2:b" : "=r"(d) : "r"(a), "r"(b)); return d; } #define mid_pred mid_pred static inline av_const int mid_pred(int a, int b, int c) { int m; __asm__ ("mov %0, %2 \n\t" "cp.w %1, %2 \n\t" "movgt %0, %1 \n\t" "movgt %1, %2 \n\t" "cp.w %1, %3 \n\t" "movle %1, %3 \n\t" "cp.w %0, %1 \n\t" "movgt %0, %1 \n\t" : "=&r"(m), "+r"(a) : "r"(b), "r"(c)); return m; } #endif /* HAVE_INLINE_ASM */ #endif /* AVCODEC_AVR32_MATHOPS_H */