Mercurial > libavcodec.hg
changeset 12100:db9ef48dc0e4 libavcodec
Move SSE optimized 32-point DCT to its own file. Should fix breakage with YASM
disabled.
author | vitor |
---|---|
date | Tue, 06 Jul 2010 17:48:23 +0000 |
parents | 1bf322283429 |
children | ab330d7112ad |
files | x86/Makefile x86/dct32_sse.c x86/fft.c x86/fft_sse.c |
diffstat | 4 files changed, 298 insertions(+), 266 deletions(-) [+] |
line wrap: on
line diff
--- a/x86/Makefile Tue Jul 06 16:58:54 2010 +0000 +++ b/x86/Makefile Tue Jul 06 17:48:23 2010 +0000 @@ -46,3 +46,5 @@ x86/motion_est_mmx.o \ x86/mpegvideo_mmx.o \ x86/simple_idct_mmx.o \ + +MMX-OBJS-$(CONFIG_DCT) += x86/dct32_sse.o
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/x86/dct32_sse.c Tue Jul 06 17:48:23 2010 +0000 @@ -0,0 +1,294 @@ +/* + * 32 point SSE-optimized DCT transform + * Copyright (c) 2010 Vitor Sessak + * + * This file is part of FFmpeg. + * + * FFmpeg is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * FFmpeg is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with FFmpeg; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + +#include <stdint.h> + +#include "libavutil/x86_cpu.h" +#include "libavutil/mem.h" +#include "libavcodec/dsputil.h" +#include "fft.h" + +DECLARE_ALIGNED(16, static const float, b1)[] = { + 0.500603, 0.505471, 0.515447, 0.531043, + 0.553104, 0.582935, 0.622504, 0.674808, + -1.169440, -0.972568, -0.839350, -0.744536, + -10.190008, -3.407609, -2.057781, -1.484165, + 0.502419, 0.522499, 0.566944, 0.646822, + 0.788155, 1.060678, 1.722447, 5.101149, + 0.509796, 0.601345, 0.899976, 2.562916, + 1.000000, 1.000000, 1.306563, 0.541196, + 1.000000, 0.707107, 1.000000, -0.707107 +}; + +DECLARE_ALIGNED(16, static const int32_t, smask)[4] = { + 0, 0, 0x80000000, 0x80000000 +}; + +/* butterfly operator */ +#define BUTTERFLY(a,b,c,tmp) \ + "movaps %%" #a ", %%" #tmp " \n\t" \ + "subps %%" #b ", %%" #a " \n\t" \ + "addps %%" #tmp ", %%" #b " \n\t" \ + "mulps " #c ", %%" #a " \n\t" + +///* Same as BUTTERFLY when vectors a and b overlap */ +#define BUTTERFLY0(val, mask, cos, tmp, shuf) \ + "movaps %%" #val ", %%" #tmp " \n\t" \ + "shufps " #shuf ", %%" #val ",%%" #val " \n\t" \ + "xorps %%" #mask ", %%" #tmp " \n\t" /* flip signs */ \ + "addps %%" #tmp ", %%" #val " \n\t" \ + "mulps %%" #cos ", %%" #val " \n\t" + +#define BUTTERFLY2(val, mask, cos, tmp) BUTTERFLY0(val, mask, cos, tmp, $0x1b) +#define BUTTERFLY3(val, mask, cos, tmp) BUTTERFLY0(val, mask, cos, tmp, $0xb1) + +void ff_dct32_float_sse(FFTSample *out, const FFTSample *in) +{ + int32_t tmp1 = 0; + __asm__ volatile( + /* pass 1 */ + + "movaps (%4), %%xmm0 \n\t" + "movaps 112(%4), %%xmm1 \n\t" + "shufps $0x1b, %%xmm1, %%xmm1 \n\t" + BUTTERFLY(xmm0, xmm1, (%2), xmm3) + + "movaps 64(%4), %%xmm7 \n\t" + "movaps 48(%4), %%xmm4 \n\t" + "shufps $0x1b, %%xmm4, %%xmm4 \n\t" + BUTTERFLY(xmm7, xmm4, 48(%2), xmm3) + + + /* pass 2 */ + "movaps 64(%2), %%xmm2 \n\t" + BUTTERFLY(xmm1, xmm4, %%xmm2, xmm3) + "movaps %%xmm1, 48(%1) \n\t" + "movaps %%xmm4, (%1) \n\t" + + /* pass 1 */ + "movaps 16(%4), %%xmm1 \n\t" + "movaps 96(%4), %%xmm6 \n\t" + "shufps $0x1b, %%xmm6, %%xmm6 \n\t" + BUTTERFLY(xmm1, xmm6, 16(%2), xmm3) + + "movaps 80(%4), %%xmm4 \n\t" + "movaps 32(%4), %%xmm5 \n\t" + "shufps $0x1b, %%xmm5, %%xmm5 \n\t" + BUTTERFLY(xmm4, xmm5, 32(%2), xmm3) + + /* pass 2 */ + BUTTERFLY(xmm0, xmm7, %%xmm2, xmm3) + + "movaps 80(%2), %%xmm2 \n\t" + BUTTERFLY(xmm6, xmm5, %%xmm2, xmm3) + + BUTTERFLY(xmm1, xmm4, %%xmm2, xmm3) + + /* pass 3 */ + "movaps 96(%2), %%xmm2 \n\t" + "shufps $0x1b, %%xmm1, %%xmm1 \n\t" + BUTTERFLY(xmm0, xmm1, %%xmm2, xmm3) + "movaps %%xmm0, 112(%1) \n\t" + "movaps %%xmm1, 96(%1) \n\t" + + "movaps 0(%1), %%xmm0 \n\t" + "shufps $0x1b, %%xmm5, %%xmm5 \n\t" + BUTTERFLY(xmm0, xmm5, %%xmm2, xmm3) + + "movaps 48(%1), %%xmm1 \n\t" + "shufps $0x1b, %%xmm6, %%xmm6 \n\t" + BUTTERFLY(xmm1, xmm6, %%xmm2, xmm3) + "movaps %%xmm1, 48(%1) \n\t" + + "shufps $0x1b, %%xmm4, %%xmm4 \n\t" + BUTTERFLY(xmm7, xmm4, %%xmm2, xmm3) + + /* pass 4 */ + "movaps (%3), %%xmm3 \n\t" + "movaps 112(%2), %%xmm2 \n\t" + + BUTTERFLY2(xmm5, xmm3, xmm2, xmm1) + + BUTTERFLY2(xmm0, xmm3, xmm2, xmm1) + "movaps %%xmm0, 16(%1) \n\t" + + BUTTERFLY2(xmm6, xmm3, xmm2, xmm1) + "movaps %%xmm6, 32(%1) \n\t" + + "movaps 48(%1), %%xmm0 \n\t" + BUTTERFLY2(xmm0, xmm3, xmm2, xmm1) + "movaps %%xmm0, 48(%1) \n\t" + + BUTTERFLY2(xmm4, xmm3, xmm2, xmm1) + + BUTTERFLY2(xmm7, xmm3, xmm2, xmm1) + + "movaps 96(%1), %%xmm6 \n\t" + BUTTERFLY2(xmm6, xmm3, xmm2, xmm1) + + "movaps 112(%1), %%xmm0 \n\t" + BUTTERFLY2(xmm0, xmm3, xmm2, xmm1) + + /* pass 5 */ + "movaps 128(%2), %%xmm2 \n\t" + "shufps $0xCC, %%xmm3,%%xmm3 \n\t" + + BUTTERFLY3(xmm5, xmm3, xmm2, xmm1) + "movaps %%xmm5, (%1) \n\t" + + "movaps 16(%1), %%xmm1 \n\t" + BUTTERFLY3(xmm1, xmm3, xmm2, xmm5) + "movaps %%xmm1, 16(%1) \n\t" + + BUTTERFLY3(xmm4, xmm3, xmm2, xmm5) + "movaps %%xmm4, 64(%1) \n\t" + + BUTTERFLY3(xmm7, xmm3, xmm2, xmm5) + "movaps %%xmm7, 80(%1) \n\t" + + "movaps 32(%1), %%xmm5 \n\t" + BUTTERFLY3(xmm5, xmm3, xmm2, xmm7) + "movaps %%xmm5, 32(%1) \n\t" + + "movaps 48(%1), %%xmm4 \n\t" + BUTTERFLY3(xmm4, xmm3, xmm2, xmm7) + "movaps %%xmm4, 48(%1) \n\t" + + BUTTERFLY3(xmm6, xmm3, xmm2, xmm7) + "movaps %%xmm6, 96(%1) \n\t" + + BUTTERFLY3(xmm0, xmm3, xmm2, xmm7) + "movaps %%xmm0, 112(%1) \n\t" + + + /* pass 6, no SIMD... */ + "movss 56(%1), %%xmm3 \n\t" + "movl 4(%1), %0 \n\t" + "addss 60(%1), %%xmm3 \n\t" + "movss 72(%1), %%xmm7 \n\t" + "addss %%xmm3, %%xmm4 \n\t" + "movss 52(%1), %%xmm2 \n\t" + "addss %%xmm3, %%xmm2 \n\t" + "movss 24(%1), %%xmm3 \n\t" + "addss 28(%1), %%xmm3 \n\t" + "addss 76(%1), %%xmm7 \n\t" + "addss %%xmm3, %%xmm1 \n\t" + "addss %%xmm4, %%xmm5 \n\t" + "movss %%xmm1, 16(%1) \n\t" + "movss 20(%1), %%xmm1 \n\t" + "addss %%xmm3, %%xmm1 \n\t" + "movss 40(%1), %%xmm3 \n\t" + "movss %%xmm1, 48(%1) \n\t" + "addss 44(%1), %%xmm3 \n\t" + "movss 20(%1), %%xmm1 \n\t" + "addss %%xmm3, %%xmm4 \n\t" + "addss %%xmm2, %%xmm3 \n\t" + "addss 28(%1), %%xmm1 \n\t" + "movss %%xmm3, 40(%1) \n\t" + "addss 36(%1), %%xmm2 \n\t" + "movss 8(%1), %%xmm3 \n\t" + "movss %%xmm2, 56(%1) \n\t" + "addss 12(%1), %%xmm3 \n\t" + "movss %%xmm5, 8(%1) \n\t" + "movss %%xmm3, 32(%1) \n\t" + "movss 52(%1), %%xmm2 \n\t" + "movss 80(%1), %%xmm3 \n\t" + "movss 120(%1), %%xmm5 \n\t" + "movss %%xmm1, 80(%1) \n\t" + "movss %%xmm4, 24(%1) \n\t" + "addss 124(%1), %%xmm5 \n\t" + "movss 64(%1), %%xmm1 \n\t" + "addss 60(%1), %%xmm2 \n\t" + "addss %%xmm5, %%xmm0 \n\t" + "addss 116(%1), %%xmm5 \n\t" + "movl %0, 64(%1) \n\t" + "addss %%xmm0, %%xmm6 \n\t" + "addss %%xmm6, %%xmm1 \n\t" + "movl 12(%1), %0 \n\t" + "movss %%xmm1, 4(%1) \n\t" + "movss 88(%1), %%xmm1 \n\t" + "movl %0, 96(%1) \n\t" + "addss 92(%1), %%xmm1 \n\t" + "movss 104(%1), %%xmm4 \n\t" + "movl 28(%1), %0 \n\t" + "addss 108(%1), %%xmm4 \n\t" + "addss %%xmm4, %%xmm0 \n\t" + "addss %%xmm1, %%xmm3 \n\t" + "addss 84(%1), %%xmm1 \n\t" + "addss %%xmm5, %%xmm4 \n\t" + "addss %%xmm3, %%xmm6 \n\t" + "addss %%xmm0, %%xmm3 \n\t" + "addss %%xmm7, %%xmm0 \n\t" + "addss 100(%1), %%xmm5 \n\t" + "addss %%xmm4, %%xmm7 \n\t" + "movl %0, 112(%1) \n\t" + "movss %%xmm0, 28(%1) \n\t" + "movss 36(%1), %%xmm0 \n\t" + "movss %%xmm7, 36(%1) \n\t" + "addss %%xmm1, %%xmm4 \n\t" + "movss 116(%1), %%xmm7 \n\t" + "addss %%xmm2, %%xmm0 \n\t" + "addss 124(%1), %%xmm7 \n\t" + "movss %%xmm0, 72(%1) \n\t" + "movss 44(%1), %%xmm0 \n\t" + "movss %%xmm6, 12(%1) \n\t" + "movss %%xmm3, 20(%1) \n\t" + "addss %%xmm0, %%xmm2 \n\t" + "movss %%xmm4, 44(%1) \n\t" + "movss %%xmm2, 88(%1) \n\t" + "addss 60(%1), %%xmm0 \n\t" + "movl 60(%1), %0 \n\t" + "movl %0, 120(%1) \n\t" + "movss %%xmm0, 104(%1) \n\t" + "addss %%xmm5, %%xmm1 \n\t" + "addss 68(%1), %%xmm5 \n\t" + "movss %%xmm1, 52(%1) \n\t" + "movss %%xmm5, 60(%1) \n\t" + "movss 68(%1), %%xmm1 \n\t" + "movss 100(%1), %%xmm5 \n\t" + "addss %%xmm7, %%xmm5 \n\t" + "addss 108(%1), %%xmm7 \n\t" + "addss %%xmm5, %%xmm1 \n\t" + "movss 84(%1), %%xmm2 \n\t" + "addss 92(%1), %%xmm2 \n\t" + "addss %%xmm2, %%xmm5 \n\t" + "movss %%xmm1, 68(%1) \n\t" + "addss %%xmm7, %%xmm2 \n\t" + "movss 76(%1), %%xmm1 \n\t" + "movss %%xmm2, 84(%1) \n\t" + "movss %%xmm5, 76(%1) \n\t" + "movss 108(%1), %%xmm2 \n\t" + "addss %%xmm1, %%xmm7 \n\t" + "addss 124(%1), %%xmm2 \n\t" + "addss %%xmm2, %%xmm1 \n\t" + "addss 92(%1), %%xmm2 \n\t" + "movss %%xmm1, 100(%1) \n\t" + "movss %%xmm2, 108(%1) \n\t" + "movss 92(%1), %%xmm2 \n\t" + "movss %%xmm7, 92(%1) \n\t" + "addss 124(%1), %%xmm2 \n\t" + "movss %%xmm2, 116(%1) \n\t" + :"+&r"(tmp1) + :"r"(out), "r"(b1), "r"(smask), "r"(in) + :"memory" + ); +} +
--- a/x86/fft.c Tue Jul 06 16:58:54 2010 +0000 +++ b/x86/fft.c Tue Jul 06 17:48:23 2010 +0000 @@ -43,10 +43,12 @@ #endif } +#if CONFIG_DCT av_cold void ff_dct_init_mmx(DCTContext *s) { int has_vectors = mm_support(); if (has_vectors & FF_MM_SSE && HAVE_SSE) s->dct32 = ff_dct32_float_sse; } +#endif
--- a/x86/fft_sse.c Tue Jul 06 16:58:54 2010 +0000 +++ b/x86/fft_sse.c Tue Jul 06 17:48:23 2010 +0000 @@ -20,7 +20,6 @@ */ #include "libavutil/x86_cpu.h" -#include "libavutil/common.h" #include "libavcodec/dsputil.h" #include "fft.h" @@ -202,268 +201,3 @@ ); } -DECLARE_ALIGNED(16, static const float, b1)[] = { - 0.500603, 0.505471, 0.515447, 0.531043, - 0.553104, 0.582935, 0.622504, 0.674808, - -1.169440, -0.972568, -0.839350, -0.744536, - -10.190008, -3.407609, -2.057781, -1.484165, - 0.502419, 0.522499, 0.566944, 0.646822, - 0.788155, 1.060678, 1.722447, 5.101149, - 0.509796, 0.601345, 0.899976, 2.562916, - 1.000000, 1.000000, 1.306563, 0.541196, - 1.000000, 0.707107, 1.000000, -0.707107 -}; - -DECLARE_ALIGNED(16, static const int32_t, smask)[4] = { - 0, 0, 0x80000000, 0x80000000 -}; - -/* butterfly operator */ -#define BUTTERFLY(a,b,c,tmp) \ - "movaps %%" #a ", %%" #tmp " \n\t" \ - "subps %%" #b ", %%" #a " \n\t" \ - "addps %%" #tmp ", %%" #b " \n\t" \ - "mulps " #c ", %%" #a " \n\t" - -///* Same as BUTTERFLY when vectors a and b overlap */ -#define BUTTERFLY0(val, mask, cos, tmp, shuf) \ - "movaps %%" #val ", %%" #tmp " \n\t" \ - "shufps " #shuf ", %%" #val ",%%" #val " \n\t" \ - "xorps %%" #mask ", %%" #tmp " \n\t" /* flip signs */ \ - "addps %%" #tmp ", %%" #val " \n\t" \ - "mulps %%" #cos ", %%" #val " \n\t" - -#define BUTTERFLY2(val, mask, cos, tmp) BUTTERFLY0(val, mask, cos, tmp, $0x1b) -#define BUTTERFLY3(val, mask, cos, tmp) BUTTERFLY0(val, mask, cos, tmp, $0xb1) - -void ff_dct32_float_sse(FFTSample *out, const FFTSample *in) -{ - int32_t tmp1 = 0; - __asm__ volatile( - /* pass 1 */ - - "movaps (%4), %%xmm0 \n\t" - "movaps 112(%4), %%xmm1 \n\t" - "shufps $0x1b, %%xmm1, %%xmm1 \n\t" - BUTTERFLY(xmm0, xmm1, (%2), xmm3) - - "movaps 64(%4), %%xmm7 \n\t" - "movaps 48(%4), %%xmm4 \n\t" - "shufps $0x1b, %%xmm4, %%xmm4 \n\t" - BUTTERFLY(xmm7, xmm4, 48(%2), xmm3) - - - /* pass 2 */ - "movaps 64(%2), %%xmm2 \n\t" - BUTTERFLY(xmm1, xmm4, %%xmm2, xmm3) - "movaps %%xmm1, 48(%1) \n\t" - "movaps %%xmm4, (%1) \n\t" - - /* pass 1 */ - "movaps 16(%4), %%xmm1 \n\t" - "movaps 96(%4), %%xmm6 \n\t" - "shufps $0x1b, %%xmm6, %%xmm6 \n\t" - BUTTERFLY(xmm1, xmm6, 16(%2), xmm3) - - "movaps 80(%4), %%xmm4 \n\t" - "movaps 32(%4), %%xmm5 \n\t" - "shufps $0x1b, %%xmm5, %%xmm5 \n\t" - BUTTERFLY(xmm4, xmm5, 32(%2), xmm3) - - /* pass 2 */ - BUTTERFLY(xmm0, xmm7, %%xmm2, xmm3) - - "movaps 80(%2), %%xmm2 \n\t" - BUTTERFLY(xmm6, xmm5, %%xmm2, xmm3) - - BUTTERFLY(xmm1, xmm4, %%xmm2, xmm3) - - /* pass 3 */ - "movaps 96(%2), %%xmm2 \n\t" - "shufps $0x1b, %%xmm1, %%xmm1 \n\t" - BUTTERFLY(xmm0, xmm1, %%xmm2, xmm3) - "movaps %%xmm0, 112(%1) \n\t" - "movaps %%xmm1, 96(%1) \n\t" - - "movaps 0(%1), %%xmm0 \n\t" - "shufps $0x1b, %%xmm5, %%xmm5 \n\t" - BUTTERFLY(xmm0, xmm5, %%xmm2, xmm3) - - "movaps 48(%1), %%xmm1 \n\t" - "shufps $0x1b, %%xmm6, %%xmm6 \n\t" - BUTTERFLY(xmm1, xmm6, %%xmm2, xmm3) - "movaps %%xmm1, 48(%1) \n\t" - - "shufps $0x1b, %%xmm4, %%xmm4 \n\t" - BUTTERFLY(xmm7, xmm4, %%xmm2, xmm3) - - /* pass 4 */ - "movaps (%3), %%xmm3 \n\t" - "movaps 112(%2), %%xmm2 \n\t" - - BUTTERFLY2(xmm5, xmm3, xmm2, xmm1) - - BUTTERFLY2(xmm0, xmm3, xmm2, xmm1) - "movaps %%xmm0, 16(%1) \n\t" - - BUTTERFLY2(xmm6, xmm3, xmm2, xmm1) - "movaps %%xmm6, 32(%1) \n\t" - - "movaps 48(%1), %%xmm0 \n\t" - BUTTERFLY2(xmm0, xmm3, xmm2, xmm1) - "movaps %%xmm0, 48(%1) \n\t" - - BUTTERFLY2(xmm4, xmm3, xmm2, xmm1) - - BUTTERFLY2(xmm7, xmm3, xmm2, xmm1) - - "movaps 96(%1), %%xmm6 \n\t" - BUTTERFLY2(xmm6, xmm3, xmm2, xmm1) - - "movaps 112(%1), %%xmm0 \n\t" - BUTTERFLY2(xmm0, xmm3, xmm2, xmm1) - - /* pass 5 */ - "movaps 128(%2), %%xmm2 \n\t" - "shufps $0xCC, %%xmm3,%%xmm3 \n\t" - - BUTTERFLY3(xmm5, xmm3, xmm2, xmm1) - "movaps %%xmm5, (%1) \n\t" - - "movaps 16(%1), %%xmm1 \n\t" - BUTTERFLY3(xmm1, xmm3, xmm2, xmm5) - "movaps %%xmm1, 16(%1) \n\t" - - BUTTERFLY3(xmm4, xmm3, xmm2, xmm5) - "movaps %%xmm4, 64(%1) \n\t" - - BUTTERFLY3(xmm7, xmm3, xmm2, xmm5) - "movaps %%xmm7, 80(%1) \n\t" - - "movaps 32(%1), %%xmm5 \n\t" - BUTTERFLY3(xmm5, xmm3, xmm2, xmm7) - "movaps %%xmm5, 32(%1) \n\t" - - "movaps 48(%1), %%xmm4 \n\t" - BUTTERFLY3(xmm4, xmm3, xmm2, xmm7) - "movaps %%xmm4, 48(%1) \n\t" - - BUTTERFLY3(xmm6, xmm3, xmm2, xmm7) - "movaps %%xmm6, 96(%1) \n\t" - - BUTTERFLY3(xmm0, xmm3, xmm2, xmm7) - "movaps %%xmm0, 112(%1) \n\t" - - - /* pass 6, no SIMD... */ - "movss 56(%1), %%xmm3 \n\t" - "movl 4(%1), %0 \n\t" - "addss 60(%1), %%xmm3 \n\t" - "movss 72(%1), %%xmm7 \n\t" - "addss %%xmm3, %%xmm4 \n\t" - "movss 52(%1), %%xmm2 \n\t" - "addss %%xmm3, %%xmm2 \n\t" - "movss 24(%1), %%xmm3 \n\t" - "addss 28(%1), %%xmm3 \n\t" - "addss 76(%1), %%xmm7 \n\t" - "addss %%xmm3, %%xmm1 \n\t" - "addss %%xmm4, %%xmm5 \n\t" - "movss %%xmm1, 16(%1) \n\t" - "movss 20(%1), %%xmm1 \n\t" - "addss %%xmm3, %%xmm1 \n\t" - "movss 40(%1), %%xmm3 \n\t" - "movss %%xmm1, 48(%1) \n\t" - "addss 44(%1), %%xmm3 \n\t" - "movss 20(%1), %%xmm1 \n\t" - "addss %%xmm3, %%xmm4 \n\t" - "addss %%xmm2, %%xmm3 \n\t" - "addss 28(%1), %%xmm1 \n\t" - "movss %%xmm3, 40(%1) \n\t" - "addss 36(%1), %%xmm2 \n\t" - "movss 8(%1), %%xmm3 \n\t" - "movss %%xmm2, 56(%1) \n\t" - "addss 12(%1), %%xmm3 \n\t" - "movss %%xmm5, 8(%1) \n\t" - "movss %%xmm3, 32(%1) \n\t" - "movss 52(%1), %%xmm2 \n\t" - "movss 80(%1), %%xmm3 \n\t" - "movss 120(%1), %%xmm5 \n\t" - "movss %%xmm1, 80(%1) \n\t" - "movss %%xmm4, 24(%1) \n\t" - "addss 124(%1), %%xmm5 \n\t" - "movss 64(%1), %%xmm1 \n\t" - "addss 60(%1), %%xmm2 \n\t" - "addss %%xmm5, %%xmm0 \n\t" - "addss 116(%1), %%xmm5 \n\t" - "movl %0, 64(%1) \n\t" - "addss %%xmm0, %%xmm6 \n\t" - "addss %%xmm6, %%xmm1 \n\t" - "movl 12(%1), %0 \n\t" - "movss %%xmm1, 4(%1) \n\t" - "movss 88(%1), %%xmm1 \n\t" - "movl %0, 96(%1) \n\t" - "addss 92(%1), %%xmm1 \n\t" - "movss 104(%1), %%xmm4 \n\t" - "movl 28(%1), %0 \n\t" - "addss 108(%1), %%xmm4 \n\t" - "addss %%xmm4, %%xmm0 \n\t" - "addss %%xmm1, %%xmm3 \n\t" - "addss 84(%1), %%xmm1 \n\t" - "addss %%xmm5, %%xmm4 \n\t" - "addss %%xmm3, %%xmm6 \n\t" - "addss %%xmm0, %%xmm3 \n\t" - "addss %%xmm7, %%xmm0 \n\t" - "addss 100(%1), %%xmm5 \n\t" - "addss %%xmm4, %%xmm7 \n\t" - "movl %0, 112(%1) \n\t" - "movss %%xmm0, 28(%1) \n\t" - "movss 36(%1), %%xmm0 \n\t" - "movss %%xmm7, 36(%1) \n\t" - "addss %%xmm1, %%xmm4 \n\t" - "movss 116(%1), %%xmm7 \n\t" - "addss %%xmm2, %%xmm0 \n\t" - "addss 124(%1), %%xmm7 \n\t" - "movss %%xmm0, 72(%1) \n\t" - "movss 44(%1), %%xmm0 \n\t" - "movss %%xmm6, 12(%1) \n\t" - "movss %%xmm3, 20(%1) \n\t" - "addss %%xmm0, %%xmm2 \n\t" - "movss %%xmm4, 44(%1) \n\t" - "movss %%xmm2, 88(%1) \n\t" - "addss 60(%1), %%xmm0 \n\t" - "movl 60(%1), %0 \n\t" - "movl %0, 120(%1) \n\t" - "movss %%xmm0, 104(%1) \n\t" - "addss %%xmm5, %%xmm1 \n\t" - "addss 68(%1), %%xmm5 \n\t" - "movss %%xmm1, 52(%1) \n\t" - "movss %%xmm5, 60(%1) \n\t" - "movss 68(%1), %%xmm1 \n\t" - "movss 100(%1), %%xmm5 \n\t" - "addss %%xmm7, %%xmm5 \n\t" - "addss 108(%1), %%xmm7 \n\t" - "addss %%xmm5, %%xmm1 \n\t" - "movss 84(%1), %%xmm2 \n\t" - "addss 92(%1), %%xmm2 \n\t" - "addss %%xmm2, %%xmm5 \n\t" - "movss %%xmm1, 68(%1) \n\t" - "addss %%xmm7, %%xmm2 \n\t" - "movss 76(%1), %%xmm1 \n\t" - "movss %%xmm2, 84(%1) \n\t" - "movss %%xmm5, 76(%1) \n\t" - "movss 108(%1), %%xmm2 \n\t" - "addss %%xmm1, %%xmm7 \n\t" - "addss 124(%1), %%xmm2 \n\t" - "addss %%xmm2, %%xmm1 \n\t" - "addss 92(%1), %%xmm2 \n\t" - "movss %%xmm1, 100(%1) \n\t" - "movss %%xmm2, 108(%1) \n\t" - "movss 92(%1), %%xmm2 \n\t" - "movss %%xmm7, 92(%1) \n\t" - "addss 124(%1), %%xmm2 \n\t" - "movss %%xmm2, 116(%1) \n\t" - :"+&r"(tmp1) - :"r"(out), "r"(b1), "r"(smask), "r"(in) - :"memory" - ); -}