Mercurial > mplayer.hg
changeset 4148:3b29772a4fb2
S->C
author | nick |
---|---|
date | Mon, 14 Jan 2002 10:34:38 +0000 |
parents | 4bbdda22003d |
children | 9dde9f96f860 |
files | mp3lib/Makefile mp3lib/dct64_3dnow.c mp3lib/dct64_3dnow.s mp3lib/dct64_k7.c mp3lib/dct64_k7.s |
diffstat | 5 files changed, 1708 insertions(+), 1723 deletions(-) [+] |
line wrap: on
line diff
--- a/mp3lib/Makefile Mon Jan 14 09:32:51 2002 +0000 +++ b/mp3lib/Makefile Mon Jan 14 10:34:38 2002 +0000 @@ -20,11 +20,11 @@ #OBJS += dct64_sse.o #endif ifeq ($(TARGET_3DNOW),yes) -SRCS += dct36_3dnow.s dct64_3dnow.s +SRCS += dct36_3dnow.s dct64_3dnow.c OBJS += dct36_3dnow.o dct64_3dnow.o endif ifeq ($(TARGET_3DNOWEX),yes) -SRCS += dct36_k7.s dct64_k7.s +SRCS += dct36_k7.s dct64_k7.c OBJS += dct36_k7.o dct64_k7.o endif endif
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/mp3lib/dct64_3dnow.c Mon Jan 14 10:34:38 2002 +0000 @@ -0,0 +1,914 @@ +/* +* This code was taken from http://www.mpg123.org +* See ChangeLog of mpg123-0.59s-pre.1 for detail +* Applied to mplayer by Nick Kurshev <nickols_k@mail.ru> +* Partial 3dnow! optimization by Nick Kurshev +* +* TODO: optimize scalar 3dnow! code +* Warning: Phases 7 & 8 are not tested +*/ +#define real float /* ugly - but only way */ + +static unsigned long long int __attribute__((aligned(8))) x_plus_minus_3dnow = 0x8000000000000000ULL; +static float plus_1f = 1.0; + +void __attribute__ (( __stdcall__ )) dct64_MMX_3dnow(real *a,real *b,real *c) +{ + char tmp[256]; + __asm __volatile( +" movl %2,%%eax\n\t" + +" leal 128+%3,%%edx\n\t" +" movl %0,%%esi\n\t" +" movl %1,%%edi\n\t" +" movl $costab_mmx,%%ebx\n\t" +" leal %3,%%ecx\n\t" + +/* Phase 1*/ +" movq (%%eax), %%mm0\n\t" +" movq 8(%%eax), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 120(%%eax), %%mm1\n\t" +" movq 112(%%eax), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, (%%edx)\n\t" +" movq %%mm4, 8(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul (%%ebx), %%mm3\n\t" +" pfmul 8(%%ebx), %%mm7\n\t" +" movd %%mm3, 124(%%edx)\n\t" +" movd %%mm7, 116(%%edx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 120(%%edx)\n\t" +" movd %%mm7, 112(%%edx)\n\t" + +" movq 16(%%eax), %%mm0\n\t" +" movq 24(%%eax), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 104(%%eax), %%mm1\n\t" +" movq 96(%%eax), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 16(%%edx)\n\t" +" movq %%mm4, 24(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 16(%%ebx), %%mm3\n\t" +" pfmul 24(%%ebx), %%mm7\n\t" +" movd %%mm3, 108(%%edx)\n\t" +" movd %%mm7, 100(%%edx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 104(%%edx)\n\t" +" movd %%mm7, 96(%%edx)\n\t" + +" movq 32(%%eax), %%mm0\n\t" +" movq 40(%%eax), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 88(%%eax), %%mm1\n\t" +" movq 80(%%eax), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 32(%%edx)\n\t" +" movq %%mm4, 40(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 32(%%ebx), %%mm3\n\t" +" pfmul 40(%%ebx), %%mm7\n\t" +" movd %%mm3, 92(%%edx)\n\t" +" movd %%mm7, 84(%%edx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 88(%%edx)\n\t" +" movd %%mm7, 80(%%edx)\n\t" + +" movq 48(%%eax), %%mm0\n\t" +" movq 56(%%eax), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 72(%%eax), %%mm1\n\t" +" movq 64(%%eax), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 48(%%edx)\n\t" +" movq %%mm4, 56(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 48(%%ebx), %%mm3\n\t" +" pfmul 56(%%ebx), %%mm7\n\t" +" movd %%mm3, 76(%%edx)\n\t" +" movd %%mm7, 68(%%edx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 72(%%edx)\n\t" +" movd %%mm7, 64(%%edx)\n\t" + +/* Phase 2*/ + +" movq (%%edx), %%mm0\n\t" +" movq 8(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 56(%%edx), %%mm1\n\t" +" movq 48(%%edx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, (%%ecx)\n\t" +" movq %%mm4, 8(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 64(%%ebx), %%mm3\n\t" +" pfmul 72(%%ebx), %%mm7\n\t" +" movd %%mm3, 60(%%ecx)\n\t" +" movd %%mm7, 52(%%ecx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 56(%%ecx)\n\t" +" movd %%mm7, 48(%%ecx)\n\t" + +" movq 16(%%edx), %%mm0\n\t" +" movq 24(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 40(%%edx), %%mm1\n\t" +" movq 32(%%edx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 16(%%ecx)\n\t" +" movq %%mm4, 24(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 80(%%ebx), %%mm3\n\t" +" pfmul 88(%%ebx), %%mm7\n\t" +" movd %%mm3, 44(%%ecx)\n\t" +" movd %%mm7, 36(%%ecx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 40(%%ecx)\n\t" +" movd %%mm7, 32(%%ecx)\n\t" + +/* Phase 3*/ + +" movq 64(%%edx), %%mm0\n\t" +" movq 72(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 120(%%edx), %%mm1\n\t" +" movq 112(%%edx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 64(%%ecx)\n\t" +" movq %%mm4, 72(%%ecx)\n\t" +" pfsubr %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 64(%%ebx), %%mm3\n\t" +" pfmul 72(%%ebx), %%mm7\n\t" +" movd %%mm3, 124(%%ecx)\n\t" +" movd %%mm7, 116(%%ecx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 120(%%ecx)\n\t" +" movd %%mm7, 112(%%ecx)\n\t" + +" movq 80(%%edx), %%mm0\n\t" +" movq 88(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 104(%%edx), %%mm1\n\t" +" movq 96(%%edx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 80(%%ecx)\n\t" +" movq %%mm4, 88(%%ecx)\n\t" +" pfsubr %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 80(%%ebx), %%mm3\n\t" +" pfmul 88(%%ebx), %%mm7\n\t" +" movd %%mm3, 108(%%ecx)\n\t" +" movd %%mm7, 100(%%ecx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 104(%%ecx)\n\t" +" movd %%mm7, 96(%%ecx)\n\t" + +/* Phase 4*/ + +" movq (%%ecx), %%mm0\n\t" +" movq 8(%%ecx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 24(%%ecx), %%mm1\n\t" +" movq 16(%%ecx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, (%%edx)\n\t" +" movq %%mm4, 8(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 96(%%ebx), %%mm3\n\t" +" pfmul 104(%%ebx), %%mm7\n\t" +" movd %%mm3, 28(%%edx)\n\t" +" movd %%mm7, 20(%%edx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 24(%%edx)\n\t" +" movd %%mm7, 16(%%edx)\n\t" + +" movq 32(%%ecx), %%mm0\n\t" +" movq 40(%%ecx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 56(%%ecx), %%mm1\n\t" +" movq 48(%%ecx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 32(%%edx)\n\t" +" movq %%mm4, 40(%%edx)\n\t" +" pfsubr %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 96(%%ebx), %%mm3\n\t" +" pfmul 104(%%ebx), %%mm7\n\t" +" movd %%mm3, 60(%%edx)\n\t" +" movd %%mm7, 52(%%edx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 56(%%edx)\n\t" +" movd %%mm7, 48(%%edx)\n\t" + +" movq 64(%%ecx), %%mm0\n\t" +" movq 72(%%ecx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 88(%%ecx), %%mm1\n\t" +" movq 80(%%ecx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 64(%%edx)\n\t" +" movq %%mm4, 72(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 96(%%ebx), %%mm3\n\t" +" pfmul 104(%%ebx), %%mm7\n\t" +" movd %%mm3, 92(%%edx)\n\t" +" movd %%mm7, 84(%%edx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 88(%%edx)\n\t" +" movd %%mm7, 80(%%edx)\n\t" + +" movq 96(%%ecx), %%mm0\n\t" +" movq 104(%%ecx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 120(%%ecx), %%mm1\n\t" +" movq 112(%%ecx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 96(%%edx)\n\t" +" movq %%mm4, 104(%%edx)\n\t" +" pfsubr %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 96(%%ebx), %%mm3\n\t" +" pfmul 104(%%ebx), %%mm7\n\t" +" movd %%mm3, 124(%%edx)\n\t" +" movd %%mm7, 116(%%edx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 120(%%edx)\n\t" +" movd %%mm7, 112(%%edx)\n\t" + +/* Phase 5 */ + +" movq (%%edx), %%mm0\n\t" +" movq 16(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 8(%%edx), %%mm1\n\t" +" movq 24(%%edx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, (%%ecx)\n\t" +" movq %%mm4, 16(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 112(%%ebx), %%mm3\n\t" +" pfmul 112(%%ebx), %%mm7\n\t" +" movd %%mm3, 12(%%ecx)\n\t" +" movd %%mm7, 28(%%ecx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 8(%%ecx)\n\t" +" movd %%mm7, 24(%%ecx)\n\t" + +" movq 32(%%edx), %%mm0\n\t" +" movq 48(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 40(%%edx), %%mm1\n\t" +" movq 56(%%edx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 32(%%ecx)\n\t" +" movq %%mm4, 48(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 112(%%ebx), %%mm3\n\t" +" pfmul 112(%%ebx), %%mm7\n\t" +" movd %%mm3, 44(%%ecx)\n\t" +" movd %%mm7, 60(%%ecx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 40(%%ecx)\n\t" +" movd %%mm7, 56(%%ecx)\n\t" + +" movq 64(%%edx), %%mm0\n\t" +" movq 80(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 72(%%edx), %%mm1\n\t" +" movq 88(%%edx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 64(%%ecx)\n\t" +" movq %%mm4, 80(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 112(%%ebx), %%mm3\n\t" +" pfmul 112(%%ebx), %%mm7\n\t" +" movd %%mm3, 76(%%ecx)\n\t" +" movd %%mm7, 92(%%ecx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 72(%%ecx)\n\t" +" movd %%mm7, 88(%%ecx)\n\t" + +" movq 96(%%edx), %%mm0\n\t" +" movq 112(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 104(%%edx), %%mm1\n\t" +" movq 120(%%edx), %%mm5\n\t" +/* n.b.: pswapd*/ +" movq %%mm1, %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm5\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" punpckldq %%mm6, %%mm5\n\t" +/**/ +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 96(%%ecx)\n\t" +" movq %%mm4, 112(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 112(%%ebx), %%mm3\n\t" +" pfmul 112(%%ebx), %%mm7\n\t" +" movd %%mm3, 108(%%ecx)\n\t" +" movd %%mm7, 124(%%ecx)\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm7\n\t" +" movd %%mm3, 104(%%ecx)\n\t" +" movd %%mm7, 120(%%ecx)\n\t" + +/* Phase 6. This is the end of easy road. */ +/* Code below is coded in scalar mode. Should be optimized */ + +" movd plus_1f, %%mm6\n\t" +" punpckldq 120(%%ebx), %%mm6\n\t" /* mm6 = 1.0 | 120(%%ebx)*/ +" movq x_plus_minus_3dnow, %%mm7\n\t" /* mm7 = +1 | -1 */ + +" movq 32(%%ecx), %%mm0\n\t" +" movq 64(%%ecx), %%mm2\n\t" +" movq %%mm0, %%mm1\n\t" +" movq %%mm2, %%mm3\n\t" +" pxor %%mm7, %%mm1\n\t" +" pxor %%mm7, %%mm3\n\t" +" pfacc %%mm1, %%mm0\n\t" +" pfacc %%mm3, %%mm2\n\t" +" pfmul %%mm6, %%mm0\n\t" +" pfmul %%mm6, %%mm2\n\t" +" movq %%mm0, 32(%%edx)\n\t" +" movq %%mm2, 64(%%edx)\n\t" + +" movd 44(%%ecx), %%mm0\n\t" +" movd 40(%%ecx), %%mm2\n\t" +" movd 120(%%ebx), %%mm3\n\t" +" punpckldq 76(%%ecx), %%mm0\n\t" +" punpckldq 72(%%ecx), %%mm2\n\t" +" punpckldq %%mm3, %%mm3\n\t" +" movq %%mm0, %%mm4\n\t" +" movq %%mm2, %%mm5\n\t" +" pfsub %%mm2, %%mm0\n\t" +" pfmul %%mm3, %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" +" pfadd %%mm5, %%mm0\n\t" +" pfadd %%mm4, %%mm0\n\t" +" movq %%mm0, %%mm2\n\t" +" punpckldq %%mm1, %%mm0\n\t" +" punpckhdq %%mm1, %%mm2\n\t" +" movq %%mm0, 40(%%edx)\n\t" +" movq %%mm2, 72(%%edx)\n\t" + +" movd 48(%%ecx), %%mm3\n\t" +" movd 60(%%ecx), %%mm2\n\t" +" pfsub 52(%%ecx), %%mm3\n\t" +" pfsub 56(%%ecx), %%mm2\n\t" +" pfmul 120(%%ebx), %%mm3\n\t" +" pfmul 120(%%ebx), %%mm2\n\t" +" movq %%mm2, %%mm1\n\t" + +" pfadd 56(%%ecx), %%mm1\n\t" +" pfadd 60(%%ecx), %%mm1\n\t" +" movq %%mm1, %%mm0\n\t" + +" pfadd 48(%%ecx), %%mm0\n\t" +" pfadd 52(%%ecx), %%mm0\n\t" +" pfadd %%mm3, %%mm1\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" pfadd %%mm3, %%mm2\n\t" +" punpckldq %%mm2, %%mm0\n\t" +" movq %%mm1, 56(%%edx)\n\t" +" movq %%mm0, 48(%%edx)\n\t" + +/*---*/ + +" movd 92(%%ecx), %%mm1\n\t" +" pfsub 88(%%ecx), %%mm1\n\t" +" pfmul 120(%%ebx), %%mm1\n\t" +" movd %%mm1, 92(%%edx)\n\t" +" pfadd 92(%%ecx), %%mm1\n\t" +" pfadd 88(%%ecx), %%mm1\n\t" +" movq %%mm1, %%mm0\n\t" + +" pfadd 80(%%ecx), %%mm0\n\t" +" pfadd 84(%%ecx), %%mm0\n\t" +" movd %%mm0, 80(%%edx)\n\t" + +" movd 80(%%ecx), %%mm0\n\t" +" pfsub 84(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" pfadd %%mm0, %%mm1\n\t" +" pfadd 92(%%edx), %%mm0\n\t" +" punpckldq %%mm1, %%mm0\n\t" +" movq %%mm0, 84(%%edx)\n\t" + +" movq 96(%%ecx), %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" +" pxor %%mm7, %%mm1\n\t" +" pfacc %%mm1, %%mm0\n\t" +" pfmul %%mm6, %%mm0\n\t" +" movq %%mm0, 96(%%edx)\n\t" + +" movd 108(%%ecx), %%mm0\n\t" +" pfsub 104(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movd %%mm0, 108(%%edx)\n\t" +" pfadd 104(%%ecx), %%mm0\n\t" +" pfadd 108(%%ecx), %%mm0\n\t" +" movd %%mm0, 104(%%edx)\n\t" + +" movd 124(%%ecx), %%mm1\n\t" +" pfsub 120(%%ecx), %%mm1\n\t" +" pfmul 120(%%ebx), %%mm1\n\t" +" movd %%mm1, 124(%%edx)\n\t" +" pfadd 120(%%ecx), %%mm1\n\t" +" pfadd 124(%%ecx), %%mm1\n\t" +" movq %%mm1, %%mm0\n\t" + +" pfadd 112(%%ecx), %%mm0\n\t" +" pfadd 116(%%ecx), %%mm0\n\t" +" movd %%mm0, 112(%%edx)\n\t" + +" movd 112(%%ecx), %%mm0\n\t" +" pfsub 116(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" pfadd %%mm0,%%mm1\n\t" +" pfadd 124(%%edx), %%mm0\n\t" +" punpckldq %%mm1, %%mm0\n\t" +" movq %%mm0, 116(%%edx)\n\t" + +" jnz .L01\n\t" + +/* Phase 7*/ +/* Code below is coded in scalar mode. Should be optimized */ + +" movd (%%ecx), %%mm0\n\t" +" pfadd 4(%%ecx), %%mm0\n\t" +" movd %%mm0, 1024(%%esi)\n\t" + +" movd (%%ecx), %%mm0\n\t" +" pfsub 4(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movd %%mm0, (%%esi)\n\t" +" movd %%mm0, (%%edi)\n\t" + +" movd 12(%%ecx), %%mm0\n\t" +" pfsub 8(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movd %%mm0, 512(%%edi)\n\t" +" pfadd 12(%%ecx), %%mm0\n\t" +" pfadd 8(%%ecx), %%mm0\n\t" +" movd %%mm0, 512(%%esi)\n\t" + +" movd 16(%%ecx), %%mm0\n\t" +" pfsub 20(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movq %%mm0, %%mm3\n\t" + +" movd 28(%%ecx), %%mm0\n\t" +" pfsub 24(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movd %%mm0, 768(%%edi)\n\t" +" movq %%mm0, %%mm2\n\t" + +" pfadd 24(%%ecx), %%mm0\n\t" +" pfadd 28(%%ecx), %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" + +" pfadd 16(%%ecx), %%mm0\n\t" +" pfadd 20(%%ecx), %%mm0\n\t" +" movd %%mm0, 768(%%esi)\n\t" +" pfadd %%mm3, %%mm1\n\t" +" movd %%mm1, 256(%%esi)\n\t" +" pfadd %%mm3, %%mm2\n\t" +" movd %%mm2, 256(%%edi)\n\t" + +/* Phase 8*/ + +" movq 32(%%edx), %%mm0\n\t" +" movq 48(%%edx), %%mm1\n\t" +" pfadd 48(%%edx), %%mm0\n\t" +" pfadd 40(%%edx), %%mm1\n\t" +" movd %%mm0, 896(%%esi)\n\t" +" movd %%mm1, 640(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" psrlq $32, %%mm1\n\t" +" movd %%mm0, 128(%%edi)\n\t" +" movd %%mm1, 384(%%edi)\n\t" + +" movd 40(%%edx), %%mm0\n\t" +" pfadd 56(%%edx), %%mm0\n\t" +" movd %%mm0, 384(%%esi)\n\t" + +" movd 56(%%edx), %%mm0\n\t" +" pfadd 36(%%edx), %%mm0\n\t" +" movd %%mm0, 128(%%esi)\n\t" + +" movd 60(%%edx), %%mm0\n\t" +" movd %%mm0, 896(%%edi)\n\t" +" pfadd 44(%%edx), %%mm0\n\t" +" movd %%mm0, 640(%%edi)\n\t" + +" movq 96(%%edx), %%mm0\n\t" +" movq 112(%%edx), %%mm2\n\t" +" movq 104(%%edx), %%mm4\n\t" +" pfadd 112(%%edx), %%mm0\n\t" +" pfadd 104(%%edx), %%mm2\n\t" +" pfadd 120(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm1\n\t" +" movq %%mm2, %%mm3\n\t" +" movq %%mm4, %%mm5\n\t" +" pfadd 64(%%edx), %%mm0\n\t" +" pfadd 80(%%edx), %%mm2\n\t" +" pfadd 72(%%edx), %%mm4\n\t" +" movd %%mm0, 960(%%esi)\n\t" +" movd %%mm2, 704(%%esi)\n\t" +" movd %%mm4, 448(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" psrlq $32, %%mm2\n\t" +" psrlq $32, %%mm4\n\t" +" movd %%mm0, 64(%%edi)\n\t" +" movd %%mm2, 320(%%edi)\n\t" +" movd %%mm4, 576(%%edi)\n\t" +" pfadd 80(%%edx), %%mm1\n\t" +" pfadd 72(%%edx), %%mm3\n\t" +" pfadd 88(%%edx), %%mm5\n\t" +" movd %%mm1, 832(%%esi)\n\t" +" movd %%mm3, 576(%%esi)\n\t" +" movd %%mm5, 320(%%esi)\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm5\n\t" +" movd %%mm1, 192(%%edi)\n\t" +" movd %%mm3, 448(%%edi)\n\t" +" movd %%mm5, 704(%%edi)\n\t" + +" movd 120(%%edx), %%mm0\n\t" +" pfadd 100(%%edx), %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" +" pfadd 88(%%edx), %%mm0\n\t" +" movd %%mm0, 192(%%esi)\n\t" +" pfadd 68(%%edx), %%mm1\n\t" +" movd %%mm1, 64(%%esi)\n\t" + +" movd 124(%%edx), %%mm0\n\t" +" movd %%mm0, 960(%%edi)\n\t" +" pfadd 92(%%edx), %%mm0\n\t" +" movd %%mm0, 832(%%edi)\n\t" + +" jmp .L_bye\n\t" +".L01:\n\t" +/* Phase 9*/ + +" movq (%%ecx), %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" +" pxor %%mm7, %%mm1\n\t" +" pfacc %%mm1, %%mm0\n\t" +" pfmul %%mm6, %%mm0\n\t" +" pf2id %%mm0, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" movw %%ax, 512(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" movw %%ax, (%%esi)\n\t" + +" movd 12(%%ecx), %%mm0\n\t" +" pfsub 8(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" pf2id %%mm0, %%mm7\n\t" +" movd %%mm7, %%eax\n\t" +" movw %%ax, 256(%%edi)\n\t" +" pfadd 12(%%ecx), %%mm0\n\t" +" pfadd 8(%%ecx), %%mm0\n\t" +" pf2id %%mm0, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" movw %%ax, 256(%%esi)\n\t" + +" movd 16(%%ecx), %%mm3\n\t" +" pfsub 20(%%ecx), %%mm3\n\t" +" pfmul 120(%%ebx), %%mm3\n\t" +" movq %%mm3, %%mm2\n\t" + +" movd 28(%%ecx), %%mm2\n\t" +" pfsub 24(%%ecx), %%mm2\n\t" +" pfmul 120(%%ebx), %%mm2\n\t" +" movq %%mm2, %%mm1\n\t" + +" pf2id %%mm2, %%mm7\n\t" +" movd %%mm7, %%eax\n\t" +" movw %%ax, 384(%%edi)\n\t" + +" pfadd 24(%%ecx), %%mm1\n\t" +" pfadd 28(%%ecx), %%mm1\n\t" +" movq %%mm1, %%mm0\n\t" + +" pfadd 16(%%ecx), %%mm0\n\t" +" pfadd 20(%%ecx), %%mm0\n\t" +" pf2id %%mm0, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" movw %%ax, 384(%%esi)\n\t" +" pfadd %%mm3, %%mm1\n\t" +" pf2id %%mm1, %%mm1\n\t" +" movd %%mm1, %%eax\n\t" +" movw %%ax, 128(%%esi)\n\t" +" pfadd %%mm3, %%mm2\n\t" +" pf2id %%mm2, %%mm2\n\t" +" movd %%mm2, %%eax\n\t" +" movw %%ax, 128(%%edi)\n\t" + +/* Phase 10*/ + +" movq 32(%%edx), %%mm0\n\t" +" movq 48(%%edx), %%mm1\n\t" +" pfadd 48(%%edx), %%mm0\n\t" +" pfadd 40(%%edx), %%mm1\n\t" +" pf2id %%mm0, %%mm0\n\t" +" pf2id %%mm1, %%mm1\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm1, %%ecx\n\t" +" movw %%ax, 448(%%esi)\n\t" +" movw %%cx, 320(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" psrlq $32, %%mm1\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm1, %%ecx\n\t" +" movw %%ax, 64(%%edi)\n\t" +" movw %%cx, 192(%%edi)\n\t" + +" movd 40(%%edx), %%mm3\n\t" +" movd 56(%%edx), %%mm4\n\t" +" movd 60(%%edx), %%mm0\n\t" +" movd 44(%%edx), %%mm2\n\t" +" movd 120(%%edx), %%mm5\n\t" +" punpckldq %%mm4, %%mm3\n\t" +" punpckldq 124(%%edx), %%mm0\n\t" +" pfadd 100(%%edx), %%mm5\n\t" +" punpckldq 36(%%edx), %%mm4\n\t" +" punpckldq 92(%%edx), %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" pfadd %%mm4, %%mm3\n\t" +" pf2id %%mm0, %%mm1\n\t" +" pf2id %%mm3, %%mm3\n\t" +" pfadd 88(%%edx), %%mm5\n\t" +" movd %%mm1, %%eax\n\t" +" movd %%mm3, %%ecx\n\t" +" movw %%ax, 448(%%edi)\n\t" +" movw %%cx, 192(%%esi)\n\t" +" pf2id %%mm5, %%mm5\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm3\n\t" +" movd %%mm5, %%ebx\n\t" +" movd %%mm1, %%eax\n\t" +" movd %%mm3, %%ecx\n\t" +" movw %%bx, 96(%%esi)\n\t" +" movw %%ax, 480(%%edi)\n\t" +" movw %%cx, 64(%%esi)\n\t" +" pfadd %%mm2, %%mm0\n\t" +" pf2id %%mm0, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" pfadd 68(%%edx), %%mm6\n\t" +" movw %%ax, 320(%%edi)\n\t" +" psrlq $32, %%mm0\n\t" +" pf2id %%mm6, %%mm6\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm6, %%ebx\n\t" +" movw %%ax, 416(%%edi)\n\t" +" movw %%bx, 32(%%esi)\n\t" + +" movq 96(%%edx), %%mm0\n\t" +" movq 112(%%edx), %%mm2\n\t" +" movq 104(%%edx), %%mm4\n\t" +" pfadd %%mm2, %%mm0\n\t" +" pfadd %%mm4, %%mm2\n\t" +" pfadd 120(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm1\n\t" +" movq %%mm2, %%mm3\n\t" +" movq %%mm4, %%mm5\n\t" +" pfadd 64(%%edx), %%mm0\n\t" +" pfadd 80(%%edx), %%mm2\n\t" +" pfadd 72(%%edx), %%mm4\n\t" +" pf2id %%mm0, %%mm0\n\t" +" pf2id %%mm2, %%mm2\n\t" +" pf2id %%mm4, %%mm4\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm2, %%ecx\n\t" +" movd %%mm4, %%ebx\n\t" +" movw %%ax, 480(%%esi)\n\t" +" movw %%cx, 352(%%esi)\n\t" +" movw %%bx, 224(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" psrlq $32, %%mm2\n\t" +" psrlq $32, %%mm4\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm2, %%ecx\n\t" +" movd %%mm4, %%ebx\n\t" +" movw %%ax, 32(%%edi)\n\t" +" movw %%cx, 160(%%edi)\n\t" +" movw %%bx, 288(%%edi)\n\t" +" pfadd 80(%%edx), %%mm1\n\t" +" pfadd 72(%%edx), %%mm3\n\t" +" pfadd 88(%%edx), %%mm5\n\t" +" pf2id %%mm1, %%mm1\n\t" +" pf2id %%mm3, %%mm3\n\t" +" pf2id %%mm5, %%mm5\n\t" +" movd %%mm1, %%eax\n\t" +" movd %%mm3, %%ecx\n\t" +" movd %%mm5, %%ebx\n\t" +" movw %%ax, 416(%%esi)\n\t" +" movw %%cx, 288(%%esi)\n\t" +" movw %%bx, 160(%%esi)\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm5\n\t" +" movd %%mm1, %%eax\n\t" +" movd %%mm3, %%ecx\n\t" +" movd %%mm5, %%ebx\n\t" +" movw %%ax, 96(%%edi)\n\t" +" movw %%cx, 224(%%edi)\n\t" +" movw %%bx, 352(%%edi)\n\t" + +" movsw\n\t" + +".L_bye:\n\t" +" femms\n\t" + : + :"m"(a),"m"(b),"m"(c),"m"(tmp[0]) + :"memory","%ebx","%esi","%edi"); +} \ No newline at end of file
--- a/mp3lib/dct64_3dnow.s Mon Jan 14 09:32:51 2002 +0000 +++ /dev/null Thu Jan 01 00:00:00 1970 +0000 @@ -1,921 +0,0 @@ -# This code was taken from http://www.mpg123.org -# See ChangeLog of mpg123-0.59s-pre.1 for detail -# Applied to mplayer by Nick Kurshev <nickols_k@mail.ru> -# Partial 3dnow! optimization by Nick Kurshev -# -# TODO: optimize scalar 3dnow! code -# Warning: Phases 7 & 8 are not tested -# - -.data - .align 8 -x_plus_minus_3dnow: .long 0x00000000, 0x80000000 -plus_1f: .float 1.0 - -.text - - .align 16 - -.globl dct64_MMX_3dnow -dct64_MMX_3dnow: - pushl %ebx - pushl %esi - pushl %edi - subl $256,%esp - movl 280(%esp),%eax - leal 128(%esp),%edx - movl 272(%esp),%esi - movl 276(%esp),%edi - movl $costab_mmx,%ebx - orl %ecx,%ecx - movl %esp,%ecx - -/* Phase 1*/ - movq (%eax), %mm0 - movq 8(%eax), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 120(%eax), %mm1 - movq 112(%eax), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, (%edx) - movq %mm4, 8(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul (%ebx), %mm3 - pfmul 8(%ebx), %mm7 - movd %mm3, 124(%edx) - movd %mm7, 116(%edx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 120(%edx) - movd %mm7, 112(%edx) - - movq 16(%eax), %mm0 - movq 24(%eax), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 104(%eax), %mm1 - movq 96(%eax), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 16(%edx) - movq %mm4, 24(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 16(%ebx), %mm3 - pfmul 24(%ebx), %mm7 - movd %mm3, 108(%edx) - movd %mm7, 100(%edx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 104(%edx) - movd %mm7, 96(%edx) - - movq 32(%eax), %mm0 - movq 40(%eax), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 88(%eax), %mm1 - movq 80(%eax), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 32(%edx) - movq %mm4, 40(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 32(%ebx), %mm3 - pfmul 40(%ebx), %mm7 - movd %mm3, 92(%edx) - movd %mm7, 84(%edx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 88(%edx) - movd %mm7, 80(%edx) - - movq 48(%eax), %mm0 - movq 56(%eax), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 72(%eax), %mm1 - movq 64(%eax), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 48(%edx) - movq %mm4, 56(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 48(%ebx), %mm3 - pfmul 56(%ebx), %mm7 - movd %mm3, 76(%edx) - movd %mm7, 68(%edx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 72(%edx) - movd %mm7, 64(%edx) - -/* Phase 2*/ - - movq (%edx), %mm0 - movq 8(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 56(%edx), %mm1 - movq 48(%edx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, (%ecx) - movq %mm4, 8(%ecx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 64(%ebx), %mm3 - pfmul 72(%ebx), %mm7 - movd %mm3, 60(%ecx) - movd %mm7, 52(%ecx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 56(%ecx) - movd %mm7, 48(%ecx) - - movq 16(%edx), %mm0 - movq 24(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 40(%edx), %mm1 - movq 32(%edx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 16(%ecx) - movq %mm4, 24(%ecx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 80(%ebx), %mm3 - pfmul 88(%ebx), %mm7 - movd %mm3, 44(%ecx) - movd %mm7, 36(%ecx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 40(%ecx) - movd %mm7, 32(%ecx) - -/* Phase 3*/ - - movq 64(%edx), %mm0 - movq 72(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 120(%edx), %mm1 - movq 112(%edx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 64(%ecx) - movq %mm4, 72(%ecx) - pfsubr %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 64(%ebx), %mm3 - pfmul 72(%ebx), %mm7 - movd %mm3, 124(%ecx) - movd %mm7, 116(%ecx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 120(%ecx) - movd %mm7, 112(%ecx) - - movq 80(%edx), %mm0 - movq 88(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 104(%edx), %mm1 - movq 96(%edx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 80(%ecx) - movq %mm4, 88(%ecx) - pfsubr %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 80(%ebx), %mm3 - pfmul 88(%ebx), %mm7 - movd %mm3, 108(%ecx) - movd %mm7, 100(%ecx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 104(%ecx) - movd %mm7, 96(%ecx) - -/* Phase 4*/ - - movq (%ecx), %mm0 - movq 8(%ecx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 24(%ecx), %mm1 - movq 16(%ecx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, (%edx) - movq %mm4, 8(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 96(%ebx), %mm3 - pfmul 104(%ebx), %mm7 - movd %mm3, 28(%edx) - movd %mm7, 20(%edx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 24(%edx) - movd %mm7, 16(%edx) - - movq 32(%ecx), %mm0 - movq 40(%ecx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 56(%ecx), %mm1 - movq 48(%ecx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 32(%edx) - movq %mm4, 40(%edx) - pfsubr %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 96(%ebx), %mm3 - pfmul 104(%ebx), %mm7 - movd %mm3, 60(%edx) - movd %mm7, 52(%edx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 56(%edx) - movd %mm7, 48(%edx) - - movq 64(%ecx), %mm0 - movq 72(%ecx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 88(%ecx), %mm1 - movq 80(%ecx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 64(%edx) - movq %mm4, 72(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 96(%ebx), %mm3 - pfmul 104(%ebx), %mm7 - movd %mm3, 92(%edx) - movd %mm7, 84(%edx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 88(%edx) - movd %mm7, 80(%edx) - - movq 96(%ecx), %mm0 - movq 104(%ecx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 120(%ecx), %mm1 - movq 112(%ecx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 96(%edx) - movq %mm4, 104(%edx) - pfsubr %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 96(%ebx), %mm3 - pfmul 104(%ebx), %mm7 - movd %mm3, 124(%edx) - movd %mm7, 116(%edx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 120(%edx) - movd %mm7, 112(%edx) - -/* Phase 5 */ - - movq (%edx), %mm0 - movq 16(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 8(%edx), %mm1 - movq 24(%edx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, (%ecx) - movq %mm4, 16(%ecx) - pfsub %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 112(%ebx), %mm3 - pfmul 112(%ebx), %mm7 - movd %mm3, 12(%ecx) - movd %mm7, 28(%ecx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 8(%ecx) - movd %mm7, 24(%ecx) - - movq 32(%edx), %mm0 - movq 48(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 40(%edx), %mm1 - movq 56(%edx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 32(%ecx) - movq %mm4, 48(%ecx) - pfsub %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 112(%ebx), %mm3 - pfmul 112(%ebx), %mm7 - movd %mm3, 44(%ecx) - movd %mm7, 60(%ecx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 40(%ecx) - movd %mm7, 56(%ecx) - - movq 64(%edx), %mm0 - movq 80(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 72(%edx), %mm1 - movq 88(%edx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 64(%ecx) - movq %mm4, 80(%ecx) - pfsub %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 112(%ebx), %mm3 - pfmul 112(%ebx), %mm7 - movd %mm3, 76(%ecx) - movd %mm7, 92(%ecx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 72(%ecx) - movd %mm7, 88(%ecx) - - movq 96(%edx), %mm0 - movq 112(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 104(%edx), %mm1 - movq 120(%edx), %mm5 - /* n.b.: pswapd*/ - movq %mm1, %mm2 - movq %mm5, %mm6 - psrlq $32, %mm1 - psrlq $32, %mm5 - punpckldq %mm2, %mm1 - punpckldq %mm6, %mm5 - /**/ - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 96(%ecx) - movq %mm4, 112(%ecx) - pfsub %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 112(%ebx), %mm3 - pfmul 112(%ebx), %mm7 - movd %mm3, 108(%ecx) - movd %mm7, 124(%ecx) - psrlq $32, %mm3 - psrlq $32, %mm7 - movd %mm3, 104(%ecx) - movd %mm7, 120(%ecx) - -/* Phase 6. This is the end of easy road. */ -/* Code below is coded in scalar mode. Should be optimized */ - - movd plus_1f, %mm6 - punpckldq 120(%ebx), %mm6 /* mm6 = 1.0 | 120(%ebx)*/ - movq x_plus_minus_3dnow, %mm7 /* mm7 = +1 | -1 */ - - movq 32(%ecx), %mm0 - movq 64(%ecx), %mm2 - movq %mm0, %mm1 - movq %mm2, %mm3 - pxor %mm7, %mm1 - pxor %mm7, %mm3 - pfacc %mm1, %mm0 - pfacc %mm3, %mm2 - pfmul %mm6, %mm0 - pfmul %mm6, %mm2 - movq %mm0, 32(%edx) - movq %mm2, 64(%edx) - - movd 44(%ecx), %mm0 - movd 40(%ecx), %mm2 - movd 120(%ebx), %mm3 - punpckldq 76(%ecx), %mm0 - punpckldq 72(%ecx), %mm2 - punpckldq %mm3, %mm3 - movq %mm0, %mm4 - movq %mm2, %mm5 - pfsub %mm2, %mm0 - pfmul %mm3, %mm0 - movq %mm0, %mm1 - pfadd %mm5, %mm0 - pfadd %mm4, %mm0 - movq %mm0, %mm2 - punpckldq %mm1, %mm0 - punpckhdq %mm1, %mm2 - movq %mm0, 40(%edx) - movq %mm2, 72(%edx) - - movd 48(%ecx), %mm3 - movd 60(%ecx), %mm2 - pfsub 52(%ecx), %mm3 - pfsub 56(%ecx), %mm2 - pfmul 120(%ebx), %mm3 - pfmul 120(%ebx), %mm2 - movq %mm2, %mm1 - - pfadd 56(%ecx), %mm1 - pfadd 60(%ecx), %mm1 - movq %mm1, %mm0 - - pfadd 48(%ecx), %mm0 - pfadd 52(%ecx), %mm0 - pfadd %mm3, %mm1 - punpckldq %mm2, %mm1 - pfadd %mm3, %mm2 - punpckldq %mm2, %mm0 - movq %mm1, 56(%edx) - movq %mm0, 48(%edx) - -/*---*/ - - movd 92(%ecx), %mm1 - pfsub 88(%ecx), %mm1 - pfmul 120(%ebx), %mm1 - movd %mm1, 92(%edx) - pfadd 92(%ecx), %mm1 - pfadd 88(%ecx), %mm1 - movq %mm1, %mm0 - - pfadd 80(%ecx), %mm0 - pfadd 84(%ecx), %mm0 - movd %mm0, 80(%edx) - - movd 80(%ecx), %mm0 - pfsub 84(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - pfadd %mm0, %mm1 - pfadd 92(%edx), %mm0 - punpckldq %mm1, %mm0 - movq %mm0, 84(%edx) - - movq 96(%ecx), %mm0 - movq %mm0, %mm1 - pxor %mm7, %mm1 - pfacc %mm1, %mm0 - pfmul %mm6, %mm0 - movq %mm0, 96(%edx) - - movd 108(%ecx), %mm0 - pfsub 104(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movd %mm0, 108(%edx) - pfadd 104(%ecx), %mm0 - pfadd 108(%ecx), %mm0 - movd %mm0, 104(%edx) - - movd 124(%ecx), %mm1 - pfsub 120(%ecx), %mm1 - pfmul 120(%ebx), %mm1 - movd %mm1, 124(%edx) - pfadd 120(%ecx), %mm1 - pfadd 124(%ecx), %mm1 - movq %mm1, %mm0 - - pfadd 112(%ecx), %mm0 - pfadd 116(%ecx), %mm0 - movd %mm0, 112(%edx) - - movd 112(%ecx), %mm0 - pfsub 116(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - pfadd %mm0,%mm1 - pfadd 124(%edx), %mm0 - punpckldq %mm1, %mm0 - movq %mm0, 116(%edx) - - jnz .L01 - -/* Phase 7*/ -/* Code below is coded in scalar mode. Should be optimized */ - - movd (%ecx), %mm0 - pfadd 4(%ecx), %mm0 - movd %mm0, 1024(%esi) - - movd (%ecx), %mm0 - pfsub 4(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movd %mm0, (%esi) - movd %mm0, (%edi) - - movd 12(%ecx), %mm0 - pfsub 8(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movd %mm0, 512(%edi) - pfadd 12(%ecx), %mm0 - pfadd 8(%ecx), %mm0 - movd %mm0, 512(%esi) - - movd 16(%ecx), %mm0 - pfsub 20(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movq %mm0, %mm3 - - movd 28(%ecx), %mm0 - pfsub 24(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movd %mm0, 768(%edi) - movq %mm0, %mm2 - - pfadd 24(%ecx), %mm0 - pfadd 28(%ecx), %mm0 - movq %mm0, %mm1 - - pfadd 16(%ecx), %mm0 - pfadd 20(%ecx), %mm0 - movd %mm0, 768(%esi) - pfadd %mm3, %mm1 - movd %mm1, 256(%esi) - pfadd %mm3, %mm2 - movd %mm2, 256(%edi) - -/* Phase 8*/ - - movq 32(%edx), %mm0 - movq 48(%edx), %mm1 - pfadd 48(%edx), %mm0 - pfadd 40(%edx), %mm1 - movd %mm0, 896(%esi) - movd %mm1, 640(%esi) - psrlq $32, %mm0 - psrlq $32, %mm1 - movd %mm0, 128(%edi) - movd %mm1, 384(%edi) - - movd 40(%edx), %mm0 - pfadd 56(%edx), %mm0 - movd %mm0, 384(%esi) - - movd 56(%edx), %mm0 - pfadd 36(%edx), %mm0 - movd %mm0, 128(%esi) - - movd 60(%edx), %mm0 - movd %mm0, 896(%edi) - pfadd 44(%edx), %mm0 - movd %mm0, 640(%edi) - - movq 96(%edx), %mm0 - movq 112(%edx), %mm2 - movq 104(%edx), %mm4 - pfadd 112(%edx), %mm0 - pfadd 104(%edx), %mm2 - pfadd 120(%edx), %mm4 - movq %mm0, %mm1 - movq %mm2, %mm3 - movq %mm4, %mm5 - pfadd 64(%edx), %mm0 - pfadd 80(%edx), %mm2 - pfadd 72(%edx), %mm4 - movd %mm0, 960(%esi) - movd %mm2, 704(%esi) - movd %mm4, 448(%esi) - psrlq $32, %mm0 - psrlq $32, %mm2 - psrlq $32, %mm4 - movd %mm0, 64(%edi) - movd %mm2, 320(%edi) - movd %mm4, 576(%edi) - pfadd 80(%edx), %mm1 - pfadd 72(%edx), %mm3 - pfadd 88(%edx), %mm5 - movd %mm1, 832(%esi) - movd %mm3, 576(%esi) - movd %mm5, 320(%esi) - psrlq $32, %mm1 - psrlq $32, %mm3 - psrlq $32, %mm5 - movd %mm1, 192(%edi) - movd %mm3, 448(%edi) - movd %mm5, 704(%edi) - - movd 120(%edx), %mm0 - pfadd 100(%edx), %mm0 - movq %mm0, %mm1 - pfadd 88(%edx), %mm0 - movd %mm0, 192(%esi) - pfadd 68(%edx), %mm1 - movd %mm1, 64(%esi) - - movd 124(%edx), %mm0 - movd %mm0, 960(%edi) - pfadd 92(%edx), %mm0 - movd %mm0, 832(%edi) - - jmp .L_bye -.L01: -/* Phase 9*/ - - movq (%ecx), %mm0 - movq %mm0, %mm1 - pxor %mm7, %mm1 - pfacc %mm1, %mm0 - pfmul %mm6, %mm0 - pf2id %mm0, %mm0 - movd %mm0, %eax - movw %ax, 512(%esi) - psrlq $32, %mm0 - movd %mm0, %eax - movw %ax, (%esi) - - movd 12(%ecx), %mm0 - pfsub 8(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - pf2id %mm0, %mm7 - movd %mm7, %eax - movw %ax, 256(%edi) - pfadd 12(%ecx), %mm0 - pfadd 8(%ecx), %mm0 - pf2id %mm0, %mm0 - movd %mm0, %eax - movw %ax, 256(%esi) - - movd 16(%ecx), %mm3 - pfsub 20(%ecx), %mm3 - pfmul 120(%ebx), %mm3 - movq %mm3, %mm2 - - movd 28(%ecx), %mm2 - pfsub 24(%ecx), %mm2 - pfmul 120(%ebx), %mm2 - movq %mm2, %mm1 - - pf2id %mm2, %mm7 - movd %mm7, %eax - movw %ax, 384(%edi) - - pfadd 24(%ecx), %mm1 - pfadd 28(%ecx), %mm1 - movq %mm1, %mm0 - - pfadd 16(%ecx), %mm0 - pfadd 20(%ecx), %mm0 - pf2id %mm0, %mm0 - movd %mm0, %eax - movw %ax, 384(%esi) - pfadd %mm3, %mm1 - pf2id %mm1, %mm1 - movd %mm1, %eax - movw %ax, 128(%esi) - pfadd %mm3, %mm2 - pf2id %mm2, %mm2 - movd %mm2, %eax - movw %ax, 128(%edi) - -/* Phase 10*/ - - movq 32(%edx), %mm0 - movq 48(%edx), %mm1 - pfadd 48(%edx), %mm0 - pfadd 40(%edx), %mm1 - pf2id %mm0, %mm0 - pf2id %mm1, %mm1 - movd %mm0, %eax - movd %mm1, %ecx - movw %ax, 448(%esi) - movw %cx, 320(%esi) - psrlq $32, %mm0 - psrlq $32, %mm1 - movd %mm0, %eax - movd %mm1, %ecx - movw %ax, 64(%edi) - movw %cx, 192(%edi) - - movd 40(%edx), %mm3 - movd 56(%edx), %mm4 - movd 60(%edx), %mm0 - movd 44(%edx), %mm2 - movd 120(%edx), %mm5 - punpckldq %mm4, %mm3 - punpckldq 124(%edx), %mm0 - pfadd 100(%edx), %mm5 - punpckldq 36(%edx), %mm4 - punpckldq 92(%edx), %mm2 - movq %mm5, %mm6 - pfadd %mm4, %mm3 - pf2id %mm0, %mm1 - pf2id %mm3, %mm3 - pfadd 88(%edx), %mm5 - movd %mm1, %eax - movd %mm3, %ecx - movw %ax, 448(%edi) - movw %cx, 192(%esi) - pf2id %mm5, %mm5 - psrlq $32, %mm1 - psrlq $32, %mm3 - movd %mm5, %ebx - movd %mm1, %eax - movd %mm3, %ecx - movw %bx, 96(%esi) - movw %ax, 480(%edi) - movw %cx, 64(%esi) - pfadd %mm2, %mm0 - pf2id %mm0, %mm0 - movd %mm0, %eax - pfadd 68(%edx), %mm6 - movw %ax, 320(%edi) - psrlq $32, %mm0 - pf2id %mm6, %mm6 - movd %mm0, %eax - movd %mm6, %ebx - movw %ax, 416(%edi) - movw %bx, 32(%esi) - - movq 96(%edx), %mm0 - movq 112(%edx), %mm2 - movq 104(%edx), %mm4 - pfadd %mm2, %mm0 - pfadd %mm4, %mm2 - pfadd 120(%edx), %mm4 - movq %mm0, %mm1 - movq %mm2, %mm3 - movq %mm4, %mm5 - pfadd 64(%edx), %mm0 - pfadd 80(%edx), %mm2 - pfadd 72(%edx), %mm4 - pf2id %mm0, %mm0 - pf2id %mm2, %mm2 - pf2id %mm4, %mm4 - movd %mm0, %eax - movd %mm2, %ecx - movd %mm4, %ebx - movw %ax, 480(%esi) - movw %cx, 352(%esi) - movw %bx, 224(%esi) - psrlq $32, %mm0 - psrlq $32, %mm2 - psrlq $32, %mm4 - movd %mm0, %eax - movd %mm2, %ecx - movd %mm4, %ebx - movw %ax, 32(%edi) - movw %cx, 160(%edi) - movw %bx, 288(%edi) - pfadd 80(%edx), %mm1 - pfadd 72(%edx), %mm3 - pfadd 88(%edx), %mm5 - pf2id %mm1, %mm1 - pf2id %mm3, %mm3 - pf2id %mm5, %mm5 - movd %mm1, %eax - movd %mm3, %ecx - movd %mm5, %ebx - movw %ax, 416(%esi) - movw %cx, 288(%esi) - movw %bx, 160(%esi) - psrlq $32, %mm1 - psrlq $32, %mm3 - psrlq $32, %mm5 - movd %mm1, %eax - movd %mm3, %ecx - movd %mm5, %ebx - movw %ax, 96(%edi) - movw %cx, 224(%edi) - movw %bx, 352(%edi) - - movsw - -.L_bye: - addl $256,%esp - femms - popl %edi - popl %esi - popl %ebx - ret $12
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/mp3lib/dct64_k7.c Mon Jan 14 10:34:38 2002 +0000 @@ -0,0 +1,792 @@ +/* +* This code was taken from http://www.mpg123.org +* See ChangeLog of mpg123-0.59s-pre.1 for detail +* Applied to mplayer by Nick Kurshev <nickols_k@mail.ru> +* Partial 3dnowex-DSP! optimization by Nick Kurshev +* +* TODO: optimize scalar 3dnow! code +* Warning: Phases 7 & 8 are not tested +*/ +#define real float /* ugly - but only way */ + +static unsigned long long int __attribute__((aligned(8))) x_plus_minus_3dnow = 0x8000000000000000ULL; +static float plus_1f = 1.0; + +void __attribute__ (( __stdcall__ )) dct64_MMX_3dnowex(real *a,real *b,real *c) +{ + char tmp[256]; + __asm __volatile( +" movl %2,%%eax\n\t" + +" leal 128+%3,%%edx\n\t" +" movl %0,%%esi\n\t" +" movl %1,%%edi\n\t" +" movl $costab_mmx,%%ebx\n\t" +" leal %3,%%ecx\n\t" + +/* Phase 1*/ +" movq (%%eax), %%mm0\n\t" +" movq 8(%%eax), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 120(%%eax), %%mm1\n\t" +" movq 112(%%eax), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, (%%edx)\n\t" +" movq %%mm4, 8(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul (%%ebx), %%mm3\n\t" +" pfmul 8(%%ebx), %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 120(%%edx)\n\t" +" movq %%mm7, 112(%%edx)\n\t" + +" movq 16(%%eax), %%mm0\n\t" +" movq 24(%%eax), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 104(%%eax), %%mm1\n\t" +" movq 96(%%eax), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 16(%%edx)\n\t" +" movq %%mm4, 24(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 16(%%ebx), %%mm3\n\t" +" pfmul 24(%%ebx), %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 104(%%edx)\n\t" +" movq %%mm7, 96(%%edx)\n\t" + +" movq 32(%%eax), %%mm0\n\t" +" movq 40(%%eax), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 88(%%eax), %%mm1\n\t" +" movq 80(%%eax), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 32(%%edx)\n\t" +" movq %%mm4, 40(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 32(%%ebx), %%mm3\n\t" +" pfmul 40(%%ebx), %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 88(%%edx)\n\t" +" movq %%mm7, 80(%%edx)\n\t" + +" movq 48(%%eax), %%mm0\n\t" +" movq 56(%%eax), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 72(%%eax), %%mm1\n\t" +" movq 64(%%eax), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 48(%%edx)\n\t" +" movq %%mm4, 56(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 48(%%ebx), %%mm3\n\t" +" pfmul 56(%%ebx), %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 72(%%edx)\n\t" +" movq %%mm7, 64(%%edx)\n\t" + +/* Phase 2*/ + +" movq (%%edx), %%mm0\n\t" +" movq 8(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 56(%%edx), %%mm1\n\t" +" movq 48(%%edx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, (%%ecx)\n\t" +" movq %%mm4, 8(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 64(%%ebx), %%mm3\n\t" +" pfmul 72(%%ebx), %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 56(%%ecx)\n\t" +" movq %%mm7, 48(%%ecx)\n\t" + +" movq 16(%%edx), %%mm0\n\t" +" movq 24(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 40(%%edx), %%mm1\n\t" +" movq 32(%%edx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 16(%%ecx)\n\t" +" movq %%mm4, 24(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul 80(%%ebx), %%mm3\n\t" +" pfmul 88(%%ebx), %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 40(%%ecx)\n\t" +" movq %%mm7, 32(%%ecx)\n\t" + +/* Phase 3*/ + +" movq 64(%%edx), %%mm0\n\t" +" movq 72(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 120(%%edx), %%mm1\n\t" +" movq 112(%%edx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 64(%%ecx)\n\t" +" movq %%mm4, 72(%%ecx)\n\t" +" pfsubr %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 64(%%ebx), %%mm3\n\t" +" pfmul 72(%%ebx), %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 120(%%ecx)\n\t" +" movq %%mm7, 112(%%ecx)\n\t" + +" movq 80(%%edx), %%mm0\n\t" +" movq 88(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 104(%%edx), %%mm1\n\t" +" movq 96(%%edx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 80(%%ecx)\n\t" +" movq %%mm4, 88(%%ecx)\n\t" +" pfsubr %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul 80(%%ebx), %%mm3\n\t" +" pfmul 88(%%ebx), %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 104(%%ecx)\n\t" +" movq %%mm7, 96(%%ecx)\n\t" + +/* Phase 4*/ + +" movq 96(%%ebx), %%mm2\n\t" +" movq 104(%%ebx), %%mm6\n\t" + +" movq (%%ecx), %%mm0\n\t" +" movq 8(%%ecx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 24(%%ecx), %%mm1\n\t" +" movq 16(%%ecx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, (%%edx)\n\t" +" movq %%mm4, 8(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul %%mm2, %%mm3\n\t" +" pfmul %%mm6, %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 24(%%edx)\n\t" +" movq %%mm7, 16(%%edx)\n\t" + +" movq 32(%%ecx), %%mm0\n\t" +" movq 40(%%ecx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 56(%%ecx), %%mm1\n\t" +" movq 48(%%ecx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 32(%%edx)\n\t" +" movq %%mm4, 40(%%edx)\n\t" +" pfsubr %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul %%mm2, %%mm3\n\t" +" pfmul %%mm6, %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 56(%%edx)\n\t" +" movq %%mm7, 48(%%edx)\n\t" + +" movq 64(%%ecx), %%mm0\n\t" +" movq 72(%%ecx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 88(%%ecx), %%mm1\n\t" +" movq 80(%%ecx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 64(%%edx)\n\t" +" movq %%mm4, 72(%%edx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsub %%mm5, %%mm7\n\t" +" pfmul %%mm2, %%mm3\n\t" +" pfmul %%mm6, %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 88(%%edx)\n\t" +" movq %%mm7, 80(%%edx)\n\t" + +" movq 96(%%ecx), %%mm0\n\t" +" movq 104(%%ecx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 120(%%ecx), %%mm1\n\t" +" movq 112(%%ecx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 96(%%edx)\n\t" +" movq %%mm4, 104(%%edx)\n\t" +" pfsubr %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul %%mm2, %%mm3\n\t" +" pfmul %%mm6, %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 120(%%edx)\n\t" +" movq %%mm7, 112(%%edx)\n\t" + +/* Phase 5 */ + +" movq 112(%%ebx), %%mm2\n\t" + +" movq (%%edx), %%mm0\n\t" +" movq 16(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 8(%%edx), %%mm1\n\t" +" movq 24(%%edx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, (%%ecx)\n\t" +" movq %%mm4, 16(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul %%mm2, %%mm3\n\t" +" pfmul %%mm2, %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 8(%%ecx)\n\t" +" movq %%mm7, 24(%%ecx)\n\t" + +" movq 32(%%edx), %%mm0\n\t" +" movq 48(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 40(%%edx), %%mm1\n\t" +" movq 56(%%edx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 32(%%ecx)\n\t" +" movq %%mm4, 48(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul %%mm2, %%mm3\n\t" +" pfmul %%mm2, %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 40(%%ecx)\n\t" +" movq %%mm7, 56(%%ecx)\n\t" + +" movq 64(%%edx), %%mm0\n\t" +" movq 80(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 72(%%edx), %%mm1\n\t" +" movq 88(%%edx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 64(%%ecx)\n\t" +" movq %%mm4, 80(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul %%mm2, %%mm3\n\t" +" pfmul %%mm2, %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 72(%%ecx)\n\t" +" movq %%mm7, 88(%%ecx)\n\t" + +" movq 96(%%edx), %%mm0\n\t" +" movq 112(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm3\n\t" +" movq %%mm4, %%mm7\n\t" +" movq 104(%%edx), %%mm1\n\t" +" movq 120(%%edx), %%mm5\n\t" +" pswapd %%mm1, %%mm1\n\t" +" pswapd %%mm5, %%mm5\n\t" +" pfadd %%mm1, %%mm0\n\t" +" pfadd %%mm5, %%mm4\n\t" +" movq %%mm0, 96(%%ecx)\n\t" +" movq %%mm4, 112(%%ecx)\n\t" +" pfsub %%mm1, %%mm3\n\t" +" pfsubr %%mm5, %%mm7\n\t" +" pfmul %%mm2, %%mm3\n\t" +" pfmul %%mm2, %%mm7\n\t" +" pswapd %%mm3, %%mm3\n\t" +" pswapd %%mm7, %%mm7\n\t" +" movq %%mm3, 104(%%ecx)\n\t" +" movq %%mm7, 120(%%ecx)\n\t" + + +/* Phase 6. This is the end of easy road. */ +/* Code below is coded in scalar mode. Should be optimized */ + +" movd plus_1f, %%mm6\n\t" +" punpckldq 120(%%ebx), %%mm6\n\t" /* mm6 = 1.0 | 120(%%ebx)*/ +" movq x_plus_minus_3dnow, %%mm7\n\t" /* mm7 = +1 | -1 */ + +" movq 32(%%ecx), %%mm0\n\t" +" movq 64(%%ecx), %%mm2\n\t" +" movq %%mm0, %%mm1\n\t" +" movq %%mm2, %%mm3\n\t" +" pxor %%mm7, %%mm1\n\t" +" pxor %%mm7, %%mm3\n\t" +" pfacc %%mm1, %%mm0\n\t" +" pfacc %%mm3, %%mm2\n\t" +" pfmul %%mm6, %%mm0\n\t" +" pfmul %%mm6, %%mm2\n\t" +" movq %%mm0, 32(%%edx)\n\t" +" movq %%mm2, 64(%%edx)\n\t" + +" movd 44(%%ecx), %%mm0\n\t" +" movd 40(%%ecx), %%mm2\n\t" +" movd 120(%%ebx), %%mm3\n\t" +" punpckldq 76(%%ecx), %%mm0\n\t" +" punpckldq 72(%%ecx), %%mm2\n\t" +" punpckldq %%mm3, %%mm3\n\t" +" movq %%mm0, %%mm4\n\t" +" movq %%mm2, %%mm5\n\t" +" pfsub %%mm2, %%mm0\n\t" +" pfmul %%mm3, %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" +" pfadd %%mm5, %%mm0\n\t" +" pfadd %%mm4, %%mm0\n\t" +" movq %%mm0, %%mm2\n\t" +" punpckldq %%mm1, %%mm0\n\t" +" punpckhdq %%mm1, %%mm2\n\t" +" movq %%mm0, 40(%%edx)\n\t" +" movq %%mm2, 72(%%edx)\n\t" + +" movd 48(%%ecx), %%mm3\n\t" +" movd 60(%%ecx), %%mm2\n\t" +" pfsub 52(%%ecx), %%mm3\n\t" +" pfsub 56(%%ecx), %%mm2\n\t" +" pfmul 120(%%ebx), %%mm3\n\t" +" pfmul 120(%%ebx), %%mm2\n\t" +" movq %%mm2, %%mm1\n\t" + +" pfadd 56(%%ecx), %%mm1\n\t" +" pfadd 60(%%ecx), %%mm1\n\t" +" movq %%mm1, %%mm0\n\t" + +" pfadd 48(%%ecx), %%mm0\n\t" +" pfadd 52(%%ecx), %%mm0\n\t" +" pfadd %%mm3, %%mm1\n\t" +" punpckldq %%mm2, %%mm1\n\t" +" pfadd %%mm3, %%mm2\n\t" +" punpckldq %%mm2, %%mm0\n\t" +" movq %%mm1, 56(%%edx)\n\t" +" movq %%mm0, 48(%%edx)\n\t" + +/*---*/ + +" movd 92(%%ecx), %%mm1\n\t" +" pfsub 88(%%ecx), %%mm1\n\t" +" pfmul 120(%%ebx), %%mm1\n\t" +" movd %%mm1, 92(%%edx)\n\t" +" pfadd 92(%%ecx), %%mm1\n\t" +" pfadd 88(%%ecx), %%mm1\n\t" +" movq %%mm1, %%mm0\n\t" + +" pfadd 80(%%ecx), %%mm0\n\t" +" pfadd 84(%%ecx), %%mm0\n\t" +" movd %%mm0, 80(%%edx)\n\t" + +" movd 80(%%ecx), %%mm0\n\t" +" pfsub 84(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" pfadd %%mm0, %%mm1\n\t" +" pfadd 92(%%edx), %%mm0\n\t" +" punpckldq %%mm1, %%mm0\n\t" +" movq %%mm0, 84(%%edx)\n\t" + +" movq 96(%%ecx), %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" +" pxor %%mm7, %%mm1\n\t" +" pfacc %%mm1, %%mm0\n\t" +" pfmul %%mm6, %%mm0\n\t" +" movq %%mm0, 96(%%edx)\n\t" + +" movd 108(%%ecx), %%mm0\n\t" +" pfsub 104(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movd %%mm0, 108(%%edx)\n\t" +" pfadd 104(%%ecx), %%mm0\n\t" +" pfadd 108(%%ecx), %%mm0\n\t" +" movd %%mm0, 104(%%edx)\n\t" + +" movd 124(%%ecx), %%mm1\n\t" +" pfsub 120(%%ecx), %%mm1\n\t" +" pfmul 120(%%ebx), %%mm1\n\t" +" movd %%mm1, 124(%%edx)\n\t" +" pfadd 120(%%ecx), %%mm1\n\t" +" pfadd 124(%%ecx), %%mm1\n\t" +" movq %%mm1, %%mm0\n\t" + +" pfadd 112(%%ecx), %%mm0\n\t" +" pfadd 116(%%ecx), %%mm0\n\t" +" movd %%mm0, 112(%%edx)\n\t" + +" movd 112(%%ecx), %%mm0\n\t" +" pfsub 116(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" pfadd %%mm0,%%mm1\n\t" +" pfadd 124(%%edx), %%mm0\n\t" +" punpckldq %%mm1, %%mm0\n\t" +" movq %%mm0, 116(%%edx)\n\t" + +" jnz .L01\n\t" + +/* Phase 7*/ +/* Code below is coded in scalar mode. Should be optimized */ + +" movd (%%ecx), %%mm0\n\t" +" pfadd 4(%%ecx), %%mm0\n\t" +" movd %%mm0, 1024(%%esi)\n\t" + +" movd (%%ecx), %%mm0\n\t" +" pfsub 4(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movd %%mm0, (%%esi)\n\t" +" movd %%mm0, (%%edi)\n\t" + +" movd 12(%%ecx), %%mm0\n\t" +" pfsub 8(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movd %%mm0, 512(%%edi)\n\t" +" pfadd 12(%%ecx), %%mm0\n\t" +" pfadd 8(%%ecx), %%mm0\n\t" +" movd %%mm0, 512(%%esi)\n\t" + +" movd 16(%%ecx), %%mm0\n\t" +" pfsub 20(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movq %%mm0, %%mm3\n\t" + +" movd 28(%%ecx), %%mm0\n\t" +" pfsub 24(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" movd %%mm0, 768(%%edi)\n\t" +" movq %%mm0, %%mm2\n\t" + +" pfadd 24(%%ecx), %%mm0\n\t" +" pfadd 28(%%ecx), %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" + +" pfadd 16(%%ecx), %%mm0\n\t" +" pfadd 20(%%ecx), %%mm0\n\t" +" movd %%mm0, 768(%%esi)\n\t" +" pfadd %%mm3, %%mm1\n\t" +" movd %%mm1, 256(%%esi)\n\t" +" pfadd %%mm3, %%mm2\n\t" +" movd %%mm2, 256(%%edi)\n\t" + +/* Phase 8*/ + +" movq 32(%%edx), %%mm0\n\t" +" movq 48(%%edx), %%mm1\n\t" +" pfadd 48(%%edx), %%mm0\n\t" +" pfadd 40(%%edx), %%mm1\n\t" +" movd %%mm0, 896(%%esi)\n\t" +" movd %%mm1, 640(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" psrlq $32, %%mm1\n\t" +" movd %%mm0, 128(%%edi)\n\t" +" movd %%mm1, 384(%%edi)\n\t" + +" movd 40(%%edx), %%mm0\n\t" +" pfadd 56(%%edx), %%mm0\n\t" +" movd %%mm0, 384(%%esi)\n\t" + +" movd 56(%%edx), %%mm0\n\t" +" pfadd 36(%%edx), %%mm0\n\t" +" movd %%mm0, 128(%%esi)\n\t" + +" movd 60(%%edx), %%mm0\n\t" +" movd %%mm0, 896(%%edi)\n\t" +" pfadd 44(%%edx), %%mm0\n\t" +" movd %%mm0, 640(%%edi)\n\t" + +" movq 96(%%edx), %%mm0\n\t" +" movq 112(%%edx), %%mm2\n\t" +" movq 104(%%edx), %%mm4\n\t" +" pfadd 112(%%edx), %%mm0\n\t" +" pfadd 104(%%edx), %%mm2\n\t" +" pfadd 120(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm1\n\t" +" movq %%mm2, %%mm3\n\t" +" movq %%mm4, %%mm5\n\t" +" pfadd 64(%%edx), %%mm0\n\t" +" pfadd 80(%%edx), %%mm2\n\t" +" pfadd 72(%%edx), %%mm4\n\t" +" movd %%mm0, 960(%%esi)\n\t" +" movd %%mm2, 704(%%esi)\n\t" +" movd %%mm4, 448(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" psrlq $32, %%mm2\n\t" +" psrlq $32, %%mm4\n\t" +" movd %%mm0, 64(%%edi)\n\t" +" movd %%mm2, 320(%%edi)\n\t" +" movd %%mm4, 576(%%edi)\n\t" +" pfadd 80(%%edx), %%mm1\n\t" +" pfadd 72(%%edx), %%mm3\n\t" +" pfadd 88(%%edx), %%mm5\n\t" +" movd %%mm1, 832(%%esi)\n\t" +" movd %%mm3, 576(%%esi)\n\t" +" movd %%mm5, 320(%%esi)\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm5\n\t" +" movd %%mm1, 192(%%edi)\n\t" +" movd %%mm3, 448(%%edi)\n\t" +" movd %%mm5, 704(%%edi)\n\t" + +" movd 120(%%edx), %%mm0\n\t" +" pfadd 100(%%edx), %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" +" pfadd 88(%%edx), %%mm0\n\t" +" movd %%mm0, 192(%%esi)\n\t" +" pfadd 68(%%edx), %%mm1\n\t" +" movd %%mm1, 64(%%esi)\n\t" + +" movd 124(%%edx), %%mm0\n\t" +" movd %%mm0, 960(%%edi)\n\t" +" pfadd 92(%%edx), %%mm0\n\t" +" movd %%mm0, 832(%%edi)\n\t" + +" jmp .L_bye\n\t" +".L01: \n\t" +/* Phase 9*/ + +" movq (%%ecx), %%mm0\n\t" +" movq %%mm0, %%mm1\n\t" +" pxor %%mm7, %%mm1\n\t" +" pfacc %%mm1, %%mm0\n\t" +" pfmul %%mm6, %%mm0\n\t" +" pf2id %%mm0, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" movw %%ax, 512(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" movw %%ax, (%%esi)\n\t" + +" movd 12(%%ecx), %%mm0\n\t" +" pfsub 8(%%ecx), %%mm0\n\t" +" pfmul 120(%%ebx), %%mm0\n\t" +" pf2id %%mm0, %%mm7\n\t" +" movd %%mm7, %%eax\n\t" +" movw %%ax, 256(%%edi)\n\t" +" pfadd 12(%%ecx), %%mm0\n\t" +" pfadd 8(%%ecx), %%mm0\n\t" +" pf2id %%mm0, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" movw %%ax, 256(%%esi)\n\t" + +" movd 16(%%ecx), %%mm3\n\t" +" pfsub 20(%%ecx), %%mm3\n\t" +" pfmul 120(%%ebx), %%mm3\n\t" +" movq %%mm3, %%mm2\n\t" + +" movd 28(%%ecx), %%mm2\n\t" +" pfsub 24(%%ecx), %%mm2\n\t" +" pfmul 120(%%ebx), %%mm2\n\t" +" movq %%mm2, %%mm1\n\t" + +" pf2id %%mm2, %%mm7\n\t" +" movd %%mm7, %%eax\n\t" +" movw %%ax, 384(%%edi)\n\t" + +" pfadd 24(%%ecx), %%mm1\n\t" +" pfadd 28(%%ecx), %%mm1\n\t" +" movq %%mm1, %%mm0\n\t" + +" pfadd 16(%%ecx), %%mm0\n\t" +" pfadd 20(%%ecx), %%mm0\n\t" +" pf2id %%mm0, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" movw %%ax, 384(%%esi)\n\t" +" pfadd %%mm3, %%mm1\n\t" +" pf2id %%mm1, %%mm1\n\t" +" movd %%mm1, %%eax\n\t" +" movw %%ax, 128(%%esi)\n\t" +" pfadd %%mm3, %%mm2\n\t" +" pf2id %%mm2, %%mm2\n\t" +" movd %%mm2, %%eax\n\t" +" movw %%ax, 128(%%edi)\n\t" + +/* Phase 10*/ + +" movq 32(%%edx), %%mm0\n\t" +" movq 48(%%edx), %%mm1\n\t" +" pfadd 48(%%edx), %%mm0\n\t" +" pfadd 40(%%edx), %%mm1\n\t" +" pf2id %%mm0, %%mm0\n\t" +" pf2id %%mm1, %%mm1\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm1, %%ecx\n\t" +" movw %%ax, 448(%%esi)\n\t" +" movw %%cx, 320(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" psrlq $32, %%mm1\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm1, %%ecx\n\t" +" movw %%ax, 64(%%edi)\n\t" +" movw %%cx, 192(%%edi)\n\t" + +" movd 40(%%edx), %%mm3\n\t" +" movd 56(%%edx), %%mm4\n\t" +" movd 60(%%edx), %%mm0\n\t" +" movd 44(%%edx), %%mm2\n\t" +" movd 120(%%edx), %%mm5\n\t" +" punpckldq %%mm4, %%mm3\n\t" +" punpckldq 124(%%edx), %%mm0\n\t" +" pfadd 100(%%edx), %%mm5\n\t" +" punpckldq 36(%%edx), %%mm4\n\t" +" punpckldq 92(%%edx), %%mm2\n\t" +" movq %%mm5, %%mm6\n\t" +" pfadd %%mm4, %%mm3\n\t" +" pf2id %%mm0, %%mm1\n\t" +" pf2id %%mm3, %%mm3\n\t" +" pfadd 88(%%edx), %%mm5\n\t" +" movd %%mm1, %%eax\n\t" +" movd %%mm3, %%ecx\n\t" +" movw %%ax, 448(%%edi)\n\t" +" movw %%cx, 192(%%esi)\n\t" +" pf2id %%mm5, %%mm5\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm3\n\t" +" movd %%mm5, %%ebx\n\t" +" movd %%mm1, %%eax\n\t" +" movd %%mm3, %%ecx\n\t" +" movw %%bx, 96(%%esi)\n\t" +" movw %%ax, 480(%%edi)\n\t" +" movw %%cx, 64(%%esi)\n\t" +" pfadd %%mm2, %%mm0\n\t" +" pf2id %%mm0, %%mm0\n\t" +" movd %%mm0, %%eax\n\t" +" pfadd 68(%%edx), %%mm6\n\t" +" movw %%ax, 320(%%edi)\n\t" +" psrlq $32, %%mm0\n\t" +" pf2id %%mm6, %%mm6\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm6, %%ebx\n\t" +" movw %%ax, 416(%%edi)\n\t" +" movw %%bx, 32(%%esi)\n\t" + +" movq 96(%%edx), %%mm0\n\t" +" movq 112(%%edx), %%mm2\n\t" +" movq 104(%%edx), %%mm4\n\t" +" pfadd %%mm2, %%mm0\n\t" +" pfadd %%mm4, %%mm2\n\t" +" pfadd 120(%%edx), %%mm4\n\t" +" movq %%mm0, %%mm1\n\t" +" movq %%mm2, %%mm3\n\t" +" movq %%mm4, %%mm5\n\t" +" pfadd 64(%%edx), %%mm0\n\t" +" pfadd 80(%%edx), %%mm2\n\t" +" pfadd 72(%%edx), %%mm4\n\t" +" pf2id %%mm0, %%mm0\n\t" +" pf2id %%mm2, %%mm2\n\t" +" pf2id %%mm4, %%mm4\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm2, %%ecx\n\t" +" movd %%mm4, %%ebx\n\t" +" movw %%ax, 480(%%esi)\n\t" +" movw %%cx, 352(%%esi)\n\t" +" movw %%bx, 224(%%esi)\n\t" +" psrlq $32, %%mm0\n\t" +" psrlq $32, %%mm2\n\t" +" psrlq $32, %%mm4\n\t" +" movd %%mm0, %%eax\n\t" +" movd %%mm2, %%ecx\n\t" +" movd %%mm4, %%ebx\n\t" +" movw %%ax, 32(%%edi)\n\t" +" movw %%cx, 160(%%edi)\n\t" +" movw %%bx, 288(%%edi)\n\t" +" pfadd 80(%%edx), %%mm1\n\t" +" pfadd 72(%%edx), %%mm3\n\t" +" pfadd 88(%%edx), %%mm5\n\t" +" pf2id %%mm1, %%mm1\n\t" +" pf2id %%mm3, %%mm3\n\t" +" pf2id %%mm5, %%mm5\n\t" +" movd %%mm1, %%eax\n\t" +" movd %%mm3, %%ecx\n\t" +" movd %%mm5, %%ebx\n\t" +" movw %%ax, 416(%%esi)\n\t" +" movw %%cx, 288(%%esi)\n\t" +" movw %%bx, 160(%%esi)\n\t" +" psrlq $32, %%mm1\n\t" +" psrlq $32, %%mm3\n\t" +" psrlq $32, %%mm5\n\t" +" movd %%mm1, %%eax\n\t" +" movd %%mm3, %%ecx\n\t" +" movd %%mm5, %%ebx\n\t" +" movw %%ax, 96(%%edi)\n\t" +" movw %%cx, 224(%%edi)\n\t" +" movw %%bx, 352(%%edi)\n\t" + +" movsw\n\t" + +".L_bye:\n\t" +" femms\n\t" + : + :"m"(a),"m"(b),"m"(c),"m"(tmp[0]) + :"memory","%ebx","%esi","%edi"); +} \ No newline at end of file
--- a/mp3lib/dct64_k7.s Mon Jan 14 09:32:51 2002 +0000 +++ /dev/null Thu Jan 01 00:00:00 1970 +0000 @@ -1,800 +0,0 @@ -# This code was taken from http://www.mpg123.org -# See ChangeLog of mpg123-0.59s-pre.1 for detail -# Applied to mplayer by Nick Kurshev <nickols_k@mail.ru> -# Partial 3dnowex-DSP! optimization by Nick Kurshev -# -# TODO: optimize scalar 3dnow! code -# Warning: Phases 7 & 8 are not tested -# - -.data - .align 8 -x_plus_minus_3dnow: .long 0x00000000, 0x80000000 -plus_1f: .float 1.0 - -.text - - .align 16 - -.globl dct64_MMX_3dnowex -dct64_MMX_3dnowex: - pushl %ebx - pushl %esi - pushl %edi - subl $256,%esp - movl 280(%esp),%eax - - leal 128(%esp),%edx - movl 272(%esp),%esi - movl 276(%esp),%edi - movl $costab_mmx,%ebx - orl %ecx,%ecx - movl %esp,%ecx - -/* Phase 1*/ - movq (%eax), %mm0 - movq 8(%eax), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 120(%eax), %mm1 - movq 112(%eax), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, (%edx) - movq %mm4, 8(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul (%ebx), %mm3 - pfmul 8(%ebx), %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 120(%edx) - movq %mm7, 112(%edx) - - movq 16(%eax), %mm0 - movq 24(%eax), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 104(%eax), %mm1 - movq 96(%eax), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 16(%edx) - movq %mm4, 24(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 16(%ebx), %mm3 - pfmul 24(%ebx), %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 104(%edx) - movq %mm7, 96(%edx) - - movq 32(%eax), %mm0 - movq 40(%eax), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 88(%eax), %mm1 - movq 80(%eax), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 32(%edx) - movq %mm4, 40(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 32(%ebx), %mm3 - pfmul 40(%ebx), %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 88(%edx) - movq %mm7, 80(%edx) - - movq 48(%eax), %mm0 - movq 56(%eax), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 72(%eax), %mm1 - movq 64(%eax), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 48(%edx) - movq %mm4, 56(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 48(%ebx), %mm3 - pfmul 56(%ebx), %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 72(%edx) - movq %mm7, 64(%edx) - -/* Phase 2*/ - - movq (%edx), %mm0 - movq 8(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 56(%edx), %mm1 - movq 48(%edx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, (%ecx) - movq %mm4, 8(%ecx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 64(%ebx), %mm3 - pfmul 72(%ebx), %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 56(%ecx) - movq %mm7, 48(%ecx) - - movq 16(%edx), %mm0 - movq 24(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 40(%edx), %mm1 - movq 32(%edx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 16(%ecx) - movq %mm4, 24(%ecx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul 80(%ebx), %mm3 - pfmul 88(%ebx), %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 40(%ecx) - movq %mm7, 32(%ecx) - -/* Phase 3*/ - - movq 64(%edx), %mm0 - movq 72(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 120(%edx), %mm1 - movq 112(%edx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 64(%ecx) - movq %mm4, 72(%ecx) - pfsubr %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 64(%ebx), %mm3 - pfmul 72(%ebx), %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 120(%ecx) - movq %mm7, 112(%ecx) - - movq 80(%edx), %mm0 - movq 88(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 104(%edx), %mm1 - movq 96(%edx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 80(%ecx) - movq %mm4, 88(%ecx) - pfsubr %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul 80(%ebx), %mm3 - pfmul 88(%ebx), %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 104(%ecx) - movq %mm7, 96(%ecx) - -/* Phase 4*/ - - movq 96(%ebx), %mm2 - movq 104(%ebx), %mm6 - - movq (%ecx), %mm0 - movq 8(%ecx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 24(%ecx), %mm1 - movq 16(%ecx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, (%edx) - movq %mm4, 8(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul %mm2, %mm3 - pfmul %mm6, %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 24(%edx) - movq %mm7, 16(%edx) - - movq 32(%ecx), %mm0 - movq 40(%ecx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 56(%ecx), %mm1 - movq 48(%ecx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 32(%edx) - movq %mm4, 40(%edx) - pfsubr %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul %mm2, %mm3 - pfmul %mm6, %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 56(%edx) - movq %mm7, 48(%edx) - - movq 64(%ecx), %mm0 - movq 72(%ecx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 88(%ecx), %mm1 - movq 80(%ecx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 64(%edx) - movq %mm4, 72(%edx) - pfsub %mm1, %mm3 - pfsub %mm5, %mm7 - pfmul %mm2, %mm3 - pfmul %mm6, %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 88(%edx) - movq %mm7, 80(%edx) - - movq 96(%ecx), %mm0 - movq 104(%ecx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 120(%ecx), %mm1 - movq 112(%ecx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 96(%edx) - movq %mm4, 104(%edx) - pfsubr %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul %mm2, %mm3 - pfmul %mm6, %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 120(%edx) - movq %mm7, 112(%edx) - -/* Phase 5 */ - - movq 112(%ebx), %mm2 - - movq (%edx), %mm0 - movq 16(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 8(%edx), %mm1 - movq 24(%edx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, (%ecx) - movq %mm4, 16(%ecx) - pfsub %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul %mm2, %mm3 - pfmul %mm2, %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 8(%ecx) - movq %mm7, 24(%ecx) - - movq 32(%edx), %mm0 - movq 48(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 40(%edx), %mm1 - movq 56(%edx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 32(%ecx) - movq %mm4, 48(%ecx) - pfsub %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul %mm2, %mm3 - pfmul %mm2, %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 40(%ecx) - movq %mm7, 56(%ecx) - - movq 64(%edx), %mm0 - movq 80(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 72(%edx), %mm1 - movq 88(%edx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 64(%ecx) - movq %mm4, 80(%ecx) - pfsub %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul %mm2, %mm3 - pfmul %mm2, %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 72(%ecx) - movq %mm7, 88(%ecx) - - movq 96(%edx), %mm0 - movq 112(%edx), %mm4 - movq %mm0, %mm3 - movq %mm4, %mm7 - movq 104(%edx), %mm1 - movq 120(%edx), %mm5 - pswapd %mm1, %mm1 - pswapd %mm5, %mm5 - pfadd %mm1, %mm0 - pfadd %mm5, %mm4 - movq %mm0, 96(%ecx) - movq %mm4, 112(%ecx) - pfsub %mm1, %mm3 - pfsubr %mm5, %mm7 - pfmul %mm2, %mm3 - pfmul %mm2, %mm7 - pswapd %mm3, %mm3 - pswapd %mm7, %mm7 - movq %mm3, 104(%ecx) - movq %mm7, 120(%ecx) - - -/* Phase 6. This is the end of easy road. */ -/* Code below is coded in scalar mode. Should be optimized */ - - movd plus_1f, %mm6 - punpckldq 120(%ebx), %mm6 /* mm6 = 1.0 | 120(%ebx)*/ - movq x_plus_minus_3dnow, %mm7 /* mm7 = +1 | -1 */ - - movq 32(%ecx), %mm0 - movq 64(%ecx), %mm2 - movq %mm0, %mm1 - movq %mm2, %mm3 - pxor %mm7, %mm1 - pxor %mm7, %mm3 - pfacc %mm1, %mm0 - pfacc %mm3, %mm2 - pfmul %mm6, %mm0 - pfmul %mm6, %mm2 - movq %mm0, 32(%edx) - movq %mm2, 64(%edx) - - movd 44(%ecx), %mm0 - movd 40(%ecx), %mm2 - movd 120(%ebx), %mm3 - punpckldq 76(%ecx), %mm0 - punpckldq 72(%ecx), %mm2 - punpckldq %mm3, %mm3 - movq %mm0, %mm4 - movq %mm2, %mm5 - pfsub %mm2, %mm0 - pfmul %mm3, %mm0 - movq %mm0, %mm1 - pfadd %mm5, %mm0 - pfadd %mm4, %mm0 - movq %mm0, %mm2 - punpckldq %mm1, %mm0 - punpckhdq %mm1, %mm2 - movq %mm0, 40(%edx) - movq %mm2, 72(%edx) - - movd 48(%ecx), %mm3 - movd 60(%ecx), %mm2 - pfsub 52(%ecx), %mm3 - pfsub 56(%ecx), %mm2 - pfmul 120(%ebx), %mm3 - pfmul 120(%ebx), %mm2 - movq %mm2, %mm1 - - pfadd 56(%ecx), %mm1 - pfadd 60(%ecx), %mm1 - movq %mm1, %mm0 - - pfadd 48(%ecx), %mm0 - pfadd 52(%ecx), %mm0 - pfadd %mm3, %mm1 - punpckldq %mm2, %mm1 - pfadd %mm3, %mm2 - punpckldq %mm2, %mm0 - movq %mm1, 56(%edx) - movq %mm0, 48(%edx) - -/*---*/ - - movd 92(%ecx), %mm1 - pfsub 88(%ecx), %mm1 - pfmul 120(%ebx), %mm1 - movd %mm1, 92(%edx) - pfadd 92(%ecx), %mm1 - pfadd 88(%ecx), %mm1 - movq %mm1, %mm0 - - pfadd 80(%ecx), %mm0 - pfadd 84(%ecx), %mm0 - movd %mm0, 80(%edx) - - movd 80(%ecx), %mm0 - pfsub 84(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - pfadd %mm0, %mm1 - pfadd 92(%edx), %mm0 - punpckldq %mm1, %mm0 - movq %mm0, 84(%edx) - - movq 96(%ecx), %mm0 - movq %mm0, %mm1 - pxor %mm7, %mm1 - pfacc %mm1, %mm0 - pfmul %mm6, %mm0 - movq %mm0, 96(%edx) - - movd 108(%ecx), %mm0 - pfsub 104(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movd %mm0, 108(%edx) - pfadd 104(%ecx), %mm0 - pfadd 108(%ecx), %mm0 - movd %mm0, 104(%edx) - - movd 124(%ecx), %mm1 - pfsub 120(%ecx), %mm1 - pfmul 120(%ebx), %mm1 - movd %mm1, 124(%edx) - pfadd 120(%ecx), %mm1 - pfadd 124(%ecx), %mm1 - movq %mm1, %mm0 - - pfadd 112(%ecx), %mm0 - pfadd 116(%ecx), %mm0 - movd %mm0, 112(%edx) - - movd 112(%ecx), %mm0 - pfsub 116(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - pfadd %mm0,%mm1 - pfadd 124(%edx), %mm0 - punpckldq %mm1, %mm0 - movq %mm0, 116(%edx) - - jnz .L01 - -/* Phase 7*/ -/* Code below is coded in scalar mode. Should be optimized */ - - movd (%ecx), %mm0 - pfadd 4(%ecx), %mm0 - movd %mm0, 1024(%esi) - - movd (%ecx), %mm0 - pfsub 4(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movd %mm0, (%esi) - movd %mm0, (%edi) - - movd 12(%ecx), %mm0 - pfsub 8(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movd %mm0, 512(%edi) - pfadd 12(%ecx), %mm0 - pfadd 8(%ecx), %mm0 - movd %mm0, 512(%esi) - - movd 16(%ecx), %mm0 - pfsub 20(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movq %mm0, %mm3 - - movd 28(%ecx), %mm0 - pfsub 24(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - movd %mm0, 768(%edi) - movq %mm0, %mm2 - - pfadd 24(%ecx), %mm0 - pfadd 28(%ecx), %mm0 - movq %mm0, %mm1 - - pfadd 16(%ecx), %mm0 - pfadd 20(%ecx), %mm0 - movd %mm0, 768(%esi) - pfadd %mm3, %mm1 - movd %mm1, 256(%esi) - pfadd %mm3, %mm2 - movd %mm2, 256(%edi) - -/* Phase 8*/ - - movq 32(%edx), %mm0 - movq 48(%edx), %mm1 - pfadd 48(%edx), %mm0 - pfadd 40(%edx), %mm1 - movd %mm0, 896(%esi) - movd %mm1, 640(%esi) - psrlq $32, %mm0 - psrlq $32, %mm1 - movd %mm0, 128(%edi) - movd %mm1, 384(%edi) - - movd 40(%edx), %mm0 - pfadd 56(%edx), %mm0 - movd %mm0, 384(%esi) - - movd 56(%edx), %mm0 - pfadd 36(%edx), %mm0 - movd %mm0, 128(%esi) - - movd 60(%edx), %mm0 - movd %mm0, 896(%edi) - pfadd 44(%edx), %mm0 - movd %mm0, 640(%edi) - - movq 96(%edx), %mm0 - movq 112(%edx), %mm2 - movq 104(%edx), %mm4 - pfadd 112(%edx), %mm0 - pfadd 104(%edx), %mm2 - pfadd 120(%edx), %mm4 - movq %mm0, %mm1 - movq %mm2, %mm3 - movq %mm4, %mm5 - pfadd 64(%edx), %mm0 - pfadd 80(%edx), %mm2 - pfadd 72(%edx), %mm4 - movd %mm0, 960(%esi) - movd %mm2, 704(%esi) - movd %mm4, 448(%esi) - psrlq $32, %mm0 - psrlq $32, %mm2 - psrlq $32, %mm4 - movd %mm0, 64(%edi) - movd %mm2, 320(%edi) - movd %mm4, 576(%edi) - pfadd 80(%edx), %mm1 - pfadd 72(%edx), %mm3 - pfadd 88(%edx), %mm5 - movd %mm1, 832(%esi) - movd %mm3, 576(%esi) - movd %mm5, 320(%esi) - psrlq $32, %mm1 - psrlq $32, %mm3 - psrlq $32, %mm5 - movd %mm1, 192(%edi) - movd %mm3, 448(%edi) - movd %mm5, 704(%edi) - - movd 120(%edx), %mm0 - pfadd 100(%edx), %mm0 - movq %mm0, %mm1 - pfadd 88(%edx), %mm0 - movd %mm0, 192(%esi) - pfadd 68(%edx), %mm1 - movd %mm1, 64(%esi) - - movd 124(%edx), %mm0 - movd %mm0, 960(%edi) - pfadd 92(%edx), %mm0 - movd %mm0, 832(%edi) - - jmp .L_bye -.L01: -/* Phase 9*/ - - movq (%ecx), %mm0 - movq %mm0, %mm1 - pxor %mm7, %mm1 - pfacc %mm1, %mm0 - pfmul %mm6, %mm0 - pf2id %mm0, %mm0 - movd %mm0, %eax - movw %ax, 512(%esi) - psrlq $32, %mm0 - movd %mm0, %eax - movw %ax, (%esi) - - movd 12(%ecx), %mm0 - pfsub 8(%ecx), %mm0 - pfmul 120(%ebx), %mm0 - pf2id %mm0, %mm7 - movd %mm7, %eax - movw %ax, 256(%edi) - pfadd 12(%ecx), %mm0 - pfadd 8(%ecx), %mm0 - pf2id %mm0, %mm0 - movd %mm0, %eax - movw %ax, 256(%esi) - - movd 16(%ecx), %mm3 - pfsub 20(%ecx), %mm3 - pfmul 120(%ebx), %mm3 - movq %mm3, %mm2 - - movd 28(%ecx), %mm2 - pfsub 24(%ecx), %mm2 - pfmul 120(%ebx), %mm2 - movq %mm2, %mm1 - - pf2id %mm2, %mm7 - movd %mm7, %eax - movw %ax, 384(%edi) - - pfadd 24(%ecx), %mm1 - pfadd 28(%ecx), %mm1 - movq %mm1, %mm0 - - pfadd 16(%ecx), %mm0 - pfadd 20(%ecx), %mm0 - pf2id %mm0, %mm0 - movd %mm0, %eax - movw %ax, 384(%esi) - pfadd %mm3, %mm1 - pf2id %mm1, %mm1 - movd %mm1, %eax - movw %ax, 128(%esi) - pfadd %mm3, %mm2 - pf2id %mm2, %mm2 - movd %mm2, %eax - movw %ax, 128(%edi) - -/* Phase 10*/ - - movq 32(%edx), %mm0 - movq 48(%edx), %mm1 - pfadd 48(%edx), %mm0 - pfadd 40(%edx), %mm1 - pf2id %mm0, %mm0 - pf2id %mm1, %mm1 - movd %mm0, %eax - movd %mm1, %ecx - movw %ax, 448(%esi) - movw %cx, 320(%esi) - psrlq $32, %mm0 - psrlq $32, %mm1 - movd %mm0, %eax - movd %mm1, %ecx - movw %ax, 64(%edi) - movw %cx, 192(%edi) - - movd 40(%edx), %mm3 - movd 56(%edx), %mm4 - movd 60(%edx), %mm0 - movd 44(%edx), %mm2 - movd 120(%edx), %mm5 - punpckldq %mm4, %mm3 - punpckldq 124(%edx), %mm0 - pfadd 100(%edx), %mm5 - punpckldq 36(%edx), %mm4 - punpckldq 92(%edx), %mm2 - movq %mm5, %mm6 - pfadd %mm4, %mm3 - pf2id %mm0, %mm1 - pf2id %mm3, %mm3 - pfadd 88(%edx), %mm5 - movd %mm1, %eax - movd %mm3, %ecx - movw %ax, 448(%edi) - movw %cx, 192(%esi) - pf2id %mm5, %mm5 - psrlq $32, %mm1 - psrlq $32, %mm3 - movd %mm5, %ebx - movd %mm1, %eax - movd %mm3, %ecx - movw %bx, 96(%esi) - movw %ax, 480(%edi) - movw %cx, 64(%esi) - pfadd %mm2, %mm0 - pf2id %mm0, %mm0 - movd %mm0, %eax - pfadd 68(%edx), %mm6 - movw %ax, 320(%edi) - psrlq $32, %mm0 - pf2id %mm6, %mm6 - movd %mm0, %eax - movd %mm6, %ebx - movw %ax, 416(%edi) - movw %bx, 32(%esi) - - movq 96(%edx), %mm0 - movq 112(%edx), %mm2 - movq 104(%edx), %mm4 - pfadd %mm2, %mm0 - pfadd %mm4, %mm2 - pfadd 120(%edx), %mm4 - movq %mm0, %mm1 - movq %mm2, %mm3 - movq %mm4, %mm5 - pfadd 64(%edx), %mm0 - pfadd 80(%edx), %mm2 - pfadd 72(%edx), %mm4 - pf2id %mm0, %mm0 - pf2id %mm2, %mm2 - pf2id %mm4, %mm4 - movd %mm0, %eax - movd %mm2, %ecx - movd %mm4, %ebx - movw %ax, 480(%esi) - movw %cx, 352(%esi) - movw %bx, 224(%esi) - psrlq $32, %mm0 - psrlq $32, %mm2 - psrlq $32, %mm4 - movd %mm0, %eax - movd %mm2, %ecx - movd %mm4, %ebx - movw %ax, 32(%edi) - movw %cx, 160(%edi) - movw %bx, 288(%edi) - pfadd 80(%edx), %mm1 - pfadd 72(%edx), %mm3 - pfadd 88(%edx), %mm5 - pf2id %mm1, %mm1 - pf2id %mm3, %mm3 - pf2id %mm5, %mm5 - movd %mm1, %eax - movd %mm3, %ecx - movd %mm5, %ebx - movw %ax, 416(%esi) - movw %cx, 288(%esi) - movw %bx, 160(%esi) - psrlq $32, %mm1 - psrlq $32, %mm3 - psrlq $32, %mm5 - movd %mm1, %eax - movd %mm3, %ecx - movd %mm5, %ebx - movw %ax, 96(%edi) - movw %cx, 224(%edi) - movw %bx, 352(%edi) - - movsw - -.L_bye: - addl $256,%esp - femms - popl %edi - popl %esi - popl %ebx - ret $12