changeset 23378:ef54df9f07d3

HAVE_MMX1 -> HAVE_ONLY_MMX1 (makes more sense ...)
author michael
date Sun, 27 May 2007 21:45:33 +0000
parents d43ce381e578
children 0c6e8d9da11a
files libvo/aclib_template.c
diffstat 1 files changed, 6 insertions(+), 6 deletions(-) [+]
line wrap: on
line diff
--- a/libvo/aclib_template.c	Sun May 27 21:14:32 2007 +0000
+++ b/libvo/aclib_template.c	Sun May 27 21:45:33 2007 +0000
@@ -64,14 +64,14 @@
 //  by Pontscho/fresh!mindworkz
 
 
-#undef HAVE_MMX1
+#undef HAVE_ONLY_MMX1
 #if defined(HAVE_MMX) && !defined(HAVE_MMX2) && !defined(HAVE_3DNOW) && !defined(HAVE_SSE)
 /*  means: mmx v.1. Note: Since we added alignment of destinition it speedups
     of memory copying on PentMMX, Celeron-1 and P2 upto 12% versus
     standard (non MMX-optimized) version.
     Note: on K6-2+ it speedups memory copying upto 25% and
           on K7 and P3 about 500% (5 times). */
-#define HAVE_MMX1
+#define HAVE_ONLY_MMX1
 #endif
 
 
@@ -127,7 +127,7 @@
 #endif
 
 #undef MIN_LEN
-#ifdef HAVE_MMX1
+#ifdef HAVE_ONLY_MMX1
 #define MIN_LEN 0x800  /* 2K blocks */
 #else
 #define MIN_LEN 0x40  /* 64-byte blocks */
@@ -151,7 +151,7 @@
 				printf("freq < %8d %4d\n", 1<<i, freq[i]);
 	}
 #endif
-#ifndef HAVE_MMX1
+#ifndef HAVE_ONLY_MMX1
         /* PREFETCH has effect even for MOVSB instruction ;) */
 	__asm__ __volatile__ (
 	        PREFETCH" (%0)\n"
@@ -229,7 +229,7 @@
 	for(; ((int)to & (BLOCK_SIZE-1)) && i>0; i--)
 	{
 		__asm__ __volatile__ (
-#ifndef HAVE_MMX1
+#ifndef HAVE_ONLY_MMX1
         	PREFETCH" 320(%0)\n"
 #endif
 		"movq (%0), %%mm0\n"
@@ -318,7 +318,7 @@
 	for(; i>0; i--)
 	{
 		__asm__ __volatile__ (
-#ifndef HAVE_MMX1
+#ifndef HAVE_ONLY_MMX1
         	PREFETCH" 320(%0)\n"
 #endif
 		"movq (%0), %%mm0\n"