changeset 27757:b5a46071062a

Replace all occurrences of '__volatile__' and '__volatile' by plain 'volatile'. We were using an inconsistent mix of the three variants and 'volatile' should be the most correct and portable variant.
author diego
date Thu, 16 Oct 2008 20:17:56 +0000
parents 1266470a5651
children 7929bcb921e9
files TOOLS/fastmemcpybench.c configure cpudetect.c drivers/radeon_vid.c liba52/imdct_3dnow.h liba52/srfftp_3dnow.h libmpcodecs/dec_video.c libmpcodecs/native/mmx.h libvo/aclib_template.c loader/drv.c loader/ldt_keeper.c loader/qtx/qtxload.c loader/win32.c loader/wine/windef.h mp3lib/dct36_3dnow.c mp3lib/dct64_3dnow.c mp3lib/dct64_k7.c mp3lib/dct64_mmx.c mp3lib/decode_i586.c mp3lib/decode_mmx.c vidix/nvidia_vid.c vidix/s3_regs.h vidix/sysdep/AsmMacros_powerpc.h vidix/sysdep/AsmMacros_sparc.h vidix/sysdep/AsmMacros_x86.h
diffstat 25 files changed, 101 insertions(+), 101 deletions(-) [+]
line wrap: on
line diff
--- a/TOOLS/fastmemcpybench.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/TOOLS/fastmemcpybench.c	Thu Oct 16 20:17:56 2008 +0000
@@ -86,7 +86,7 @@
 static inline unsigned long long int read_tsc( void )
 {
   unsigned long long int retval;
-  __asm__ __volatile ("rdtsc":"=A"(retval)::"memory");
+  __asm__ volatile ("rdtsc":"=A"(retval)::"memory");
   return retval;
 }
 
--- a/configure	Thu Oct 16 19:24:21 2008 +0000
+++ b/configure	Thu Oct 16 20:17:56 2008 +0000
@@ -1595,7 +1595,7 @@
 void catch() { exit(1); }
 int main(void) {
   signal(SIGILL, catch);
-  __asm__ __volatile__ ("$3":::"memory"); return 0;
+  __asm__ volatile ("$3":::"memory"); return 0;
 }
 EOF
 
@@ -2443,7 +2443,7 @@
   echocheck "ARMv5TE (Enhanced DSP Extensions)"
   if test $_armv5te = "auto" ; then
     cat > $TMPC << EOF
-int main(void) { __asm__ __volatile__ ("qadd r0, r0, r0"); return 0; }
+int main(void) { __asm__ volatile ("qadd r0, r0, r0"); return 0; }
 EOF
     _armv5te=no
     cc_check && _armv5te=yes
@@ -2453,7 +2453,7 @@
   echocheck "ARMv6 (SIMD instructions)"
   if test $_armv6 = "auto" ; then
     cat > $TMPC << EOF
-int main(void) { __asm__ __volatile__ ("sadd16 r0, r0, r0"); return 0; }
+int main(void) { __asm__ volatile ("sadd16 r0, r0, r0"); return 0; }
 EOF
     _armv6=no
     cc_check && _armv6=yes
@@ -2463,7 +2463,7 @@
   echocheck "ARM VFP"
   if test $_armvfp = "auto" ; then
     cat > $TMPC << EOF
-int main(void) { __asm__ __volatile__ ("fadds s0, s0, s0"); return 0; }
+int main(void) { __asm__ volatile ("fadds s0, s0, s0"); return 0; }
 EOF
     _armvfp=no
     cc_check && _armvfp=yes
@@ -2473,7 +2473,7 @@
   echocheck "iWMMXt (Intel XScale SIMD instructions)"
   if test $_iwmmxt = "auto" ; then
     cat > $TMPC << EOF
-int main(void) { __asm__ __volatile__ ("wunpckelub wr6, wr4"); return 0; }
+int main(void) { __asm__ volatile ("wunpckelub wr6, wr4"); return 0; }
 EOF
     _iwmmxt=no
     cc_check && _iwmmxt=yes
--- a/cpudetect.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/cpudetect.c	Thu Oct 16 20:17:56 2008 +0000
@@ -57,7 +57,7 @@
 	long a, c;
 
 // code from libavcodec:
-    __asm__ __volatile__ (
+    __asm__ volatile (
                           /* See if CPUID instruction is supported ... */
                           /* ... Get copies of EFLAGS into eax and ecx */
                           "pushf\n\t"
@@ -85,14 +85,14 @@
 do_cpuid(unsigned int ax, unsigned int *p)
 {
 #if 0
-	__asm__ __volatile(
+	__asm__ volatile(
 	"cpuid;"
 	: "=a" (p[0]), "=b" (p[1]), "=c" (p[2]), "=d" (p[3])
 	:  "0" (ax)
 	);
 #else
 // code from libavcodec:
-    __asm__ __volatile__
+    __asm__ volatile
 	("mov %%"REG_b", %%"REG_S"\n\t"
          "cpuid\n\t"
          "xchg %%"REG_b", %%"REG_S
@@ -400,7 +400,7 @@
    if ( gCpuCaps.hasSSE ) {
       mp_msg(MSGT_CPUDETECT,MSGL_V, "Testing OS support for SSE... " );
       exc_fil = SetUnhandledExceptionFilter(win32_sig_handler_sse);
-      __asm__ __volatile ("xorps %xmm0, %xmm0");
+      __asm__ volatile ("xorps %xmm0, %xmm0");
       SetUnhandledExceptionFilter(exc_fil);
       mp_msg(MSGT_CPUDETECT,MSGL_V, gCpuCaps.hasSSE ? "yes.\n" : "no!\n" );
    }
@@ -409,7 +409,7 @@
    if ( gCpuCaps.hasSSE ) {
       mp_msg(MSGT_CPUDETECT,MSGL_V, "Testing OS support for SSE... " );
       DosSetExceptionHandler( &RegRec );
-      __asm__ __volatile ("xorps %xmm0, %xmm0");
+      __asm__ volatile ("xorps %xmm0, %xmm0");
       DosUnsetExceptionHandler( &RegRec );
       mp_msg(MSGT_CPUDETECT,MSGL_V, gCpuCaps.hasSSE ? "yes.\n" : "no!\n" );
    }
@@ -432,8 +432,8 @@
    if ( gCpuCaps.hasSSE ) {
       mp_msg(MSGT_CPUDETECT,MSGL_V, "Testing OS support for SSE... " );
 
-//      __asm__ __volatile ("xorps %%xmm0, %%xmm0");
-      __asm__ __volatile ("xorps %xmm0, %xmm0");
+//      __asm__ volatile ("xorps %%xmm0, %%xmm0");
+      __asm__ volatile ("xorps %xmm0, %xmm0");
 
       mp_msg(MSGT_CPUDETECT,MSGL_V, gCpuCaps.hasSSE ? "yes.\n" : "no!\n" );
    }
--- a/drivers/radeon_vid.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/drivers/radeon_vid.c	Thu Oct 16 20:17:56 2008 +0000
@@ -129,7 +129,7 @@
 static inline double FastSin(double x)
 {
    register double res;
-   __asm__ __volatile("fsin":"=t"(res):"0"(x));
+   __asm__ volatile("fsin":"=t"(res):"0"(x));
    return res;
 }
 #undef sin
@@ -138,7 +138,7 @@
 static inline double FastCos(double x)
 {
    register double res;
-   __asm__ __volatile("fcos":"=t"(res):"0"(x));
+   __asm__ volatile("fcos":"=t"(res):"0"(x));
    return res;
 }
 #undef cos
--- a/liba52/imdct_3dnow.h	Thu Oct 16 19:24:21 2008 +0000
+++ b/liba52/imdct_3dnow.h	Thu Oct 16 20:17:56 2008 +0000
@@ -45,7 +45,7 @@
   /* delta_p = 1 here */
   /* x[k] = sum_{i=0..3} x[i] * w^{i*k}, w=e^{-2*pi/4} 
    */
-  __asm__ __volatile__(
+  __asm__ volatile(
 	"movq	24(%1), %%mm3\n\t"
 	"movq	8(%1), %%mm1\n\t"
 	"pxor	%2, %%mm3\n\t" /* mm3.re | -mm3.im */
@@ -90,7 +90,7 @@
    */
   complex_t wT1, wB1, wB2;
   
-  __asm__ __volatile__(
+  __asm__ volatile(
 	"movq	8(%2), %%mm0\n\t"
 	"movq	24(%2), %%mm1\n\t"
 	"movq	%%mm0, %0\n\t"  /* wT1 = x[1]; */
@@ -99,7 +99,7 @@
 	:"r"(x)
 	:"memory");
 
-  __asm__ __volatile__(
+  __asm__ volatile(
 	"movq	16(%0), %%mm2\n\t"
 	"movq	32(%0), %%mm3\n\t"
 	"movq	%%mm2, 8(%0)\n\t"  /* x[1] = x[2]; */
@@ -114,7 +114,7 @@
   
   /* x[0] x[4] x[2] x[6] */
   
-  __asm__ __volatile__(
+  __asm__ volatile(
       "movq	40(%1), %%mm0\n\t"
       "movq	%%mm0,	%%mm3\n\t"
       "movq	56(%1),	%%mm1\n\t"
@@ -153,7 +153,7 @@
       :"memory");
   
   /* x[1] x[5] */
-  __asm__ __volatile__ (
+  __asm__ volatile (
 	"movq	%6,	%%mm6\n\t"
 	"movq	%5,	%%mm7\n\t"
 	"movq	%1,	%%mm0\n\t"
@@ -203,7 +203,7 @@
 
 
   /* x[3] x[7] */
-  __asm__ __volatile__(
+  __asm__ volatile(
 	"movq	%1,	%%mm0\n\t"
 #ifdef HAVE_3DNOWEX
 	"pswapd	%3,	%%mm1\n\t"
@@ -358,13 +358,13 @@
 	
     /* Pre IFFT complex multiply plus IFFT cmplx conjugate & reordering*/
 #if 1
-      __asm__ __volatile__ (
+      __asm__ volatile (
 	"movq %0, %%mm7\n\t"
 	::"m"(x_plus_minus_3dnow)
 	:"memory");
 	for( i=0; i < 128; i++) {
 		int j = pm128[i];
-	__asm__ __volatile__ (
+	__asm__ volatile (
 		"movd	%1, %%mm0\n\t"
 		"movd	%3, %%mm1\n\t"
 		"punpckldq %2, %%mm0\n\t" /* mm0 = data[256-2*j-1] | data[2*j]*/
@@ -394,7 +394,7 @@
 		buf[i].im = (data[256-2*j-1] * xsin1[j] + data[2*j] * xcos1[j])*(-1.0);*/
 	}
 #else
-  __asm__ __volatile__ ("femms":::"memory");
+  __asm__ volatile ("femms":::"memory");
     for( i=0; i < 128; i++) {
 	/* z[i] = (X[256-2*i-1] + j * X[2*i]) * (xcos1[i] + j * xsin1[i]) ; */ 
 	int j= pm128[i];
@@ -435,14 +435,14 @@
     
     /* Post IFFT complex multiply  plus IFFT complex conjugate*/
 #if 1  
-  __asm__ __volatile__ (
+  __asm__ volatile (
 	"movq %0, %%mm7\n\t"
 	"movq %1, %%mm6\n\t"
 	::"m"(x_plus_minus_3dnow),
 	"m"(x_minus_plus_3dnow)
 	:"eax","memory");
 	for (i=0; i < 128; i++) {
-	    __asm__ __volatile__ (
+	    __asm__ volatile (
 		"movq %1, %%mm0\n\t" /* ac3_buf[i].re | ac3_buf[i].im */
 		"movq %%mm0, %%mm1\n\t" /* ac3_buf[i].re | ac3_buf[i].im */
 #ifndef HAVE_3DNOWEX
@@ -473,7 +473,7 @@
 		ac3_buf[i].im =(tmp_a_r * ac3_xsin1[i])  -  (tmp_a_i  * ac3_xcos1[i]);*/
 	}
 #else    
-  __asm__ __volatile__ ("femms":::"memory");
+  __asm__ volatile ("femms":::"memory");
     for( i=0; i < 128; i++) {
 	/* y[n] = z[n] * (xcos1[n] + j * xsin1[n]) ; */
 	tmp_a_r =        buf[i].real;
@@ -496,7 +496,7 @@
 	);
 	for (i=0; i< 64; i++) {
 /* merge two loops in one to enable working of 2 decoders */
-	__asm__ __volatile__ (
+	__asm__ volatile (
 		"movd	516(%1), %%mm0\n\t"
 		"movd	(%1), %%mm1\n\t" /**data_ptr++=-buf[64+i].im**window_ptr+++*delay_ptr++;*/
 		"punpckldq (%2), %%mm0\n\t"/*data_ptr[128]=-buf[i].re*window_ptr[128]+delay_ptr[128];*/
@@ -520,7 +520,7 @@
 	}
 	window_ptr += 128;
 #else    
-  __asm__ __volatile__ ("femms":::"memory");
+  __asm__ volatile ("femms":::"memory");
     for(i=0; i< 64; i++) { 
 	*data_ptr++   = -buf[64+i].imag   * *window_ptr++ + *delay_ptr++ + bias; 
 	*data_ptr++   =  buf[64-i-1].real * *window_ptr++ + *delay_ptr++ + bias; 
@@ -538,7 +538,7 @@
 	for(i=0; i< 64; i++) {
 /* merge two loops in one to enable working of 2 decoders */
 	    window_ptr -=2;
-	    __asm__ __volatile__(
+	    __asm__ volatile(
 		"movd	508(%1), %%mm0\n\t"
 		"movd	(%1), %%mm1\n\t"
 		"punpckldq (%2), %%mm0\n\t"
@@ -565,9 +565,9 @@
 		:"memory");
 		delay_ptr += 2;
 	}
-  __asm__ __volatile__ ("femms":::"memory");
+  __asm__ volatile ("femms":::"memory");
 #else    
-  __asm__ __volatile__ ("femms":::"memory");
+  __asm__ volatile ("femms":::"memory");
     for(i=0; i< 64; i++) { 
 	*delay_ptr++  = -buf[64+i].real   * *--window_ptr; 
 	*delay_ptr++  =  buf[64-i-1].imag * *--window_ptr; 
--- a/liba52/srfftp_3dnow.h	Thu Oct 16 19:24:21 2008 +0000
+++ b/liba52/srfftp_3dnow.h	Thu Oct 16 20:17:56 2008 +0000
@@ -39,7 +39,7 @@
 }i_cmplx_t;
 
 #define TRANS_FILL_MM6_MM7_3DNOW()\
-    __asm__ __volatile__(\
+    __asm__ volatile(\
 	"movq	%1, %%mm7\n\t"\
 	"movq	%0, %%mm6\n\t"\
 	::"m"(x_plus_minus_3dnow),\
@@ -66,7 +66,7 @@
 
 #define TRANSZERO_3DNOW(A0,A4,A8,A12) \
 { \
-    __asm__ __volatile__(\
+    __asm__ volatile(\
 	"movq	%4, %%mm0\n\t" /* mm0 = wTB[0]*/\
 	"movq	%5, %%mm1\n\t" /* mm1 = wTB[k*2]*/ \
 	"movq	%%mm0, %%mm5\n\t"/*u.re = wTB[0].re + wTB[k*2].re;*/\
@@ -95,7 +95,7 @@
 
 #define TRANSHALF_16_3DNOW(A2,A6,A10,A14)\
 {\
-    __asm__ __volatile__(\
+    __asm__ volatile(\
 	"movq	%4, %%mm0\n\t"/*u.re = wTB[2].im + wTB[2].re;*/\
 	"movq	%%mm0, %%mm1\n\t"\
 	"pxor	%%mm7, %%mm1\n\t"\
@@ -136,7 +136,7 @@
 
 #define TRANS_3DNOW(A1,A5,A9,A13,WT,WB,D,D3)\
 { \
-    __asm__ __volatile__(\
+    __asm__ volatile(\
 	"movq	%1,	%%mm4\n\t"\
 	"movq	%%mm4,	%%mm5\n\t"\
 	"punpckldq %%mm4, %%mm4\n\t"/*mm4 = D.re | D.re */\
@@ -166,7 +166,7 @@
 	:\
 	:"m"(WT), "m"(D), "m"(WB), "m"(D3)\
 	:"memory");\
-    __asm__ __volatile__(\
+    __asm__ volatile(\
 	"movq	%4, %%mm0\n\t"/* a1 = A1*/\
 	"movq	%5, %%mm2\n\t"/* a1 = A5*/\
 	"movq	%%mm0, %%mm1\n\t"\
--- a/libmpcodecs/dec_video.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/libmpcodecs/dec_video.c	Thu Oct 16 20:17:56 2008 +0000
@@ -374,10 +374,10 @@
     // some codecs are broken, and doesn't restore MMX state :(
     // it happens usually with broken/damaged files.
     if (gCpuCaps.has3DNow) {
-	__asm__ __volatile ("femms\n\t":::"memory");
+	__asm__ volatile ("femms\n\t":::"memory");
     }
     else if (gCpuCaps.hasMMX) {
-	__asm__ __volatile ("emms\n\t":::"memory");
+	__asm__ volatile ("emms\n\t":::"memory");
     }
 #endif
 
--- a/libmpcodecs/native/mmx.h	Thu Oct 16 19:24:21 2008 +0000
+++ b/libmpcodecs/native/mmx.h	Thu Oct 16 20:17:56 2008 +0000
@@ -57,24 +57,24 @@
 
 
 #define mmx_i2r(op,imm,reg) \
-        __asm__ __volatile__ (#op " %0, %%" #reg \
+        __asm__ volatile (#op " %0, %%" #reg \
                               : /* nothing */ \
                               : "i" (imm) )
 
 #define mmx_m2r(op, mem, reg) \
-        __asm__ __volatile__ (#op " %0, %%" #reg \
+        __asm__ volatile (#op " %0, %%" #reg \
                               : /* nothing */ \
                               : "m" (mem))
 
 #define mmx_r2m(op, reg, mem) \
-        __asm__ __volatile__ (#op " %%" #reg ", %0" \
+        __asm__ volatile (#op " %%" #reg ", %0" \
                               : "=m" (mem) \
                               : /* nothing */ )
 
 #define mmx_r2r(op, regs, regd) \
-        __asm__ __volatile__ (#op " %" #regs ", %" #regd)
+        __asm__ volatile (#op " %" #regs ", %" #regd)
 
 
-#define emms() __asm__ __volatile__ ("emms")
+#define emms() __asm__ volatile ("emms")
 
 #endif /* MPLAYER_MMX_H */
--- a/libvo/aclib_template.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/libvo/aclib_template.c	Thu Oct 16 20:17:56 2008 +0000
@@ -84,7 +84,7 @@
 #define small_memcpy(to,from,n)\
 {\
 register unsigned long int dummy;\
-__asm__ __volatile__(\
+__asm__ volatile(\
 	"rep; movsb"\
 	:"=&D"(to), "=&S"(from), "=&c"(dummy)\
 /* It's most portable way to notify compiler */\
@@ -153,7 +153,7 @@
 #endif
 #ifndef HAVE_ONLY_MMX1
         /* PREFETCH has effect even for MOVSB instruction ;) */
-	__asm__ __volatile__ (
+	__asm__ volatile (
 	        PREFETCH" (%0)\n"
 	        PREFETCH" 64(%0)\n"
 	        PREFETCH" 128(%0)\n"
@@ -188,7 +188,7 @@
 	/* if SRC is misaligned */
 	for(; i>0; i--)
 	{
-		__asm__ __volatile__ (
+		__asm__ volatile (
 		PREFETCH" 320(%0)\n"
 		"movups (%0), %%xmm0\n"
 		"movups 16(%0), %%xmm1\n"
@@ -210,7 +210,7 @@
 	*/
 	for(; i>0; i--)
 	{
-		__asm__ __volatile__ (
+		__asm__ volatile (
 		PREFETCH" 320(%0)\n"
 		"movaps (%0), %%xmm0\n"
 		"movaps 16(%0), %%xmm1\n"
@@ -228,7 +228,7 @@
 	// Align destination at BLOCK_SIZE boundary
 	for(; ((int)to & (BLOCK_SIZE-1)) && i>0; i--)
 	{
-		__asm__ __volatile__ (
+		__asm__ volatile (
 #ifndef HAVE_ONLY_MMX1
         	PREFETCH" 320(%0)\n"
 #endif
@@ -317,7 +317,7 @@
 
 	for(; i>0; i--)
 	{
-		__asm__ __volatile__ (
+		__asm__ volatile (
 #ifndef HAVE_ONLY_MMX1
         	PREFETCH" 320(%0)\n"
 #endif
@@ -346,11 +346,11 @@
 #ifdef HAVE_MMX2
                 /* since movntq is weakly-ordered, a "sfence"
 		 * is needed to become ordered again. */
-		__asm__ __volatile__ ("sfence":::"memory");
+		__asm__ volatile ("sfence":::"memory");
 #endif
 #ifndef HAVE_SSE
 		/* enables to use FPU */
-		__asm__ __volatile__ (EMMS:::"memory");
+		__asm__ volatile (EMMS:::"memory");
 #endif
 	}
 	/*
@@ -405,7 +405,7 @@
         */
 	for(; i>0; i--)
 	{
-		__asm__ __volatile__ (
+		__asm__ volatile (
         	PREFETCH" 320(%0)\n"
 		"movq (%0), %%mm0\n"
 		"movq 8(%0), %%mm1\n"
@@ -430,10 +430,10 @@
 #ifdef HAVE_MMX2
                 /* since movntq is weakly-ordered, a "sfence"
 		 * is needed to become ordered again. */
-		__asm__ __volatile__ ("sfence":::"memory");
+		__asm__ volatile ("sfence":::"memory");
 #endif
 		/* enables to use FPU */
-		__asm__ __volatile__ (EMMS:::"memory");
+		__asm__ volatile (EMMS:::"memory");
 	}
 	/*
 	 *	Now do the tail of the block
--- a/loader/drv.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/loader/drv.c	Thu Oct 16 20:17:56 2008 +0000
@@ -52,7 +52,7 @@
 #else
 // this asm code is no longer needed
 #define STORE_ALL \
-    __asm__ __volatile__ ( \
+    __asm__ volatile ( \
     "push %%ebx\n\t" \
     "push %%ecx\n\t" \
     "push %%edx\n\t" \
@@ -60,7 +60,7 @@
     "push %%edi\n\t"::)
 
 #define REST_ALL \
-    __asm__ __volatile__ ( \
+    __asm__ volatile ( \
     "pop %%edi\n\t" \
     "pop %%esi\n\t" \
     "pop %%edx\n\t" \
@@ -98,7 +98,7 @@
 #endif
     if (!module || !module->hDriverModule || !module->DriverProc) return -1;
 #ifndef __svr4__
-    __asm__ __volatile__ ("fsave (%0)\n\t": :"r"(&qw));
+    __asm__ volatile ("fsave (%0)\n\t": :"r"(&qw));
 #endif
 
 #ifdef WIN32_LOADER
@@ -110,7 +110,7 @@
     REST_ALL;
 
 #ifndef __svr4__
-    __asm__ __volatile__ ("frstor (%0)\n\t": :"r"(&qw));
+    __asm__ volatile ("frstor (%0)\n\t": :"r"(&qw));
 #endif
 
 #ifdef DETAILED_OUT
--- a/loader/ldt_keeper.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/loader/ldt_keeper.c	Thu Oct 16 20:17:56 2008 +0000
@@ -138,7 +138,7 @@
 {
     unsigned int ldt_desc = LDT_SEL(fs_ldt);
 
-    __asm__ __volatile__(
+    __asm__ volatile(
 	"movl %0,%%eax; movw %%ax, %%fs" : : "r" (ldt_desc)
 	:"eax"
     );
@@ -154,7 +154,7 @@
 {
     int res;
 #ifdef __PIC__
-    __asm__ __volatile__( "pushl %%ebx\n\t"
+    __asm__ volatile( "pushl %%ebx\n\t"
 			  "movl %2,%%ebx\n\t"
 			  "int $0x80\n\t"
 			  "popl %%ebx"
@@ -165,7 +165,7 @@
 			  "d"(16)//sizeof(*ptr) from kernel point of view
 			  :"esi"     );
 #else
-    __asm__ __volatile__("int $0x80"
+    __asm__ volatile("int $0x80"
 			 : "=a" (res)
 			 : "0" (__NR_modify_ldt),
 			 "b" (func),
--- a/loader/qtx/qtxload.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/loader/qtx/qtxload.c	Thu Oct 16 20:17:56 2008 +0000
@@ -104,7 +104,7 @@
     printf("params: flags: %d, paramSize: %d, what: %d, params[0] = %x\n",
         params->flags, params->paramSize, params->what, params->params[0]);
 
-//    __asm__ __volatile__ ("movl %%esp, %0\n\t" : "=a" (esp) :: "memory" );
+//    __asm__ volatile ("movl %%esp, %0\n\t" : "=a" (esp) :: "memory" );
 //    printf("ESP=%p\n",esp);
 
     *((void**)0x62b7d640) = &x_table[0]; //malloc(0x00001837 * 4); // ugly hack?
@@ -113,7 +113,7 @@
 
     ret = dispatcher(params, &globals);
 
-//    __asm__ __volatile__ ("movl %%esp, %0\n\t" : "=a" (esp) :: "memory" );
+//    __asm__ volatile ("movl %%esp, %0\n\t" : "=a" (esp) :: "memory" );
 //    printf("ESP=%p\n",esp);
 
     printf("!!! CDComponentDispatch() => %d  glob=%p\n",ret,globals);
--- a/loader/win32.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/loader/win32.c	Thu Oct 16 20:17:56 2008 +0000
@@ -79,7 +79,7 @@
 
 static void do_cpuid(unsigned int ax, unsigned int *regs)
 {
-    __asm__ __volatile__
+    __asm__ volatile
 	(
 	 "pushl %%ebx; pushl %%ecx; pushl %%edx;"
 	 ".byte  0x0f, 0xa2;"
@@ -95,7 +95,7 @@
 static unsigned int c_localcount_tsc()
 {
     int a;
-    __asm__ __volatile__
+    __asm__ volatile
 	(
 	 "rdtsc\n\t"
 	 :"=a"(a)
@@ -106,7 +106,7 @@
 }
 static void c_longcount_tsc(long long* z)
 {
-    __asm__ __volatile__
+    __asm__ volatile
 	(
 	 "pushl %%ebx\n\t"
 	 "movl %%eax, %%ebx\n\t"
@@ -2867,7 +2867,7 @@
 static long WINAPI expInterlockedExchangeAdd( long* dest, long incr )
 {
     long ret;
-    __asm__ __volatile__
+    __asm__ volatile
 	(
 	 "lock; xaddl %0,(%1)"
 	 : "=r" (ret)
@@ -4006,7 +4006,7 @@
 	    // ok this trick with push/pop is necessary as otherwice
 	    // edi/esi registers are being trashed
 	    void* p = *start;
-	    __asm__ __volatile__
+	    __asm__ volatile
 		(
 		 "pushl %%ebx		\n\t"
 		 "pushl %%ecx		\n\t"
@@ -4272,7 +4272,7 @@
 
 static void explog10(void)
 {
-    __asm__ __volatile__
+    __asm__ volatile
 	(
 	 "fldl 8(%esp)	\n\t"
 	 "fldln2	\n\t"
@@ -4283,7 +4283,7 @@
 
 static void expcos(void)
 {
-    __asm__ __volatile__
+    __asm__ volatile
 	(
 	 "fldl 8(%esp)	\n\t"
 	 "fcos		\n\t"
@@ -4300,7 +4300,7 @@
 
 static void exp_ftol(void)
 {
-    __asm__ __volatile__
+    __asm__ volatile
 	(
 	 "sub $12, %esp		\n\t"
 	 "fstcw   -2(%ebp)	\n\t"
@@ -4319,8 +4319,8 @@
 }
 
 #define FPU_DOUBLES(var1,var2) double var1,var2; \
-  __asm__ __volatile__( "fstpl %0;fwait" : "=m" (var2) : ); \
-  __asm__ __volatile__( "fstpl %0;fwait" : "=m" (var1) : )
+  __asm__ volatile( "fstpl %0;fwait" : "=m" (var2) : ); \
+  __asm__ volatile( "fstpl %0;fwait" : "=m" (var1) : )
 
 static double exp_CIpow(void)
 {
@@ -4361,7 +4361,7 @@
 {
     //dbgprintf("!!!!UNIMPLEMENTED: setjmp3(%p, %d) => 0\n", jmpbuf, x);
     //return 0;
-    __asm__ __volatile__
+    __asm__ volatile
 	(
 	 //"mov 4(%%esp), %%edx	\n\t"
 	 "mov (%%esp), %%eax   \n\t"
@@ -4383,7 +4383,7 @@
 	 : "eax"
 	);
 #if 1
-    __asm__ __volatile__
+    __asm__ volatile
 	(
 	 "mov %%fs:0, %%eax	\n\t" // unsure
 	 "mov %%eax, 24(%%edx)	\n\t"
@@ -4762,7 +4762,7 @@
 }
 
 #define FPU_DOUBLE(var) double var; \
-  __asm__ __volatile__( "fstpl %0;fwait" : "=m" (var) : )
+  __asm__ volatile( "fstpl %0;fwait" : "=m" (var) : )
 
 static double exp_CIcos(void)
 {
--- a/loader/wine/windef.h	Thu Oct 16 19:24:21 2008 +0000
+++ b/loader/wine/windef.h	Thu Oct 16 20:17:56 2008 +0000
@@ -94,7 +94,7 @@
 #  ifndef _EGCS_ 
 #define __stdcall __attribute__((__stdcall__))
 #define __cdecl   __attribute__((__cdecl__))
-#  define RESTORE_ES  __asm__ __volatile__("pushl %ds\n\tpopl %es")
+#  define RESTORE_ES  __asm__ volatile("pushl %ds\n\tpopl %es")
 #  endif
 # else
 // #  error You need gcc >= 2.7 to build Wine on a 386
--- a/mp3lib/dct36_3dnow.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/mp3lib/dct36_3dnow.c	Thu Oct 16 20:17:56 2008 +0000
@@ -39,7 +39,7 @@
     real *o2, real *wintab, real *tsbuf)
 #endif
 {
-    __asm__ __volatile__(
+    __asm__ volatile(
 	"movq (%%eax),%%mm0\n\t"
 	"movq 4(%%eax),%%mm1\n\t"
 	"pfadd %%mm1,%%mm0\n\t"
--- a/mp3lib/dct64_3dnow.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/mp3lib/dct64_3dnow.c	Thu Oct 16 20:17:56 2008 +0000
@@ -18,7 +18,7 @@
 void dct64_MMX_3dnow(short *a,short *b,real *c)
 {
   char tmp[256];
-    __asm__ __volatile(
+    __asm__ volatile(
 "	movl %2,%%eax\n\t"
 
 "	leal 128+%3,%%edx\n\t"
--- a/mp3lib/dct64_k7.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/mp3lib/dct64_k7.c	Thu Oct 16 20:17:56 2008 +0000
@@ -18,7 +18,7 @@
 void dct64_MMX_3dnowex(short *a,short *b,real *c)
 {
   char tmp[256];
-    __asm__ __volatile(
+    __asm__ volatile(
 "	movl %2,%%eax\n\t"
 
 "	leal 128+%3,%%edx\n\t"
--- a/mp3lib/dct64_mmx.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/mp3lib/dct64_mmx.c	Thu Oct 16 20:17:56 2008 +0000
@@ -9,7 +9,7 @@
 void dct64_MMX(short *a,short *b,real *c)
 {
     char tmp[256];
-    __asm__ __volatile(
+    __asm__ volatile(
 "	movl %2,%%eax\n\t"
 /* Phase 1*/
 "	flds     (%%eax)\n\t"
--- a/mp3lib/decode_i586.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/mp3lib/decode_i586.c	Thu Oct 16 20:17:56 2008 +0000
@@ -41,7 +41,7 @@
 {
   real tmp[3];
   register int retval;
-    __asm__ __volatile(
+    __asm__ volatile(
 "        movl %%ebp,"MANGLE(saved_ebp)"\n\t"
 "        movl %1,%%eax\n\t"/*bandPtr*/
 "        movl %3,%%esi\n\t"
--- a/mp3lib/decode_mmx.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/mp3lib/decode_mmx.c	Thu Oct 16 20:17:56 2008 +0000
@@ -217,7 +217,7 @@
     dct64_MMX_func(a, b, bandPtr);
     window = mp3lib_decwins + 16 - bo1;
     //printf("DEBUG: channel %d, bo %d, off %d\n", channel, bo, 16 - bo1);
-__asm__ __volatile(
+__asm__ volatile(
 ASMALIGN(4)
 ".L03:\n\t"
         "movq  (%1),%%mm0\n\t"
--- a/vidix/nvidia_vid.c	Thu Oct 16 19:24:21 2008 +0000
+++ b/vidix/nvidia_vid.c	Thu Oct 16 20:17:56 2008 +0000
@@ -372,7 +372,7 @@
  * PCI-Memory IO access macros.
  */
 
-#define MEM_BARRIER() __asm__ __volatile__ ("" : : : "memory")
+#define MEM_BARRIER() __asm__ volatile ("" : : : "memory")
 
 #undef	VID_WR08
 #define VID_WR08(p,i,val) ({ MEM_BARRIER(); ((uint8_t *)(p))[(i)]=(val); })
--- a/vidix/s3_regs.h	Thu Oct 16 19:24:21 2008 +0000
+++ b/vidix/s3_regs.h	Thu Oct 16 20:17:56 2008 +0000
@@ -196,7 +196,7 @@
 #ifndef USE_RMW_CYCLES
 
 /* Can be used to inhibit READ-MODIFY-WRITE cycles. On by default. */
-#define MEM_BARRIER() __asm__ __volatile__ ("" : : : "memory")
+#define MEM_BARRIER() __asm__ volatile ("" : : : "memory")
 
 #undef  VID_WR08
 #define VID_WR08(p,i,val) ({ MEM_BARRIER(); ((uint8_t *)(p))[(i)]=(val); })
--- a/vidix/sysdep/AsmMacros_powerpc.h	Thu Oct 16 19:24:21 2008 +0000
+++ b/vidix/sysdep/AsmMacros_powerpc.h	Thu Oct 16 20:17:56 2008 +0000
@@ -68,7 +68,7 @@
 
 static __inline__ volatile void eieio()
 {
-	__asm__ __volatile__ ("eieio");
+	__asm__ volatile ("eieio");
 }
 
 static __inline__ void outb(short port, unsigned char value)
--- a/vidix/sysdep/AsmMacros_sparc.h	Thu Oct 16 19:24:21 2008 +0000
+++ b/vidix/sysdep/AsmMacros_sparc.h	Thu Oct 16 20:17:56 2008 +0000
@@ -68,37 +68,37 @@
 
 static __inline__ void outb(unsigned long port, char val)
 {
-  __asm__ __volatile__("stba %0, [%1] %2" : : "r" (val), "r" (port), "i" (ASI_PL));
+  __asm__ volatile("stba %0, [%1] %2" : : "r" (val), "r" (port), "i" (ASI_PL));
 }
 
 static __inline__ void outw(unsigned long port, char val)
 {
-  __asm__ __volatile__("stha %0, [%1] %2" : : "r" (val), "r" (port), "i" (ASI_PL));
+  __asm__ volatile("stha %0, [%1] %2" : : "r" (val), "r" (port), "i" (ASI_PL));
 }
 
 static __inline__ void outl(unsigned long port, char val)
 {
-  __asm__ __volatile__("sta %0, [%1] %2" : : "r" (val), "r" (port), "i" (ASI_PL));
+  __asm__ volatile("sta %0, [%1] %2" : : "r" (val), "r" (port), "i" (ASI_PL));
 }
 
 static __inline__ unsigned int inb(unsigned long port)
 {
    unsigned char ret;
-   __asm__ __volatile__("lduba [%1] %2, %0" : "=r" (ret) : "r" (port), "i" (ASI_PL));
+   __asm__ volatile("lduba [%1] %2, %0" : "=r" (ret) : "r" (port), "i" (ASI_PL));
    return ret;
 }
 
 static __inline__ unsigned int inw(unsigned long port)
 {
    unsigned char ret;
-   __asm__ __volatile__("lduha [%1] %2, %0" : "=r" (ret) : "r" (port), "i" (ASI_PL));
+   __asm__ volatile("lduha [%1] %2, %0" : "=r" (ret) : "r" (port), "i" (ASI_PL));
    return ret;
 }
 
 static __inline__ unsigned int inl(unsigned long port)
 {
    unsigned char ret;
-   __asm__ __volatile__("lda [%1] %2, %0" : "=r" (ret) : "r" (port), "i" (ASI_PL));
+   __asm__ volatile("lda [%1] %2, %0" : "=r" (ret) : "r" (port), "i" (ASI_PL));
    return ret;
 }
 
--- a/vidix/sysdep/AsmMacros_x86.h	Thu Oct 16 19:24:21 2008 +0000
+++ b/vidix/sysdep/AsmMacros_x86.h	Thu Oct 16 20:17:56 2008 +0000
@@ -171,7 +171,7 @@
     }
     else
 #endif
-   __asm__ __volatile__("outb %0,%1" : :"a" (val), "d" (port));
+   __asm__ volatile("outb %0,%1" : :"a" (val), "d" (port));
     return;
 }
 
@@ -199,7 +199,7 @@
     }
     else
 #endif
-   __asm__ __volatile__("outw %0,%1" : :"a" (val), "d" (port));
+   __asm__ volatile("outw %0,%1" : :"a" (val), "d" (port));
     return;
 }
 
@@ -227,7 +227,7 @@
     }
     else
 #endif
-   __asm__ __volatile__("outl %0,%1" : :"a" (val), "d" (port));
+   __asm__ volatile("outl %0,%1" : :"a" (val), "d" (port));
     return;
 }
 
@@ -255,7 +255,7 @@
     }
     else
 #endif
-   __asm__ __volatile__("inb %1,%0" :
+   __asm__ volatile("inb %1,%0" :
        "=a" (ret) :
        "d" (port));
    return ret;
@@ -285,7 +285,7 @@
     }
     else
 #endif
-   __asm__ __volatile__("inw %1,%0" :
+   __asm__ volatile("inw %1,%0" :
        "=a" (ret) :
        "d" (port));
    return ret;
@@ -315,7 +315,7 @@
     }
     else
 #endif
-   __asm__ __volatile__("inl %1,%0" :
+   __asm__ volatile("inl %1,%0" :
        "=a" (ret) :
        "d" (port));
    return ret;
@@ -327,7 +327,7 @@
     if (svgahelper_initialized == 1)
 	return;
 #endif
-  __asm__ __volatile__("cli");
+  __asm__ volatile("cli");
 }
 
 static __inline__ void intr_enable()
@@ -336,7 +336,7 @@
     if (svgahelper_initialized == 1)
 	return;
 #endif
-  __asm__ __volatile__("sti");
+  __asm__ volatile("sti");
 }
 
 #endif /* MPLAYER_ASMMACROS_X86_H */