# HG changeset patch # User rathann # Date 1225735680 0 # Node ID 0d108ec856205d0716e933a8c41008ee39533e34 # Parent 069d7a8e2e75c5227416a56ccb692929e70af818 Remove duplicated MM_* macros for CPU capabilities from dsputil.h. Add missing one for FF_MM_ALTIVEC to avcodec.h. Rename all the occurences of MM_* to the corresponding FF_MM_*. diff -r 069d7a8e2e75 -r 0d108ec85620 armv4l/dsputil_arm.c --- a/armv4l/dsputil_arm.c Sun Nov 02 21:51:27 2008 +0000 +++ b/armv4l/dsputil_arm.c Mon Nov 03 18:08:00 2008 +0000 @@ -117,7 +117,7 @@ int mm_support(void) { - return ENABLE_IWMMXT * MM_IWMMXT; + return ENABLE_IWMMXT * FF_MM_IWMMXT; } void dsputil_init_armv4l(DSPContext* c, AVCodecContext *avctx) diff -r 069d7a8e2e75 -r 0d108ec85620 armv4l/dsputil_iwmmxt.c --- a/armv4l/dsputil_iwmmxt.c Sun Nov 02 21:51:27 2008 +0000 +++ b/armv4l/dsputil_iwmmxt.c Mon Nov 03 18:08:00 2008 +0000 @@ -150,7 +150,7 @@ /* A run time test is not simple. If this file is compiled in * then we should install the functions */ -int mm_flags = MM_IWMMXT; /* multimedia extension flags */ +int mm_flags = FF_MM_IWMMXT; /* multimedia extension flags */ void dsputil_init_iwmmxt(DSPContext* c, AVCodecContext *avctx) { @@ -161,7 +161,7 @@ mm_flags &= ~(avctx->dsp_mask & 0xffff); } - if (!(mm_flags & MM_IWMMXT)) return; + if (!(mm_flags & FF_MM_IWMMXT)) return; c->add_pixels_clamped = add_pixels_clamped_iwmmxt; diff -r 069d7a8e2e75 -r 0d108ec85620 armv4l/mpegvideo_iwmmxt.c --- a/armv4l/mpegvideo_iwmmxt.c Sun Nov 02 21:51:27 2008 +0000 +++ b/armv4l/mpegvideo_iwmmxt.c Mon Nov 03 18:08:00 2008 +0000 @@ -110,7 +110,7 @@ void MPV_common_init_iwmmxt(MpegEncContext *s) { - if (!(mm_flags & MM_IWMMXT)) return; + if (!(mm_flags & FF_MM_IWMMXT)) return; s->dct_unquantize_h263_intra = dct_unquantize_h263_intra_iwmmxt; #if 0 diff -r 069d7a8e2e75 -r 0d108ec85620 avcodec.h --- a/avcodec.h Sun Nov 02 21:51:27 2008 +0000 +++ b/avcodec.h Mon Nov 03 18:08:00 2008 +0000 @@ -1431,6 +1431,7 @@ #define FF_MM_SSE3 0x0040 ///< Prescott SSE3 functions #define FF_MM_SSSE3 0x0080 ///< Conroe SSSE3 functions #define FF_MM_IWMMXT 0x0100 ///< XScale IWMMXT +#define FF_MM_ALTIVEC 0x0001 ///< standard AltiVec /** * bits per sample/pixel from the demuxer (needed for huffyuv). diff -r 069d7a8e2e75 -r 0d108ec85620 dct-test.c --- a/dct-test.c Sun Nov 02 21:51:27 2008 +0000 +++ b/dct-test.c Mon Nov 03 18:08:00 2008 +0000 @@ -91,23 +91,23 @@ {"SIMPLE-C", 1, ff_simple_idct, idct, NO_PERM}, #ifdef HAVE_MMX - {"MMX", 0, ff_fdct_mmx, fdct, NO_PERM, MM_MMX}, + {"MMX", 0, ff_fdct_mmx, fdct, NO_PERM, FF_MM_MMX}, #ifdef HAVE_MMX2 - {"MMX2", 0, ff_fdct_mmx2, fdct, NO_PERM, MM_MMXEXT}, + {"MMX2", 0, ff_fdct_mmx2, fdct, NO_PERM, FF_MM_MMXEXT}, #endif #ifdef CONFIG_GPL - {"LIBMPEG2-MMX", 1, ff_mmx_idct, idct, MMX_PERM, MM_MMX}, - {"LIBMPEG2-MMXEXT", 1, ff_mmxext_idct, idct, MMX_PERM, MM_MMXEXT}, + {"LIBMPEG2-MMX", 1, ff_mmx_idct, idct, MMX_PERM, FF_MM_MMX}, + {"LIBMPEG2-MMXEXT", 1, ff_mmxext_idct, idct, MMX_PERM, FF_MM_MMXEXT}, #endif - {"SIMPLE-MMX", 1, ff_simple_idct_mmx, idct, MMX_SIMPLE_PERM, MM_MMX}, - {"XVID-MMX", 1, ff_idct_xvid_mmx, idct, NO_PERM, MM_MMX}, - {"XVID-MMX2", 1, ff_idct_xvid_mmx2, idct, NO_PERM, MM_MMXEXT}, - {"XVID-SSE2", 1, ff_idct_xvid_sse2, idct, SSE2_PERM, MM_SSE2}, + {"SIMPLE-MMX", 1, ff_simple_idct_mmx, idct, MMX_SIMPLE_PERM, FF_MM_MMX}, + {"XVID-MMX", 1, ff_idct_xvid_mmx, idct, NO_PERM, FF_MM_MMX}, + {"XVID-MMX2", 1, ff_idct_xvid_mmx2, idct, NO_PERM, FF_MM_MMXEXT}, + {"XVID-SSE2", 1, ff_idct_xvid_sse2, idct, SSE2_PERM, FF_MM_SSE2}, #endif #ifdef HAVE_ALTIVEC - {"altivecfdct", 0, fdct_altivec, fdct, NO_PERM, MM_ALTIVEC}, + {"altivecfdct", 0, fdct_altivec, fdct, NO_PERM, FF_MM_ALTIVEC}, #endif #ifdef ARCH_BFIN @@ -176,7 +176,7 @@ static inline void mmx_emms(void) { #ifdef HAVE_MMX - if (cpu_flags & MM_MMX) + if (cpu_flags & FF_MM_MMX) __asm__ volatile ("emms\n\t"); #endif } diff -r 069d7a8e2e75 -r 0d108ec85620 dsputil.h --- a/dsputil.h Sun Nov 02 21:51:27 2008 +0000 +++ b/dsputil.h Mon Nov 03 18:08:00 2008 +0000 @@ -562,15 +562,6 @@ #undef emms_c -#define MM_MMX 0x0001 /* standard MMX */ -#define MM_3DNOW 0x0004 /* AMD 3DNOW */ -#define MM_MMXEXT 0x0002 /* SSE integer functions or AMD MMX ext */ -#define MM_SSE 0x0008 /* SSE functions */ -#define MM_SSE2 0x0010 /* PIV SSE2 functions */ -#define MM_3DNOWEXT 0x0020 /* AMD 3DNowExt */ -#define MM_SSE3 0x0040 /* Prescott SSE3 functions */ -#define MM_SSSE3 0x0080 /* Conroe SSSE3 functions */ - extern int mm_flags; void add_pixels_clamped_mmx(const DCTELEM *block, uint8_t *pixels, int line_size); @@ -585,7 +576,7 @@ #define emms_c() \ {\ - if (mm_flags & MM_MMX)\ + if (mm_flags & FF_MM_MMX)\ emms();\ } @@ -593,8 +584,6 @@ #elif defined(ARCH_ARMV4L) -#define MM_IWMMXT 0x0100 /* XScale IWMMXT */ - extern int mm_flags; #ifdef HAVE_NEON @@ -604,8 +593,6 @@ #elif defined(ARCH_POWERPC) -#define MM_ALTIVEC 0x0001 /* standard AltiVec */ - extern int mm_flags; #define DECLARE_ALIGNED_8(t, v) DECLARE_ALIGNED(16, t, v) diff -r 069d7a8e2e75 -r 0d108ec85620 fft.c --- a/fft.c Sun Nov 02 21:51:27 2008 +0000 +++ b/fft.c Mon Nov 03 18:08:00 2008 +0000 @@ -93,18 +93,18 @@ #if defined HAVE_MMX && defined HAVE_YASM has_vectors = mm_support(); - if (has_vectors & MM_SSE) { + if (has_vectors & FF_MM_SSE) { /* SSE for P3/P4/K8 */ s->imdct_calc = ff_imdct_calc_sse; s->imdct_half = ff_imdct_half_sse; s->fft_permute = ff_fft_permute_sse; s->fft_calc = ff_fft_calc_sse; - } else if (has_vectors & MM_3DNOWEXT) { + } else if (has_vectors & FF_MM_3DNOWEXT) { /* 3DNowEx for K7 */ s->imdct_calc = ff_imdct_calc_3dn2; s->imdct_half = ff_imdct_half_3dn2; s->fft_calc = ff_fft_calc_3dn2; - } else if (has_vectors & MM_3DNOW) { + } else if (has_vectors & FF_MM_3DNOW) { /* 3DNow! for K6-2/3 */ s->imdct_calc = ff_imdct_calc_3dn; s->imdct_half = ff_imdct_half_3dn; @@ -112,7 +112,7 @@ } #elif defined HAVE_ALTIVEC && !defined ALTIVEC_USE_REFERENCE_C_CODE has_vectors = mm_support(); - if (has_vectors & MM_ALTIVEC) { + if (has_vectors & FF_MM_ALTIVEC) { s->fft_calc = ff_fft_calc_altivec; split_radix = 0; } diff -r 069d7a8e2e75 -r 0d108ec85620 h263dec.c --- a/h263dec.c Sun Nov 02 21:51:27 2008 +0000 +++ b/h263dec.c Mon Nov 03 18:08:00 2008 +0000 @@ -549,7 +549,7 @@ #endif #if defined(HAVE_MMX) - if(s->codec_id == CODEC_ID_MPEG4 && s->xvid_build && avctx->idct_algo == FF_IDCT_AUTO && (mm_flags & MM_MMX)){ + if(s->codec_id == CODEC_ID_MPEG4 && s->xvid_build && avctx->idct_algo == FF_IDCT_AUTO && (mm_flags & FF_MM_MMX)){ avctx->idct_algo= FF_IDCT_XVIDMMX; avctx->coded_width= 0; // force reinit // dsputil_init(&s->dsp, avctx); diff -r 069d7a8e2e75 -r 0d108ec85620 i386/dsputil_mmx.c --- a/i386/dsputil_mmx.c Sun Nov 02 21:51:27 2008 +0000 +++ b/i386/dsputil_mmx.c Mon Nov 03 18:08:00 2008 +0000 @@ -2498,20 +2498,20 @@ #if 0 av_log(avctx, AV_LOG_INFO, "libavcodec: CPU flags:"); - if (mm_flags & MM_MMX) + if (mm_flags & FF_MM_MMX) av_log(avctx, AV_LOG_INFO, " mmx"); - if (mm_flags & MM_MMXEXT) + if (mm_flags & FF_MM_MMXEXT) av_log(avctx, AV_LOG_INFO, " mmxext"); - if (mm_flags & MM_3DNOW) + if (mm_flags & FF_MM_3DNOW) av_log(avctx, AV_LOG_INFO, " 3dnow"); - if (mm_flags & MM_SSE) + if (mm_flags & FF_MM_SSE) av_log(avctx, AV_LOG_INFO, " sse"); - if (mm_flags & MM_SSE2) + if (mm_flags & FF_MM_SSE2) av_log(avctx, AV_LOG_INFO, " sse2"); av_log(avctx, AV_LOG_INFO, "\n"); #endif - if (mm_flags & MM_MMX) { + if (mm_flags & FF_MM_MMX) { const int idct_algo= avctx->idct_algo; if(avctx->lowres==0){ @@ -2522,7 +2522,7 @@ c->idct_permutation_type= FF_SIMPLE_IDCT_PERM; #ifdef CONFIG_GPL }else if(idct_algo==FF_IDCT_LIBMPEG2MMX){ - if(mm_flags & MM_MMXEXT){ + if(mm_flags & FF_MM_MMXEXT){ c->idct_put= ff_libmpeg2mmx2_idct_put; c->idct_add= ff_libmpeg2mmx2_idct_add; c->idct = ff_mmxext_idct; @@ -2535,7 +2535,7 @@ #endif }else if((ENABLE_VP3_DECODER || ENABLE_VP5_DECODER || ENABLE_VP6_DECODER || ENABLE_THEORA_DECODER) && idct_algo==FF_IDCT_VP3){ - if(mm_flags & MM_SSE2){ + if(mm_flags & FF_MM_SSE2){ c->idct_put= ff_vp3_idct_put_sse2; c->idct_add= ff_vp3_idct_add_sse2; c->idct = ff_vp3_idct_sse2; @@ -2549,12 +2549,12 @@ }else if(idct_algo==FF_IDCT_CAVS){ c->idct_permutation_type= FF_TRANSPOSE_IDCT_PERM; }else if(idct_algo==FF_IDCT_XVIDMMX){ - if(mm_flags & MM_SSE2){ + if(mm_flags & FF_MM_SSE2){ c->idct_put= ff_idct_xvid_sse2_put; c->idct_add= ff_idct_xvid_sse2_add; c->idct = ff_idct_xvid_sse2; c->idct_permutation_type= FF_SSE2_IDCT_PERM; - }else if(mm_flags & MM_MMXEXT){ + }else if(mm_flags & FF_MM_MMXEXT){ c->idct_put= ff_idct_xvid_mmx2_put; c->idct_add= ff_idct_xvid_mmx2_add; c->idct = ff_idct_xvid_mmx2; @@ -2605,10 +2605,10 @@ c->h264_idct_add= ff_h264_idct_add_mmx; c->h264_idct8_dc_add= c->h264_idct8_add= ff_h264_idct8_add_mmx; - if (mm_flags & MM_SSE2) + if (mm_flags & FF_MM_SSE2) c->h264_idct8_add= ff_h264_idct8_add_sse2; - if (mm_flags & MM_MMXEXT) { + if (mm_flags & FF_MM_MMXEXT) { c->prefetch = prefetch_mmx2; c->put_pixels_tab[0][1] = put_pixels16_x2_mmx2; @@ -2716,7 +2716,7 @@ ff_vc1dsp_init_mmx(c, avctx); c->add_png_paeth_prediction= add_png_paeth_prediction_mmx2; - } else if (mm_flags & MM_3DNOW) { + } else if (mm_flags & FF_MM_3DNOW) { c->prefetch = prefetch_3dnow; c->put_pixels_tab[0][1] = put_pixels16_x2_3dnow; @@ -2774,7 +2774,7 @@ c->put_h264_qpel_pixels_tab[1][x+y*4] = put_h264_qpel8_mc##x##y##_##CPU;\ c->avg_h264_qpel_pixels_tab[0][x+y*4] = avg_h264_qpel16_mc##x##y##_##CPU;\ c->avg_h264_qpel_pixels_tab[1][x+y*4] = avg_h264_qpel8_mc##x##y##_##CPU; - if((mm_flags & MM_SSE2) && !(mm_flags & MM_3DNOW)){ + if((mm_flags & FF_MM_SSE2) && !(mm_flags & FF_MM_3DNOW)){ // these functions are slower than mmx on AMD, but faster on Intel /* FIXME works in most codecs, but crashes svq1 due to unaligned chroma c->put_pixels_tab[0][0] = put_pixels16_sse2; @@ -2782,7 +2782,7 @@ */ H264_QPEL_FUNCS(0, 0, sse2); } - if(mm_flags & MM_SSE2){ + if(mm_flags & FF_MM_SSE2){ H264_QPEL_FUNCS(0, 1, sse2); H264_QPEL_FUNCS(0, 2, sse2); H264_QPEL_FUNCS(0, 3, sse2); @@ -2797,7 +2797,7 @@ H264_QPEL_FUNCS(3, 3, sse2); } #ifdef HAVE_SSSE3 - if(mm_flags & MM_SSSE3){ + if(mm_flags & FF_MM_SSSE3){ H264_QPEL_FUNCS(1, 0, ssse3); H264_QPEL_FUNCS(1, 1, ssse3); H264_QPEL_FUNCS(1, 2, ssse3); @@ -2820,7 +2820,7 @@ #endif #ifdef CONFIG_SNOW_DECODER - if(mm_flags & MM_SSE2 & 0){ + if(mm_flags & FF_MM_SSE2 & 0){ c->horizontal_compose97i = ff_snow_horizontal_compose97i_sse2; #ifdef HAVE_7REGS c->vertical_compose97i = ff_snow_vertical_compose97i_sse2; @@ -2828,7 +2828,7 @@ c->inner_add_yblock = ff_snow_inner_add_yblock_sse2; } else{ - if(mm_flags & MM_MMXEXT){ + if(mm_flags & FF_MM_MMXEXT){ c->horizontal_compose97i = ff_snow_horizontal_compose97i_mmx; #ifdef HAVE_7REGS c->vertical_compose97i = ff_snow_vertical_compose97i_mmx; @@ -2838,7 +2838,7 @@ } #endif - if(mm_flags & MM_3DNOW){ + if(mm_flags & FF_MM_3DNOW){ c->vorbis_inverse_coupling = vorbis_inverse_coupling_3dnow; c->vector_fmul = vector_fmul_3dnow; if(!(avctx->flags & CODEC_FLAG_BITEXACT)){ @@ -2846,14 +2846,14 @@ c->float_to_int16_interleave = float_to_int16_interleave_3dnow; } } - if(mm_flags & MM_3DNOWEXT){ + if(mm_flags & FF_MM_3DNOWEXT){ c->vector_fmul_reverse = vector_fmul_reverse_3dnow2; c->vector_fmul_window = vector_fmul_window_3dnow2; if(!(avctx->flags & CODEC_FLAG_BITEXACT)){ c->float_to_int16_interleave = float_to_int16_interleave_3dn2; } } - if(mm_flags & MM_SSE){ + if(mm_flags & FF_MM_SSE){ c->vorbis_inverse_coupling = vorbis_inverse_coupling_sse; c->ac3_downmix = ac3_downmix_sse; c->vector_fmul = vector_fmul_sse; @@ -2864,9 +2864,9 @@ c->float_to_int16 = float_to_int16_sse; c->float_to_int16_interleave = float_to_int16_interleave_sse; } - if(mm_flags & MM_3DNOW) + if(mm_flags & FF_MM_3DNOW) c->vector_fmul_add_add = vector_fmul_add_add_3dnow; // faster than sse - if(mm_flags & MM_SSE2){ + if(mm_flags & FF_MM_SSE2){ c->int32_to_float_fmul_scalar = int32_to_float_fmul_scalar_sse2; c->float_to_int16 = float_to_int16_sse2; c->float_to_int16_interleave = float_to_int16_interleave_sse2; diff -r 069d7a8e2e75 -r 0d108ec85620 i386/dsputilenc_mmx.c --- a/i386/dsputilenc_mmx.c Sun Nov 02 21:51:27 2008 +0000 +++ b/i386/dsputilenc_mmx.c Mon Nov 03 18:08:00 2008 +0000 @@ -1354,12 +1354,12 @@ void dsputilenc_init_mmx(DSPContext* c, AVCodecContext *avctx) { - if (mm_flags & MM_MMX) { + if (mm_flags & FF_MM_MMX) { const int dct_algo = avctx->dct_algo; if(dct_algo==FF_DCT_AUTO || dct_algo==FF_DCT_MMX){ - if(mm_flags & MM_SSE2){ + if(mm_flags & FF_MM_SSE2){ c->fdct = ff_fdct_sse2; - }else if(mm_flags & MM_MMXEXT){ + }else if(mm_flags & FF_MM_MMXEXT){ c->fdct = ff_fdct_mmx2; }else{ c->fdct = ff_fdct_mmx; @@ -1377,7 +1377,7 @@ c->hadamard8_diff[1]= hadamard8_diff_mmx; c->pix_norm1 = pix_norm1_mmx; - c->sse[0] = (mm_flags & MM_SSE2) ? sse16_sse2 : sse16_mmx; + c->sse[0] = (mm_flags & FF_MM_SSE2) ? sse16_sse2 : sse16_mmx; c->sse[1] = sse8_mmx; c->vsad[4]= vsad_intra16_mmx; @@ -1395,7 +1395,7 @@ c->ssd_int8_vs_int16 = ssd_int8_vs_int16_mmx; - if (mm_flags & MM_MMXEXT) { + if (mm_flags & FF_MM_MMXEXT) { c->sum_abs_dctelem= sum_abs_dctelem_mmx2; c->hadamard8_diff[0]= hadamard8_diff16_mmx2; c->hadamard8_diff[1]= hadamard8_diff_mmx2; @@ -1408,7 +1408,7 @@ c->sub_hfyu_median_prediction= sub_hfyu_median_prediction_mmx2; } - if(mm_flags & MM_SSE2){ + if(mm_flags & FF_MM_SSE2){ c->get_pixels = get_pixels_sse2; c->sum_abs_dctelem= sum_abs_dctelem_sse2; c->hadamard8_diff[0]= hadamard8_diff16_sse2; @@ -1418,7 +1418,7 @@ } #ifdef HAVE_SSSE3 - if(mm_flags & MM_SSSE3){ + if(mm_flags & FF_MM_SSSE3){ if(!(avctx->flags & CODEC_FLAG_BITEXACT)){ c->try_8x8basis= try_8x8basis_ssse3; } @@ -1429,7 +1429,7 @@ } #endif - if(mm_flags & MM_3DNOW){ + if(mm_flags & FF_MM_3DNOW){ if(!(avctx->flags & CODEC_FLAG_BITEXACT)){ c->try_8x8basis= try_8x8basis_3dnow; } diff -r 069d7a8e2e75 -r 0d108ec85620 i386/motion_est_mmx.c --- a/i386/motion_est_mmx.c Sun Nov 02 21:51:27 2008 +0000 +++ b/i386/motion_est_mmx.c Mon Nov 03 18:08:00 2008 +0000 @@ -426,7 +426,7 @@ void dsputil_init_pix_mmx(DSPContext* c, AVCodecContext *avctx) { - if (mm_flags & MM_MMX) { + if (mm_flags & FF_MM_MMX) { c->pix_abs[0][0] = sad16_mmx; c->pix_abs[0][1] = sad16_x2_mmx; c->pix_abs[0][2] = sad16_y2_mmx; @@ -439,7 +439,7 @@ c->sad[0]= sad16_mmx; c->sad[1]= sad8_mmx; } - if (mm_flags & MM_MMXEXT) { + if (mm_flags & FF_MM_MMXEXT) { c->pix_abs[0][0] = sad16_mmx2; c->pix_abs[1][0] = sad8_mmx2; @@ -455,7 +455,7 @@ c->pix_abs[1][3] = sad8_xy2_mmx2; } } - if ((mm_flags & MM_SSE2) && !(mm_flags & MM_3DNOW)) { + if ((mm_flags & FF_MM_SSE2) && !(mm_flags & FF_MM_3DNOW)) { c->sad[0]= sad16_sse2; } } diff -r 069d7a8e2e75 -r 0d108ec85620 i386/mpegvideo_mmx.c --- a/i386/mpegvideo_mmx.c Sun Nov 02 21:51:27 2008 +0000 +++ b/i386/mpegvideo_mmx.c Mon Nov 03 18:08:00 2008 +0000 @@ -619,7 +619,7 @@ void MPV_common_init_mmx(MpegEncContext *s) { - if (mm_flags & MM_MMX) { + if (mm_flags & FF_MM_MMX) { const int dct_algo = s->avctx->dct_algo; s->dct_unquantize_h263_intra = dct_unquantize_h263_intra_mmx; @@ -630,7 +630,7 @@ s->dct_unquantize_mpeg2_intra = dct_unquantize_mpeg2_intra_mmx; s->dct_unquantize_mpeg2_inter = dct_unquantize_mpeg2_inter_mmx; - if (mm_flags & MM_SSE2) { + if (mm_flags & FF_MM_SSE2) { s->denoise_dct= denoise_dct_sse2; } else { s->denoise_dct= denoise_dct_mmx; @@ -638,13 +638,13 @@ if(dct_algo==FF_DCT_AUTO || dct_algo==FF_DCT_MMX){ #ifdef HAVE_SSSE3 - if(mm_flags & MM_SSSE3){ + if(mm_flags & FF_MM_SSSE3){ s->dct_quantize= dct_quantize_SSSE3; } else #endif - if(mm_flags & MM_SSE2){ + if(mm_flags & FF_MM_SSE2){ s->dct_quantize= dct_quantize_SSE2; - } else if(mm_flags & MM_MMXEXT){ + } else if(mm_flags & FF_MM_MMXEXT){ s->dct_quantize= dct_quantize_MMX2; } else { s->dct_quantize= dct_quantize_MMX; diff -r 069d7a8e2e75 -r 0d108ec85620 imgresample.c --- a/imgresample.c Sun Nov 02 21:51:27 2008 +0000 +++ b/imgresample.c Mon Nov 03 18:08:00 2008 +0000 @@ -343,7 +343,7 @@ n = dst_width; } #ifdef HAVE_MMX - if ((mm_flags & MM_MMX) && NB_TAPS == 4) + if ((mm_flags & FF_MM_MMX) && NB_TAPS == 4) h_resample_fast4_mmx(dst, n, src, src_width, src_start, src_incr, filters); else @@ -401,14 +401,14 @@ phase_y = get_phase(src_y); #ifdef HAVE_MMX /* desactivated MMX because loss of precision */ - if ((mm_flags & MM_MMX) && NB_TAPS == 4 && 0) + if ((mm_flags & FF_MM_MMX) && NB_TAPS == 4 && 0) v_resample4_mmx(output, owidth, s->line_buf + (ring_y - NB_TAPS + 1) * owidth, owidth, &s->v_filters[phase_y][0]); else #endif #ifdef HAVE_ALTIVEC - if ((mm_flags & MM_ALTIVEC) && NB_TAPS == 4 && FILTER_BITS <= 6) + if ((mm_flags & FF_MM_ALTIVEC) && NB_TAPS == 4 && FILTER_BITS <= 6) v_resample16_altivec(output, owidth, s->line_buf + (ring_y - NB_TAPS + 1) * owidth, owidth, &s->v_filters[phase_y][0]); @@ -811,7 +811,7 @@ fact = 0.72; xsize = (int)(XSIZE * fact); ysize = (int)(YSIZE * fact); - mm_flags = MM_MMX; + mm_flags = FF_MM_MMX; s = img_resample_init(xsize, ysize, XSIZE, YSIZE); component_resample(s, img1, xsize, xsize, ysize, img, XSIZE, XSIZE, YSIZE); diff -r 069d7a8e2e75 -r 0d108ec85620 ppc/dsputil_ppc.c --- a/ppc/dsputil_ppc.c Sun Nov 02 21:51:27 2008 +0000 +++ b/ppc/dsputil_ppc.c Mon Nov 03 18:08:00 2008 +0000 @@ -50,7 +50,7 @@ int result = 0; #ifdef HAVE_ALTIVEC if (has_altivec()) { - result |= MM_ALTIVEC; + result |= FF_MM_ALTIVEC; } #endif /* result */ return result; @@ -265,7 +265,7 @@ if(ENABLE_H264_DECODER) dsputil_h264_init_ppc(c, avctx); if (has_altivec()) { - mm_flags |= MM_ALTIVEC; + mm_flags |= FF_MM_ALTIVEC; dsputil_init_altivec(c, avctx); if(ENABLE_SNOW_DECODER) snow_init_altivec(c, avctx); diff -r 069d7a8e2e75 -r 0d108ec85620 ppc/mpegvideo_altivec.c --- a/ppc/mpegvideo_altivec.c Sun Nov 02 21:51:27 2008 +0000 +++ b/ppc/mpegvideo_altivec.c Mon Nov 03 18:08:00 2008 +0000 @@ -590,7 +590,7 @@ void MPV_common_init_altivec(MpegEncContext *s) { - if ((mm_flags & MM_ALTIVEC) == 0) return; + if ((mm_flags & FF_MM_ALTIVEC) == 0) return; if (s->avctx->lowres==0) { if ((s->avctx->idct_algo == FF_IDCT_AUTO) ||