Mercurial > libavcodec.hg
changeset 4196:fbac0859753d libavcodec
move some CFLAGS settings away from config.* writing section
author | mru |
---|---|
date | Tue, 14 Nov 2006 03:12:29 +0000 |
parents | 6b784f65ce16 |
children | bbe0bc387a19 |
files | ac3.h ac3enc.c armv4l/dsputil_iwmmxt.c armv4l/mpegvideo_iwmmxt.c dct-test.c dsputil.c dsputil.h h263dec.c i386/dsputil_mmx.c i386/motion_est_mmx.c i386/mpegvideo_mmx.c i386/mpegvideo_mmx_template.c imgresample.c ppc/dsputil_ppc.c snow.c snow.h utils.c |
diffstat | 17 files changed, 64 insertions(+), 73 deletions(-) [+] |
line wrap: on
line diff
--- a/ac3.h Tue Nov 14 00:56:56 2006 +0000 +++ b/ac3.h Tue Nov 14 03:12:29 2006 +0000 @@ -57,9 +57,3 @@ extern const uint16_t fgaintab[8]; #endif -void ac3_common_init(void); -void ac3_parametric_bit_allocation(AC3BitAllocParameters *s, uint8_t *bap, - int8_t *exp, int start, int end, - int snroffset, int fgain, int is_lfe, - int deltbae,int deltnseg, - uint8_t *deltoffst, uint8_t *deltlen, uint8_t *deltba);
--- a/ac3enc.c Tue Nov 14 00:56:56 2006 +0000 +++ b/ac3enc.c Tue Nov 14 03:12:29 2006 +0000 @@ -119,6 +119,7 @@ /* AC3 bit allocation. The algorithm is the one described in the AC3 spec. */ +static void ac3_parametric_bit_allocation(AC3BitAllocParameters *s, uint8_t *bap, int8_t *exp, int start, int end, int snroffset, int fgain, int is_lfe, @@ -794,7 +795,7 @@ return 0; } -void ac3_common_init(void) +static void ac3_common_init(void) { int i, j, k, l, v; /* compute bndtab and masktab from bandsz */
--- a/armv4l/dsputil_iwmmxt.c Tue Nov 14 00:56:56 2006 +0000 +++ b/armv4l/dsputil_iwmmxt.c Tue Nov 14 03:12:29 2006 +0000 @@ -128,7 +128,7 @@ return; } -int mm_flags; /* multimedia extension flags */ +int ff_mm_flags; /* multimedia extension flags */ int mm_support(void) { @@ -137,16 +137,16 @@ void dsputil_init_iwmmxt(DSPContext* c, AVCodecContext *avctx) { - mm_flags = mm_support(); + ff_mm_flags = mm_support(); if (avctx->dsp_mask) { if (avctx->dsp_mask & FF_MM_FORCE) - mm_flags |= (avctx->dsp_mask & 0xffff); + ff_mm_flags |= (avctx->dsp_mask & 0xffff); else - mm_flags &= ~(avctx->dsp_mask & 0xffff); + ff_mm_flags &= ~(avctx->dsp_mask & 0xffff); } - if (!(mm_flags & MM_IWMMXT)) return; + if (!(ff_mm_flags & MM_IWMMXT)) return; c->add_pixels_clamped = add_pixels_clamped_iwmmxt;
--- a/armv4l/mpegvideo_iwmmxt.c Tue Nov 14 00:56:56 2006 +0000 +++ b/armv4l/mpegvideo_iwmmxt.c Tue Nov 14 03:12:29 2006 +0000 @@ -110,7 +110,7 @@ void MPV_common_init_iwmmxt(MpegEncContext *s) { - if (!(mm_flags & MM_IWMMXT)) return; + if (!(ff_mm_flags & MM_IWMMXT)) return; s->dct_unquantize_h263_intra = dct_unquantize_h263_intra_iwmmxt; #if 0
--- a/dct-test.c Tue Nov 14 00:56:56 2006 +0000 +++ b/dct-test.c Tue Nov 14 03:12:29 2006 +0000 @@ -69,7 +69,7 @@ 4520, 6270, 5906, 5315, 4520, 3552, 2446, 1247 }; -uint8_t cropTbl[256 + 2 * MAX_NEG_CROP]; +uint8_t ff_cropTbl[256 + 2 * MAX_NEG_CROP]; int64_t gettime(void) { @@ -473,10 +473,10 @@ init_fdct(); idct_mmx_init(); - for(i=0;i<256;i++) cropTbl[i + MAX_NEG_CROP] = i; + for(i=0;i<256;i++) ff_cropTbl[i + MAX_NEG_CROP] = i; for(i=0;i<MAX_NEG_CROP;i++) { - cropTbl[i] = 0; - cropTbl[i + MAX_NEG_CROP + 256] = 255; + ff_cropTbl[i] = 0; + ff_cropTbl[i + MAX_NEG_CROP + 256] = 255; } for(;;) {
--- a/dsputil.c Tue Nov 14 00:56:56 2006 +0000 +++ b/dsputil.c Tue Nov 14 03:12:29 2006 +0000 @@ -68,7 +68,7 @@ }; /* not permutated inverse zigzag_direct + 1 for MMX quantizer */ -DECLARE_ALIGNED_8(uint16_t, inv_zigzag_direct16[64]) = {0, }; +DECLARE_ALIGNED_8(uint16_t, ff_inv_zigzag_direct16[64]) = {0, }; const uint8_t ff_alternate_horizontal_scan[64] = { 0, 1, 2, 3, 8, 9, 16, 17, @@ -383,11 +383,11 @@ return w_c(v, pix1, pix2, line_size, 16, h, 0); } -int w53_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h){ +int ff_w53_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h){ return w_c(v, pix1, pix2, line_size, 32, h, 1); } -int w97_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h){ +int ff_w97_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h){ return w_c(v, pix1, pix2, line_size, 32, h, 0); } #endif @@ -3861,7 +3861,7 @@ static void just_return() { return; } /* init static data */ -void dsputil_static_init(void) +void ff_dsputil_static_init(void) { int i; @@ -3875,7 +3875,7 @@ ff_squareTbl[i] = (i - 256) * (i - 256); } - for(i=0; i<64; i++) inv_zigzag_direct16[ff_zigzag_direct[i]]= i+1; + for(i=0; i<64; i++) ff_inv_zigzag_direct16[ff_zigzag_direct[i]]= i+1; }
--- a/dsputil.h Tue Nov 14 00:56:56 2006 +0000 +++ b/dsputil.h Tue Nov 14 03:12:29 2006 +0000 @@ -405,7 +405,7 @@ op_pixels_func put_vc1_mspel_pixels_tab[16]; } DSPContext; -void dsputil_static_init(void); +void ff_dsputil_static_init(void); void dsputil_init(DSPContext* p, AVCodecContext *avctx); /** @@ -463,6 +463,8 @@ one or more MultiMedia extension */ int mm_support(void); +extern int ff_mm_flags; + #ifdef __GNUC__ #define DECLARE_ALIGNED_16(t,v) t v __attribute__ ((aligned (16))) #else @@ -481,8 +483,6 @@ #define MM_3DNOWEXT 0x0020 /* AMD 3DNowExt */ #define MM_SSE3 0x0040 /* Prescott SSE3 functions */ -extern int mm_flags; - void add_pixels_clamped_mmx(const DCTELEM *block, uint8_t *pixels, int line_size); void put_pixels_clamped_mmx(const DCTELEM *block, uint8_t *pixels, int line_size); void put_signed_pixels_clamped_mmx(const DCTELEM *block, uint8_t *pixels, int line_size); @@ -495,7 +495,7 @@ #define emms_c() \ {\ - if (mm_flags & MM_MMX)\ + if (ff_mm_flags & MM_MMX)\ emms();\ } @@ -519,8 +519,6 @@ #define MM_IWMMXT 0x0100 /* XScale IWMMXT */ -extern int mm_flags; - void dsputil_init_armv4l(DSPContext* c, AVCodecContext *avctx); #elif defined(HAVE_MLIB) @@ -549,8 +547,6 @@ #define MM_ALTIVEC 0x0001 /* standard AltiVec */ -extern int mm_flags; - #if defined(HAVE_ALTIVEC) && !defined(CONFIG_DARWIN) #define pixel altivec_pixel #include <altivec.h>
--- a/h263dec.c Tue Nov 14 00:56:56 2006 +0000 +++ b/h263dec.c Tue Nov 14 03:12:29 2006 +0000 @@ -643,7 +643,7 @@ #endif #if defined(HAVE_MMX) && defined(CONFIG_GPL) - if(s->codec_id == CODEC_ID_MPEG4 && s->xvid_build && avctx->idct_algo == FF_IDCT_AUTO && (mm_flags & MM_MMX)){ + if(s->codec_id == CODEC_ID_MPEG4 && s->xvid_build && avctx->idct_algo == FF_IDCT_AUTO && (ff_mm_flags & MM_MMX)){ avctx->idct_algo= FF_IDCT_XVIDMMX; avctx->coded_width= 0; // force reinit // dsputil_init(&s->dsp, avctx);
--- a/i386/dsputil_mmx.c Tue Nov 14 00:56:56 2006 +0000 +++ b/i386/dsputil_mmx.c Tue Nov 14 03:12:29 2006 +0000 @@ -34,7 +34,7 @@ extern void ff_idct_xvid_mmx(short *block); extern void ff_idct_xvid_mmx2(short *block); -int mm_flags; /* multimedia extension flags */ +int ff_mm_flags; /* multimedia extension flags */ /* pixel operations */ static const uint64_t mm_bone attribute_used __attribute__ ((aligned(8))) = 0x0101010101010101ULL; @@ -3050,39 +3050,39 @@ void dsputil_init_mmx(DSPContext* c, AVCodecContext *avctx) { - mm_flags = mm_support(); + ff_mm_flags = mm_support(); if (avctx->dsp_mask) { if (avctx->dsp_mask & FF_MM_FORCE) - mm_flags |= (avctx->dsp_mask & 0xffff); + ff_mm_flags |= (avctx->dsp_mask & 0xffff); else - mm_flags &= ~(avctx->dsp_mask & 0xffff); + ff_mm_flags &= ~(avctx->dsp_mask & 0xffff); } #if 0 av_log(avctx, AV_LOG_INFO, "libavcodec: CPU flags:"); - if (mm_flags & MM_MMX) + if (ff_mm_flags & MM_MMX) av_log(avctx, AV_LOG_INFO, " mmx"); - if (mm_flags & MM_MMXEXT) + if (ff_mm_flags & MM_MMXEXT) av_log(avctx, AV_LOG_INFO, " mmxext"); - if (mm_flags & MM_3DNOW) + if (ff_mm_flags & MM_3DNOW) av_log(avctx, AV_LOG_INFO, " 3dnow"); - if (mm_flags & MM_SSE) + if (ff_mm_flags & MM_SSE) av_log(avctx, AV_LOG_INFO, " sse"); - if (mm_flags & MM_SSE2) + if (ff_mm_flags & MM_SSE2) av_log(avctx, AV_LOG_INFO, " sse2"); av_log(avctx, AV_LOG_INFO, "\n"); #endif - if (mm_flags & MM_MMX) { + if (ff_mm_flags & MM_MMX) { const int idct_algo= avctx->idct_algo; #ifdef CONFIG_ENCODERS const int dct_algo = avctx->dct_algo; if(dct_algo==FF_DCT_AUTO || dct_algo==FF_DCT_MMX){ - if(mm_flags & MM_SSE2){ + if(ff_mm_flags & MM_SSE2){ c->fdct = ff_fdct_sse2; - }else if(mm_flags & MM_MMXEXT){ + }else if(ff_mm_flags & MM_MMXEXT){ c->fdct = ff_fdct_mmx2; }else{ c->fdct = ff_fdct_mmx; @@ -3097,7 +3097,7 @@ c->idct_permutation_type= FF_SIMPLE_IDCT_PERM; #ifdef CONFIG_GPL }else if(idct_algo==FF_IDCT_LIBMPEG2MMX){ - if(mm_flags & MM_MMXEXT){ + if(ff_mm_flags & MM_MMXEXT){ c->idct_put= ff_libmpeg2mmx2_idct_put; c->idct_add= ff_libmpeg2mmx2_idct_add; c->idct = ff_mmxext_idct; @@ -3111,7 +3111,7 @@ }else if(idct_algo==FF_IDCT_VP3 && avctx->codec->id!=CODEC_ID_THEORA && !(avctx->flags & CODEC_FLAG_BITEXACT)){ - if(mm_flags & MM_SSE2){ + if(ff_mm_flags & MM_SSE2){ c->idct_put= ff_vp3_idct_put_sse2; c->idct_add= ff_vp3_idct_add_sse2; c->idct = ff_vp3_idct_sse2; @@ -3126,7 +3126,7 @@ }else if(idct_algo==FF_IDCT_CAVS){ c->idct_permutation_type= FF_TRANSPOSE_IDCT_PERM; }else if(idct_algo==FF_IDCT_XVIDMMX){ - if(mm_flags & MM_MMXEXT){ + if(ff_mm_flags & MM_MMXEXT){ c->idct_put= ff_idct_xvid_mmx2_put; c->idct_add= ff_idct_xvid_mmx2_add; c->idct = ff_idct_xvid_mmx2; @@ -3200,7 +3200,7 @@ c->hadamard8_diff[1]= hadamard8_diff_mmx; c->pix_norm1 = pix_norm1_mmx; - c->sse[0] = (mm_flags & MM_SSE2) ? sse16_sse2 : sse16_mmx; + c->sse[0] = (ff_mm_flags & MM_SSE2) ? sse16_sse2 : sse16_mmx; c->sse[1] = sse8_mmx; c->vsad[4]= vsad_intra16_mmx; @@ -3227,7 +3227,7 @@ c->h264_idct8_dc_add= c->h264_idct8_add= ff_h264_idct8_add_mmx; - if (mm_flags & MM_MMXEXT) { + if (ff_mm_flags & MM_MMXEXT) { c->prefetch = prefetch_mmx2; c->put_pixels_tab[0][1] = put_pixels16_x2_mmx2; @@ -3369,7 +3369,7 @@ #ifdef CONFIG_ENCODERS c->sub_hfyu_median_prediction= sub_hfyu_median_prediction_mmx2; #endif //CONFIG_ENCODERS - } else if (mm_flags & MM_3DNOW) { + } else if (ff_mm_flags & MM_3DNOW) { c->prefetch = prefetch_3dnow; c->put_pixels_tab[0][1] = put_pixels16_x2_3dnow; @@ -3463,7 +3463,7 @@ } #ifdef CONFIG_SNOW_ENCODER - if(mm_flags & MM_SSE2){ + if(ff_mm_flags & MM_SSE2){ c->horizontal_compose97i = ff_snow_horizontal_compose97i_sse2; c->vertical_compose97i = ff_snow_vertical_compose97i_sse2; c->inner_add_yblock = ff_snow_inner_add_yblock_sse2; @@ -3475,22 +3475,22 @@ } #endif - if(mm_flags & MM_3DNOW){ + if(ff_mm_flags & MM_3DNOW){ c->vorbis_inverse_coupling = vorbis_inverse_coupling_3dnow; c->vector_fmul = vector_fmul_3dnow; if(!(avctx->flags & CODEC_FLAG_BITEXACT)) c->float_to_int16 = float_to_int16_3dnow; } - if(mm_flags & MM_3DNOWEXT) + if(ff_mm_flags & MM_3DNOWEXT) c->vector_fmul_reverse = vector_fmul_reverse_3dnow2; - if(mm_flags & MM_SSE){ + if(ff_mm_flags & MM_SSE){ c->vorbis_inverse_coupling = vorbis_inverse_coupling_sse; c->vector_fmul = vector_fmul_sse; c->float_to_int16 = float_to_int16_sse; c->vector_fmul_reverse = vector_fmul_reverse_sse; c->vector_fmul_add_add = vector_fmul_add_add_sse; } - if(mm_flags & MM_3DNOW) + if(ff_mm_flags & MM_3DNOW) c->vector_fmul_add_add = vector_fmul_add_add_3dnow; // faster than sse }
--- a/i386/motion_est_mmx.c Tue Nov 14 00:56:56 2006 +0000 +++ b/i386/motion_est_mmx.c Tue Nov 14 03:12:29 2006 +0000 @@ -376,7 +376,7 @@ void dsputil_init_pix_mmx(DSPContext* c, AVCodecContext *avctx) { - if (mm_flags & MM_MMX) { + if (ff_mm_flags & MM_MMX) { c->pix_abs[0][0] = sad16_mmx; c->pix_abs[0][1] = sad16_x2_mmx; c->pix_abs[0][2] = sad16_y2_mmx; @@ -389,7 +389,7 @@ c->sad[0]= sad16_mmx; c->sad[1]= sad8_mmx; } - if (mm_flags & MM_MMXEXT) { + if (ff_mm_flags & MM_MMXEXT) { c->pix_abs[0][0] = sad16_mmx2; c->pix_abs[1][0] = sad8_mmx2;
--- a/i386/mpegvideo_mmx.c Tue Nov 14 00:56:56 2006 +0000 +++ b/i386/mpegvideo_mmx.c Tue Nov 14 03:12:29 2006 +0000 @@ -27,7 +27,7 @@ #include "../avcodec.h" #include "x86_cpu.h" -extern uint16_t inv_zigzag_direct16[64]; +extern uint16_t ff_inv_zigzag_direct16[64]; static const unsigned long long int mm_wabs __attribute__ ((aligned(8))) = 0xffffffffffffffffULL; static const unsigned long long int mm_wone __attribute__ ((aligned(8))) = 0x0001000100010001ULL; @@ -693,7 +693,7 @@ void MPV_common_init_mmx(MpegEncContext *s) { - if (mm_flags & MM_MMX) { + if (ff_mm_flags & MM_MMX) { const int dct_algo = s->avctx->dct_algo; s->dct_unquantize_h263_intra = dct_unquantize_h263_intra_mmx; @@ -706,16 +706,16 @@ draw_edges = draw_edges_mmx; - if (mm_flags & MM_SSE2) { + if (ff_mm_flags & MM_SSE2) { s->denoise_dct= denoise_dct_sse2; } else { s->denoise_dct= denoise_dct_mmx; } if(dct_algo==FF_DCT_AUTO || dct_algo==FF_DCT_MMX){ - if(mm_flags & MM_SSE2){ + if(ff_mm_flags & MM_SSE2){ s->dct_quantize= dct_quantize_SSE2; - } else if(mm_flags & MM_MMXEXT){ + } else if(ff_mm_flags & MM_MMXEXT){ s->dct_quantize= dct_quantize_MMX2; } else { s->dct_quantize= dct_quantize_MMX;
--- a/i386/mpegvideo_mmx_template.c Tue Nov 14 00:56:56 2006 +0000 +++ b/i386/mpegvideo_mmx_template.c Tue Nov 14 03:12:29 2006 +0000 @@ -139,7 +139,7 @@ "movzb %%al, %%"REG_a" \n\t" // last_non_zero_p1 : "+a" (last_non_zero_p1) : "r" (block+64), "r" (qmat), "r" (bias), - "r" (inv_zigzag_direct16+64), "r" (temp_block+64) + "r" (ff_inv_zigzag_direct16+64), "r" (temp_block+64) ); // note the asm is split cuz gcc doesnt like that many operands ... asm volatile( @@ -185,7 +185,7 @@ "movzb %%al, %%"REG_a" \n\t" // last_non_zero_p1 : "+a" (last_non_zero_p1) : "r" (block+64), "r" (qmat+64), "r" (bias+64), - "r" (inv_zigzag_direct16+64), "r" (temp_block+64) + "r" (ff_inv_zigzag_direct16+64), "r" (temp_block+64) ); // note the asm is split cuz gcc doesnt like that many operands ... asm volatile(
--- a/imgresample.c Tue Nov 14 00:56:56 2006 +0000 +++ b/imgresample.c Tue Nov 14 03:12:29 2006 +0000 @@ -469,7 +469,7 @@ n = dst_width; } #ifdef HAVE_MMX - if ((mm_flags & MM_MMX) && NB_TAPS == 4) + if ((ff_mm_flags & MM_MMX) && NB_TAPS == 4) h_resample_fast4_mmx(dst, n, src, src_width, src_start, src_incr, filters); else @@ -527,14 +527,14 @@ phase_y = get_phase(src_y); #ifdef HAVE_MMX /* desactivated MMX because loss of precision */ - if ((mm_flags & MM_MMX) && NB_TAPS == 4 && 0) + if ((ff_mm_flags & MM_MMX) && NB_TAPS == 4 && 0) v_resample4_mmx(output, owidth, s->line_buf + (ring_y - NB_TAPS + 1) * owidth, owidth, &s->v_filters[phase_y][0]); else #endif #ifdef HAVE_ALTIVEC - if ((mm_flags & MM_ALTIVEC) && NB_TAPS == 4 && FILTER_BITS <= 6) + if ((ff_mm_flags & MM_ALTIVEC) && NB_TAPS == 4 && FILTER_BITS <= 6) v_resample16_altivec(output, owidth, s->line_buf + (ring_y - NB_TAPS + 1) * owidth, owidth, &s->v_filters[phase_y][0]);
--- a/ppc/dsputil_ppc.c Tue Nov 14 00:56:56 2006 +0000 +++ b/ppc/dsputil_ppc.c Tue Nov 14 03:12:29 2006 +0000 @@ -42,7 +42,7 @@ #endif -int mm_flags = 0; +int ff_mm_flags = 0; int mm_support(void) { @@ -277,7 +277,7 @@ dsputil_h264_init_ppc(c, avctx); if (has_altivec()) { - mm_flags |= MM_ALTIVEC; + ff_mm_flags |= MM_ALTIVEC; dsputil_init_altivec(c, avctx); snow_init_altivec(c, avctx);
--- a/snow.c Tue Nov 14 00:56:56 2006 +0000 +++ b/snow.c Tue Nov 14 03:12:29 2006 +0000 @@ -3001,9 +3001,9 @@ * improve the score of the whole frame, so iterative motion est * doesn't always converge. */ if(s->avctx->me_cmp == FF_CMP_W97) - distortion = w97_32_c(&s->m, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, 32); + distortion = ff_w97_32_c(&s->m, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, 32); else if(s->avctx->me_cmp == FF_CMP_W53) - distortion = w53_32_c(&s->m, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, 32); + distortion = ff_w53_32_c(&s->m, src + sx + sy*ref_stride, dst + sx + sy*ref_stride, ref_stride, 32); else{ distortion = 0; for(i=0; i<4; i++){
--- a/snow.h Tue Nov 14 00:56:56 2006 +0000 +++ b/snow.h Tue Nov 14 03:12:29 2006 +0000 @@ -128,8 +128,8 @@ extern void ff_snow_inner_add_yblock(uint8_t *obmc, const int obmc_stride, uint8_t * * block, int b_w, int b_h, int src_x, int src_y, int src_stride, slice_buffer * sb, int add, uint8_t * dst8); #ifdef CONFIG_SNOW_ENCODER -int w53_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h); -int w97_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h); +int ff_w53_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h); +int ff_w97_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h); #else static int w53_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h) {assert (0);} static int w97_32_c(void *v, uint8_t * pix1, uint8_t * pix2, int line_size, int h) {assert (0);}