Mercurial > mplayer.hg
view libmpcodecs/vf_spp.c @ 16718:044260623695
makes demux_lavf (-demuxer 35) use the framerate specified in the container
if it's set and only fall back to the codec framerate if the former is not
set.
This solves the issue of some avi's playing at 30000/1 fps instead of the
correct framerate.
Patch by Ivo < ivop AH euronet POIS nl >
Original thread:
Date: Sep 25, 2005 12:34 AM
Subject: [MPlayer-dev-eng] [PATCH] make demux_lavf use container framerate
author | gpoirier |
---|---|
date | Mon, 10 Oct 2005 05:45:38 +0000 |
parents | bdf1b4ecb906 |
children | 6ff3379a0862 |
line wrap: on
line source
/* Copyright (C) 2003 Michael Niedermayer <michaelni@gmx.at> This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA */ /* * This implementation is based on an algorithm described in * "Aria Nosratinia Embedded Post-Processing for * Enhancement of Compressed Images (1999)" * (http://citeseer.nj.nec.com/nosratinia99embedded.html) */ #include <stdio.h> #include <stdlib.h> #include <string.h> #include <inttypes.h> #include <math.h> #include "../config.h" #ifdef USE_LIBAVCODEC #include "../mp_msg.h" #include "../cpudetect.h" #ifdef USE_LIBAVCODEC_SO #include <ffmpeg/avcodec.h> #include <ffmpeg/dsputil.h> #else #include "../libavcodec/avcodec.h" #include "../libavcodec/dsputil.h" #endif #ifdef HAVE_MALLOC_H #include <malloc.h> #endif #include "img_format.h" #include "mp_image.h" #include "vf.h" #include "../libvo/fastmemcpy.h" #define XMIN(a,b) ((a) < (b) ? (a) : (b)) //===========================================================================// static const uint8_t __attribute__((aligned(8))) dither[8][8]={ { 0, 48, 12, 60, 3, 51, 15, 63, }, { 32, 16, 44, 28, 35, 19, 47, 31, }, { 8, 56, 4, 52, 11, 59, 7, 55, }, { 40, 24, 36, 20, 43, 27, 39, 23, }, { 2, 50, 14, 62, 1, 49, 13, 61, }, { 34, 18, 46, 30, 33, 17, 45, 29, }, { 10, 58, 6, 54, 9, 57, 5, 53, }, { 42, 26, 38, 22, 41, 25, 37, 21, }, }; static const uint8_t offset[127][2]= { {0,0}, {0,0}, {4,4}, {0,0}, {2,2}, {6,4}, {4,6}, {0,0}, {5,1}, {2,2}, {7,3}, {4,4}, {1,5}, {6,6}, {3,7}, {0,0}, {4,0}, {1,1}, {5,1}, {3,2}, {7,2}, {2,3}, {6,3}, {0,4}, {4,4}, {1,5}, {5,5}, {3,6}, {7,6}, {2,7}, {6,7}, {0,0}, {0,2}, {0,4}, {0,6}, {1,1}, {1,3}, {1,5}, {1,7}, {2,0}, {2,2}, {2,4}, {2,6}, {3,1}, {3,3}, {3,5}, {3,7}, {4,0}, {4,2}, {4,4}, {4,6}, {5,1}, {5,3}, {5,5}, {5,7}, {6,0}, {6,2}, {6,4}, {6,6}, {7,1}, {7,3}, {7,5}, {7,7}, {0,0}, {4,4}, {0,4}, {4,0}, {2,2}, {6,6}, {2,6}, {6,2}, {0,2}, {4,6}, {0,6}, {4,2}, {2,0}, {6,4}, {2,4}, {6,0}, {1,1}, {5,5}, {1,5}, {5,1}, {3,3}, {7,7}, {3,7}, {7,3}, {1,3}, {5,7}, {1,7}, {5,3}, {3,1}, {7,5}, {3,5}, {7,1}, {0,1}, {4,5}, {0,5}, {4,1}, {2,3}, {6,7}, {2,7}, {6,3}, {0,3}, {4,7}, {0,7}, {4,3}, {2,1}, {6,5}, {2,5}, {6,1}, {1,0}, {5,4}, {1,4}, {5,0}, {3,2}, {7,6}, {3,6}, {7,2}, {1,2}, {5,6}, {1,6}, {5,2}, {3,0}, {7,4}, {3,4}, {7,0}, }; struct vf_priv_s { int log2_count; int qp; int mode; int mpeg2; int temp_stride; uint8_t *src; int16_t *temp; AVCodecContext *avctx; DSPContext dsp; }; #define SHIFT 22 static void hardthresh_c(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){ int i; int bias= 0; //FIXME unsigned int threshold1, threshold2; threshold1= qp*((1<<4) - bias) - 1; threshold2= (threshold1<<1); memset(dst, 0, 64*sizeof(DCTELEM)); dst[0]= (src[0] + 4)>>3; for(i=1; i<64; i++){ int level= src[i]; if(((unsigned)(level+threshold1))>threshold2){ const int j= permutation[i]; dst[j]= (level + 4)>>3; } } } static void softthresh_c(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){ int i; int bias= 0; //FIXME unsigned int threshold1, threshold2; threshold1= qp*((1<<4) - bias) - 1; threshold2= (threshold1<<1); memset(dst, 0, 64*sizeof(DCTELEM)); dst[0]= (src[0] + 4)>>3; for(i=1; i<64; i++){ int level= src[i]; if(((unsigned)(level+threshold1))>threshold2){ const int j= permutation[i]; if(level>0) dst[j]= (level - threshold1 + 4)>>3; else dst[j]= (level + threshold1 + 4)>>3; } } } #ifdef HAVE_MMX static void hardthresh_mmx(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){ int bias= 0; //FIXME unsigned int threshold1; threshold1= qp*((1<<4) - bias) - 1; asm volatile( #define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \ "movq " #src0 ", %%mm0 \n\t"\ "movq " #src1 ", %%mm1 \n\t"\ "movq " #src2 ", %%mm2 \n\t"\ "movq " #src3 ", %%mm3 \n\t"\ "psubw %%mm4, %%mm0 \n\t"\ "psubw %%mm4, %%mm1 \n\t"\ "psubw %%mm4, %%mm2 \n\t"\ "psubw %%mm4, %%mm3 \n\t"\ "paddusw %%mm5, %%mm0 \n\t"\ "paddusw %%mm5, %%mm1 \n\t"\ "paddusw %%mm5, %%mm2 \n\t"\ "paddusw %%mm5, %%mm3 \n\t"\ "paddw %%mm6, %%mm0 \n\t"\ "paddw %%mm6, %%mm1 \n\t"\ "paddw %%mm6, %%mm2 \n\t"\ "paddw %%mm6, %%mm3 \n\t"\ "psubusw %%mm6, %%mm0 \n\t"\ "psubusw %%mm6, %%mm1 \n\t"\ "psubusw %%mm6, %%mm2 \n\t"\ "psubusw %%mm6, %%mm3 \n\t"\ "psraw $3, %%mm0 \n\t"\ "psraw $3, %%mm1 \n\t"\ "psraw $3, %%mm2 \n\t"\ "psraw $3, %%mm3 \n\t"\ \ "movq %%mm0, %%mm7 \n\t"\ "punpcklwd %%mm2, %%mm0 \n\t" /*A*/\ "punpckhwd %%mm2, %%mm7 \n\t" /*C*/\ "movq %%mm1, %%mm2 \n\t"\ "punpcklwd %%mm3, %%mm1 \n\t" /*B*/\ "punpckhwd %%mm3, %%mm2 \n\t" /*D*/\ "movq %%mm0, %%mm3 \n\t"\ "punpcklwd %%mm1, %%mm0 \n\t" /*A*/\ "punpckhwd %%mm7, %%mm3 \n\t" /*C*/\ "punpcklwd %%mm2, %%mm7 \n\t" /*B*/\ "punpckhwd %%mm2, %%mm1 \n\t" /*D*/\ \ "movq %%mm0, " #dst0 " \n\t"\ "movq %%mm7, " #dst1 " \n\t"\ "movq %%mm3, " #dst2 " \n\t"\ "movq %%mm1, " #dst3 " \n\t" "movd %2, %%mm4 \n\t" "movd %3, %%mm5 \n\t" "movd %4, %%mm6 \n\t" "packssdw %%mm4, %%mm4 \n\t" "packssdw %%mm5, %%mm5 \n\t" "packssdw %%mm6, %%mm6 \n\t" "packssdw %%mm4, %%mm4 \n\t" "packssdw %%mm5, %%mm5 \n\t" "packssdw %%mm6, %%mm6 \n\t" REQUANT_CORE( (%1), 8(%1), 16(%1), 24(%1), (%0), 8(%0), 64(%0), 72(%0)) REQUANT_CORE(32(%1), 40(%1), 48(%1), 56(%1),16(%0),24(%0), 48(%0), 56(%0)) REQUANT_CORE(64(%1), 72(%1), 80(%1), 88(%1),32(%0),40(%0), 96(%0),104(%0)) REQUANT_CORE(96(%1),104(%1),112(%1),120(%1),80(%0),88(%0),112(%0),120(%0)) : : "r" (src), "r" (dst), "g" (threshold1+1), "g" (threshold1+5), "g" (threshold1-4) //FIXME maybe more accurate then needed? ); dst[0]= (src[0] + 4)>>3; } static void softthresh_mmx(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){ int bias= 0; //FIXME unsigned int threshold1; threshold1= qp*((1<<4) - bias) - 1; asm volatile( #undef REQUANT_CORE #define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \ "movq " #src0 ", %%mm0 \n\t"\ "movq " #src1 ", %%mm1 \n\t"\ "pxor %%mm6, %%mm6 \n\t"\ "pxor %%mm7, %%mm7 \n\t"\ "pcmpgtw %%mm0, %%mm6 \n\t"\ "pcmpgtw %%mm1, %%mm7 \n\t"\ "pxor %%mm6, %%mm0 \n\t"\ "pxor %%mm7, %%mm1 \n\t"\ "psubusw %%mm4, %%mm0 \n\t"\ "psubusw %%mm4, %%mm1 \n\t"\ "pxor %%mm6, %%mm0 \n\t"\ "pxor %%mm7, %%mm1 \n\t"\ "movq " #src2 ", %%mm2 \n\t"\ "movq " #src3 ", %%mm3 \n\t"\ "pxor %%mm6, %%mm6 \n\t"\ "pxor %%mm7, %%mm7 \n\t"\ "pcmpgtw %%mm2, %%mm6 \n\t"\ "pcmpgtw %%mm3, %%mm7 \n\t"\ "pxor %%mm6, %%mm2 \n\t"\ "pxor %%mm7, %%mm3 \n\t"\ "psubusw %%mm4, %%mm2 \n\t"\ "psubusw %%mm4, %%mm3 \n\t"\ "pxor %%mm6, %%mm2 \n\t"\ "pxor %%mm7, %%mm3 \n\t"\ \ "paddsw %%mm5, %%mm0 \n\t"\ "paddsw %%mm5, %%mm1 \n\t"\ "paddsw %%mm5, %%mm2 \n\t"\ "paddsw %%mm5, %%mm3 \n\t"\ "psraw $3, %%mm0 \n\t"\ "psraw $3, %%mm1 \n\t"\ "psraw $3, %%mm2 \n\t"\ "psraw $3, %%mm3 \n\t"\ \ "movq %%mm0, %%mm7 \n\t"\ "punpcklwd %%mm2, %%mm0 \n\t" /*A*/\ "punpckhwd %%mm2, %%mm7 \n\t" /*C*/\ "movq %%mm1, %%mm2 \n\t"\ "punpcklwd %%mm3, %%mm1 \n\t" /*B*/\ "punpckhwd %%mm3, %%mm2 \n\t" /*D*/\ "movq %%mm0, %%mm3 \n\t"\ "punpcklwd %%mm1, %%mm0 \n\t" /*A*/\ "punpckhwd %%mm7, %%mm3 \n\t" /*C*/\ "punpcklwd %%mm2, %%mm7 \n\t" /*B*/\ "punpckhwd %%mm2, %%mm1 \n\t" /*D*/\ \ "movq %%mm0, " #dst0 " \n\t"\ "movq %%mm7, " #dst1 " \n\t"\ "movq %%mm3, " #dst2 " \n\t"\ "movq %%mm1, " #dst3 " \n\t" "movd %2, %%mm4 \n\t" "movd %3, %%mm5 \n\t" "packssdw %%mm4, %%mm4 \n\t" "packssdw %%mm5, %%mm5 \n\t" "packssdw %%mm4, %%mm4 \n\t" "packssdw %%mm5, %%mm5 \n\t" REQUANT_CORE( (%1), 8(%1), 16(%1), 24(%1), (%0), 8(%0), 64(%0), 72(%0)) REQUANT_CORE(32(%1), 40(%1), 48(%1), 56(%1),16(%0),24(%0), 48(%0), 56(%0)) REQUANT_CORE(64(%1), 72(%1), 80(%1), 88(%1),32(%0),40(%0), 96(%0),104(%0)) REQUANT_CORE(96(%1),104(%1),112(%1),120(%1),80(%0),88(%0),112(%0),120(%0)) : : "r" (src), "r" (dst), "g" (threshold1), "rm" (4) //FIXME maybe more accurate then needed? ); dst[0]= (src[0] + 4)>>3; } #endif static inline void add_block(int16_t *dst, int stride, DCTELEM block[64]){ int y; for(y=0; y<8; y++){ *(uint32_t*)&dst[0 + y*stride]+= *(uint32_t*)&block[0 + y*8]; *(uint32_t*)&dst[2 + y*stride]+= *(uint32_t*)&block[2 + y*8]; *(uint32_t*)&dst[4 + y*stride]+= *(uint32_t*)&block[4 + y*8]; *(uint32_t*)&dst[6 + y*stride]+= *(uint32_t*)&block[6 + y*8]; } } static void store_slice_c(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale){ int y, x; #define STORE(pos) \ temp= ((src[x + y*src_stride + pos]<<log2_scale) + d[pos])>>6;\ if(temp & 0x100) temp= ~(temp>>31);\ dst[x + y*dst_stride + pos]= temp; for(y=0; y<height; y++){ const uint8_t *d= dither[y]; for(x=0; x<width; x+=8){ int temp; STORE(0); STORE(1); STORE(2); STORE(3); STORE(4); STORE(5); STORE(6); STORE(7); } } } #ifdef HAVE_MMX static void store_slice_mmx(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale){ int y; for(y=0; y<height; y++){ uint8_t *dst1= dst; int16_t *src1= src; asm volatile( "movq (%3), %%mm3 \n\t" "movq (%3), %%mm4 \n\t" "movd %4, %%mm2 \n\t" "pxor %%mm0, %%mm0 \n\t" "punpcklbw %%mm0, %%mm3 \n\t" "punpckhbw %%mm0, %%mm4 \n\t" "psraw %%mm2, %%mm3 \n\t" "psraw %%mm2, %%mm4 \n\t" "movd %5, %%mm2 \n\t" "1: \n\t" "movq (%0), %%mm0 \n\t" "movq 8(%0), %%mm1 \n\t" "paddw %%mm3, %%mm0 \n\t" "paddw %%mm4, %%mm1 \n\t" "psraw %%mm2, %%mm0 \n\t" "psraw %%mm2, %%mm1 \n\t" "packuswb %%mm1, %%mm0 \n\t" "movq %%mm0, (%1) \n\t" "add $16, %0 \n\t" "add $8, %1 \n\t" "cmp %2, %1 \n\t" " jb 1b \n\t" : "+r" (src1), "+r"(dst1) : "r"(dst + width), "r"(dither[y]), "g"(log2_scale), "g"(6-log2_scale) ); src += src_stride; dst += dst_stride; } // if(width != mmxw) // store_slice_c(dst + mmxw, src + mmxw, dst_stride, src_stride, width - mmxw, log2_scale); } #endif static void (*store_slice)(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale)= store_slice_c; static void (*requantize)(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation)= hardthresh_c; static void filter(struct vf_priv_s *p, uint8_t *dst, uint8_t *src, int dst_stride, int src_stride, int width, int height, uint8_t *qp_store, int qp_stride, int is_luma){ int x, y, i; const int count= 1<<p->log2_count; const int stride= is_luma ? p->temp_stride : ((width+16+15)&(~15)); uint64_t block_align[32]; DCTELEM *block = (DCTELEM *)block_align; DCTELEM *block2= (DCTELEM *)(block_align+16); if (!src || !dst) return; // HACK avoid crash for Y8 colourspace for(y=0; y<height; y++){ int index= 8 + 8*stride + y*stride; memcpy(p->src + index, src + y*src_stride, width); for(x=0; x<8; x++){ p->src[index - x - 1]= p->src[index + x ]; p->src[index + width + x ]= p->src[index + width - x - 1]; } } for(y=0; y<8; y++){ memcpy(p->src + ( 7-y)*stride, p->src + ( y+8)*stride, stride); memcpy(p->src + (height+8+y)*stride, p->src + (height-y+7)*stride, stride); } //FIXME (try edge emu) for(y=0; y<height+8; y+=8){ memset(p->temp + (8+y)*stride, 0, 8*stride*sizeof(int16_t)); for(x=0; x<width+8; x+=8){ const int qps= 3 + is_luma; int qp; if(p->qp) qp= p->qp; else{ qp= qp_store[ (XMIN(x, width-1)>>qps) + (XMIN(y, height-1)>>qps) * qp_stride]; if(p->mpeg2) qp>>=1; } for(i=0; i<count; i++){ const int x1= x + offset[i+count-1][0]; const int y1= y + offset[i+count-1][1]; const int index= x1 + y1*stride; p->dsp.get_pixels(block, p->src + index, stride); p->dsp.fdct(block); requantize(block2, block, qp, p->dsp.idct_permutation); p->dsp.idct(block2); add_block(p->temp + index, stride, block2); } } if(y) store_slice(dst + (y-8)*dst_stride, p->temp + 8 + y*stride, dst_stride, stride, width, XMIN(8, height+8-y), 6-p->log2_count); } #if 0 for(y=0; y<height; y++){ for(x=0; x<width; x++){ if((((x>>6) ^ (y>>6)) & 1) == 0) dst[x + y*dst_stride]= p->src[8 + 8*stride + x + y*stride]; if((x&63) == 0 || (y&63)==0) dst[x + y*dst_stride] += 128; } } #endif //FIXME reorder for better caching } static int config(struct vf_instance_s* vf, int width, int height, int d_width, int d_height, unsigned int flags, unsigned int outfmt){ int h= (height+16+15)&(~15); vf->priv->temp_stride= (width+16+15)&(~15); vf->priv->temp= malloc(vf->priv->temp_stride*h*sizeof(int16_t)); vf->priv->src = malloc(vf->priv->temp_stride*h*sizeof(uint8_t)); return vf_next_config(vf,width,height,d_width,d_height,flags,outfmt); } static void get_image(struct vf_instance_s* vf, mp_image_t *mpi){ if(mpi->flags&MP_IMGFLAG_PRESERVE) return; // don't change // ok, we can do pp in-place (or pp disabled): vf->dmpi=vf_get_image(vf->next,mpi->imgfmt, mpi->type, mpi->flags | MP_IMGFLAG_READABLE, mpi->width, mpi->height); mpi->planes[0]=vf->dmpi->planes[0]; mpi->stride[0]=vf->dmpi->stride[0]; mpi->width=vf->dmpi->width; if(mpi->flags&MP_IMGFLAG_PLANAR){ mpi->planes[1]=vf->dmpi->planes[1]; mpi->planes[2]=vf->dmpi->planes[2]; mpi->stride[1]=vf->dmpi->stride[1]; mpi->stride[2]=vf->dmpi->stride[2]; } mpi->flags|=MP_IMGFLAG_DIRECT; } static int put_image(struct vf_instance_s* vf, mp_image_t *mpi){ mp_image_t *dmpi; if(!(mpi->flags&MP_IMGFLAG_DIRECT)){ // no DR, so get a new image! hope we'll get DR buffer: dmpi=vf_get_image(vf->next,mpi->imgfmt, MP_IMGTYPE_TEMP, MP_IMGFLAG_ACCEPT_STRIDE|MP_IMGFLAG_PREFER_ALIGNED_STRIDE, mpi->width,mpi->height); vf_clone_mpi_attributes(dmpi, mpi); }else{ dmpi=vf->dmpi; } vf->priv->mpeg2= mpi->qscale_type; if(vf->priv->log2_count || !(mpi->flags&MP_IMGFLAG_DIRECT)){ if(mpi->qscale || vf->priv->qp){ filter(vf->priv, dmpi->planes[0], mpi->planes[0], dmpi->stride[0], mpi->stride[0], mpi->w, mpi->h, mpi->qscale, mpi->qstride, 1); filter(vf->priv, dmpi->planes[1], mpi->planes[1], dmpi->stride[1], mpi->stride[1], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, mpi->qscale, mpi->qstride, 0); filter(vf->priv, dmpi->planes[2], mpi->planes[2], dmpi->stride[2], mpi->stride[2], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, mpi->qscale, mpi->qstride, 0); }else{ memcpy_pic(dmpi->planes[0], mpi->planes[0], mpi->w, mpi->h, dmpi->stride[0], mpi->stride[0]); memcpy_pic(dmpi->planes[1], mpi->planes[1], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, dmpi->stride[1], mpi->stride[1]); memcpy_pic(dmpi->planes[2], mpi->planes[2], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, dmpi->stride[2], mpi->stride[2]); } } #ifdef HAVE_MMX if(gCpuCaps.hasMMX) asm volatile ("emms\n\t"); #endif #ifdef HAVE_MMX2 if(gCpuCaps.hasMMX2) asm volatile ("sfence\n\t"); #endif return vf_next_put_image(vf,dmpi); } static void uninit(struct vf_instance_s* vf){ if(!vf->priv) return; if(vf->priv->temp) free(vf->priv->temp); vf->priv->temp= NULL; if(vf->priv->src) free(vf->priv->src); vf->priv->src= NULL; if(vf->priv->avctx) free(vf->priv->avctx); vf->priv->avctx= NULL; free(vf->priv); vf->priv=NULL; } //===========================================================================// static int query_format(struct vf_instance_s* vf, unsigned int fmt){ switch(fmt){ case IMGFMT_YVU9: case IMGFMT_IF09: case IMGFMT_YV12: case IMGFMT_I420: case IMGFMT_IYUV: case IMGFMT_CLPL: case IMGFMT_Y800: case IMGFMT_Y8: case IMGFMT_444P: case IMGFMT_422P: case IMGFMT_411P: return vf_next_query_format(vf,fmt); } return 0; } static unsigned int fmt_list[]={ IMGFMT_YVU9, IMGFMT_IF09, IMGFMT_YV12, IMGFMT_I420, IMGFMT_IYUV, IMGFMT_CLPL, IMGFMT_Y800, IMGFMT_Y8, IMGFMT_444P, IMGFMT_422P, IMGFMT_411P, 0 }; static int control(struct vf_instance_s* vf, int request, void* data){ switch(request){ case VFCTRL_QUERY_MAX_PP_LEVEL: return 6; case VFCTRL_SET_PP_LEVEL: vf->priv->log2_count= *((unsigned int*)data); return CONTROL_TRUE; } return vf_next_control(vf,request,data); } static int open(vf_instance_t *vf, char* args){ int log2c=-1; vf->config=config; vf->put_image=put_image; vf->get_image=get_image; vf->query_format=query_format; vf->uninit=uninit; vf->control= control; vf->priv=malloc(sizeof(struct vf_priv_s)); memset(vf->priv, 0, sizeof(struct vf_priv_s)); avcodec_init(); vf->priv->avctx= avcodec_alloc_context(); dsputil_init(&vf->priv->dsp, vf->priv->avctx); vf->priv->log2_count= 3; if (args) sscanf(args, "%d:%d:%d", &log2c, &vf->priv->qp, &vf->priv->mode); if( log2c >=0 && log2c <=6 ) vf->priv->log2_count = log2c; if(vf->priv->qp < 0) vf->priv->qp = 0; switch(vf->priv->mode){ default: case 0: requantize= hardthresh_c; break; case 1: requantize= softthresh_c; break; } #ifdef HAVE_MMX if(gCpuCaps.hasMMX){ store_slice= store_slice_mmx; switch(vf->priv->mode){ case 0: requantize= hardthresh_mmx; break; case 1: requantize= softthresh_mmx; break; } } #endif return 1; } vf_info_t vf_info_spp = { "simple postprocess", "spp", "Michael Niedermayer", "", open, NULL }; #endif //USE_LIBAVCODEC