Mercurial > mplayer.hg
view libmpcodecs/vf_spp.c @ 23510:a6c619ee9d30
Teletext support for tv:// (v4l and v4l2 only)
modified patch from Otvos Attila oattila at chello dot hu
Module uses zvbi library for all low-level VBI operations (like I/O with vbi
device, converting vbi pages into usefull vbi_page stuctures, rendering them
into RGB32 images).
All teletext related stuff (except properties, slave commands and rendering
osd in text mode or RGB32 rendered teletext pages in spu mode) is implemented
in tvi_vbi.c
New properties:
teletext_page - switching between pages
teletext_mode - switch between on/off/opaque/transparent modes
teletext_format - (currently read-only) allows to get format info
(black/white,gray,text)
teletext_half_page - trivial zooming (displaying top/bottom half of teletext
page)
New slave commands:
teletext_add_dec - user interface for jumping to any page by editing page number
interactively
teletext_go_link - goes though links, specified on current page
author | voroshil |
---|---|
date | Sun, 10 Jun 2007 00:06:12 +0000 |
parents | a124f3abc1ec |
children | c9c2e550a449 |
line wrap: on
line source
/* Copyright (C) 2003 Michael Niedermayer <michaelni@gmx.at> This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA */ /* * This implementation is based on an algorithm described in * "Aria Nosratinia Embedded Post-Processing for * Enhancement of Compressed Images (1999)" * (http://citeseer.nj.nec.com/nosratinia99embedded.html) */ #include <stdio.h> #include <stdlib.h> #include <string.h> #include <inttypes.h> #include <math.h> #include "config.h" #include "mp_msg.h" #include "cpudetect.h" #include "libavcodec/avcodec.h" #include "libavcodec/dsputil.h" #ifdef HAVE_MALLOC_H #include <malloc.h> #endif #include "img_format.h" #include "mp_image.h" #include "vf.h" #include "libvo/fastmemcpy.h" #define XMIN(a,b) ((a) < (b) ? (a) : (b)) //===========================================================================// static const uint8_t __attribute__((aligned(8))) dither[8][8]={ { 0, 48, 12, 60, 3, 51, 15, 63, }, { 32, 16, 44, 28, 35, 19, 47, 31, }, { 8, 56, 4, 52, 11, 59, 7, 55, }, { 40, 24, 36, 20, 43, 27, 39, 23, }, { 2, 50, 14, 62, 1, 49, 13, 61, }, { 34, 18, 46, 30, 33, 17, 45, 29, }, { 10, 58, 6, 54, 9, 57, 5, 53, }, { 42, 26, 38, 22, 41, 25, 37, 21, }, }; static const uint8_t offset[127][2]= { {0,0}, {0,0}, {4,4}, {0,0}, {2,2}, {6,4}, {4,6}, {0,0}, {5,1}, {2,2}, {7,3}, {4,4}, {1,5}, {6,6}, {3,7}, {0,0}, {4,0}, {1,1}, {5,1}, {3,2}, {7,2}, {2,3}, {6,3}, {0,4}, {4,4}, {1,5}, {5,5}, {3,6}, {7,6}, {2,7}, {6,7}, {0,0}, {0,2}, {0,4}, {0,6}, {1,1}, {1,3}, {1,5}, {1,7}, {2,0}, {2,2}, {2,4}, {2,6}, {3,1}, {3,3}, {3,5}, {3,7}, {4,0}, {4,2}, {4,4}, {4,6}, {5,1}, {5,3}, {5,5}, {5,7}, {6,0}, {6,2}, {6,4}, {6,6}, {7,1}, {7,3}, {7,5}, {7,7}, {0,0}, {4,4}, {0,4}, {4,0}, {2,2}, {6,6}, {2,6}, {6,2}, {0,2}, {4,6}, {0,6}, {4,2}, {2,0}, {6,4}, {2,4}, {6,0}, {1,1}, {5,5}, {1,5}, {5,1}, {3,3}, {7,7}, {3,7}, {7,3}, {1,3}, {5,7}, {1,7}, {5,3}, {3,1}, {7,5}, {3,5}, {7,1}, {0,1}, {4,5}, {0,5}, {4,1}, {2,3}, {6,7}, {2,7}, {6,3}, {0,3}, {4,7}, {0,7}, {4,3}, {2,1}, {6,5}, {2,5}, {6,1}, {1,0}, {5,4}, {1,4}, {5,0}, {3,2}, {7,6}, {3,6}, {7,2}, {1,2}, {5,6}, {1,6}, {5,2}, {3,0}, {7,4}, {3,4}, {7,0}, }; struct vf_priv_s { int log2_count; int qp; int mode; int mpeg2; int temp_stride; uint8_t *src; int16_t *temp; AVCodecContext *avctx; DSPContext dsp; char *non_b_qp; }; #define SHIFT 22 static void hardthresh_c(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){ int i; int bias= 0; //FIXME unsigned int threshold1, threshold2; threshold1= qp*((1<<4) - bias) - 1; threshold2= (threshold1<<1); memset(dst, 0, 64*sizeof(DCTELEM)); dst[0]= (src[0] + 4)>>3; for(i=1; i<64; i++){ int level= src[i]; if(((unsigned)(level+threshold1))>threshold2){ const int j= permutation[i]; dst[j]= (level + 4)>>3; } } } static void softthresh_c(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){ int i; int bias= 0; //FIXME unsigned int threshold1, threshold2; threshold1= qp*((1<<4) - bias) - 1; threshold2= (threshold1<<1); memset(dst, 0, 64*sizeof(DCTELEM)); dst[0]= (src[0] + 4)>>3; for(i=1; i<64; i++){ int level= src[i]; if(((unsigned)(level+threshold1))>threshold2){ const int j= permutation[i]; if(level>0) dst[j]= (level - threshold1 + 4)>>3; else dst[j]= (level + threshold1 + 4)>>3; } } } #ifdef HAVE_MMX static void hardthresh_mmx(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){ int bias= 0; //FIXME unsigned int threshold1; threshold1= qp*((1<<4) - bias) - 1; asm volatile( #define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \ "movq " #src0 ", %%mm0 \n\t"\ "movq " #src1 ", %%mm1 \n\t"\ "movq " #src2 ", %%mm2 \n\t"\ "movq " #src3 ", %%mm3 \n\t"\ "psubw %%mm4, %%mm0 \n\t"\ "psubw %%mm4, %%mm1 \n\t"\ "psubw %%mm4, %%mm2 \n\t"\ "psubw %%mm4, %%mm3 \n\t"\ "paddusw %%mm5, %%mm0 \n\t"\ "paddusw %%mm5, %%mm1 \n\t"\ "paddusw %%mm5, %%mm2 \n\t"\ "paddusw %%mm5, %%mm3 \n\t"\ "paddw %%mm6, %%mm0 \n\t"\ "paddw %%mm6, %%mm1 \n\t"\ "paddw %%mm6, %%mm2 \n\t"\ "paddw %%mm6, %%mm3 \n\t"\ "psubusw %%mm6, %%mm0 \n\t"\ "psubusw %%mm6, %%mm1 \n\t"\ "psubusw %%mm6, %%mm2 \n\t"\ "psubusw %%mm6, %%mm3 \n\t"\ "psraw $3, %%mm0 \n\t"\ "psraw $3, %%mm1 \n\t"\ "psraw $3, %%mm2 \n\t"\ "psraw $3, %%mm3 \n\t"\ \ "movq %%mm0, %%mm7 \n\t"\ "punpcklwd %%mm2, %%mm0 \n\t" /*A*/\ "punpckhwd %%mm2, %%mm7 \n\t" /*C*/\ "movq %%mm1, %%mm2 \n\t"\ "punpcklwd %%mm3, %%mm1 \n\t" /*B*/\ "punpckhwd %%mm3, %%mm2 \n\t" /*D*/\ "movq %%mm0, %%mm3 \n\t"\ "punpcklwd %%mm1, %%mm0 \n\t" /*A*/\ "punpckhwd %%mm7, %%mm3 \n\t" /*C*/\ "punpcklwd %%mm2, %%mm7 \n\t" /*B*/\ "punpckhwd %%mm2, %%mm1 \n\t" /*D*/\ \ "movq %%mm0, " #dst0 " \n\t"\ "movq %%mm7, " #dst1 " \n\t"\ "movq %%mm3, " #dst2 " \n\t"\ "movq %%mm1, " #dst3 " \n\t" "movd %2, %%mm4 \n\t" "movd %3, %%mm5 \n\t" "movd %4, %%mm6 \n\t" "packssdw %%mm4, %%mm4 \n\t" "packssdw %%mm5, %%mm5 \n\t" "packssdw %%mm6, %%mm6 \n\t" "packssdw %%mm4, %%mm4 \n\t" "packssdw %%mm5, %%mm5 \n\t" "packssdw %%mm6, %%mm6 \n\t" REQUANT_CORE( (%1), 8(%1), 16(%1), 24(%1), (%0), 8(%0), 64(%0), 72(%0)) REQUANT_CORE(32(%1), 40(%1), 48(%1), 56(%1),16(%0),24(%0), 48(%0), 56(%0)) REQUANT_CORE(64(%1), 72(%1), 80(%1), 88(%1),32(%0),40(%0), 96(%0),104(%0)) REQUANT_CORE(96(%1),104(%1),112(%1),120(%1),80(%0),88(%0),112(%0),120(%0)) : : "r" (src), "r" (dst), "g" (threshold1+1), "g" (threshold1+5), "g" (threshold1-4) //FIXME maybe more accurate then needed? ); dst[0]= (src[0] + 4)>>3; } static void softthresh_mmx(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){ int bias= 0; //FIXME unsigned int threshold1; threshold1= qp*((1<<4) - bias) - 1; asm volatile( #undef REQUANT_CORE #define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \ "movq " #src0 ", %%mm0 \n\t"\ "movq " #src1 ", %%mm1 \n\t"\ "pxor %%mm6, %%mm6 \n\t"\ "pxor %%mm7, %%mm7 \n\t"\ "pcmpgtw %%mm0, %%mm6 \n\t"\ "pcmpgtw %%mm1, %%mm7 \n\t"\ "pxor %%mm6, %%mm0 \n\t"\ "pxor %%mm7, %%mm1 \n\t"\ "psubusw %%mm4, %%mm0 \n\t"\ "psubusw %%mm4, %%mm1 \n\t"\ "pxor %%mm6, %%mm0 \n\t"\ "pxor %%mm7, %%mm1 \n\t"\ "movq " #src2 ", %%mm2 \n\t"\ "movq " #src3 ", %%mm3 \n\t"\ "pxor %%mm6, %%mm6 \n\t"\ "pxor %%mm7, %%mm7 \n\t"\ "pcmpgtw %%mm2, %%mm6 \n\t"\ "pcmpgtw %%mm3, %%mm7 \n\t"\ "pxor %%mm6, %%mm2 \n\t"\ "pxor %%mm7, %%mm3 \n\t"\ "psubusw %%mm4, %%mm2 \n\t"\ "psubusw %%mm4, %%mm3 \n\t"\ "pxor %%mm6, %%mm2 \n\t"\ "pxor %%mm7, %%mm3 \n\t"\ \ "paddsw %%mm5, %%mm0 \n\t"\ "paddsw %%mm5, %%mm1 \n\t"\ "paddsw %%mm5, %%mm2 \n\t"\ "paddsw %%mm5, %%mm3 \n\t"\ "psraw $3, %%mm0 \n\t"\ "psraw $3, %%mm1 \n\t"\ "psraw $3, %%mm2 \n\t"\ "psraw $3, %%mm3 \n\t"\ \ "movq %%mm0, %%mm7 \n\t"\ "punpcklwd %%mm2, %%mm0 \n\t" /*A*/\ "punpckhwd %%mm2, %%mm7 \n\t" /*C*/\ "movq %%mm1, %%mm2 \n\t"\ "punpcklwd %%mm3, %%mm1 \n\t" /*B*/\ "punpckhwd %%mm3, %%mm2 \n\t" /*D*/\ "movq %%mm0, %%mm3 \n\t"\ "punpcklwd %%mm1, %%mm0 \n\t" /*A*/\ "punpckhwd %%mm7, %%mm3 \n\t" /*C*/\ "punpcklwd %%mm2, %%mm7 \n\t" /*B*/\ "punpckhwd %%mm2, %%mm1 \n\t" /*D*/\ \ "movq %%mm0, " #dst0 " \n\t"\ "movq %%mm7, " #dst1 " \n\t"\ "movq %%mm3, " #dst2 " \n\t"\ "movq %%mm1, " #dst3 " \n\t" "movd %2, %%mm4 \n\t" "movd %3, %%mm5 \n\t" "packssdw %%mm4, %%mm4 \n\t" "packssdw %%mm5, %%mm5 \n\t" "packssdw %%mm4, %%mm4 \n\t" "packssdw %%mm5, %%mm5 \n\t" REQUANT_CORE( (%1), 8(%1), 16(%1), 24(%1), (%0), 8(%0), 64(%0), 72(%0)) REQUANT_CORE(32(%1), 40(%1), 48(%1), 56(%1),16(%0),24(%0), 48(%0), 56(%0)) REQUANT_CORE(64(%1), 72(%1), 80(%1), 88(%1),32(%0),40(%0), 96(%0),104(%0)) REQUANT_CORE(96(%1),104(%1),112(%1),120(%1),80(%0),88(%0),112(%0),120(%0)) : : "r" (src), "r" (dst), "g" (threshold1), "rm" (4) //FIXME maybe more accurate then needed? ); dst[0]= (src[0] + 4)>>3; } #endif static inline void add_block(int16_t *dst, int stride, DCTELEM block[64]){ int y; for(y=0; y<8; y++){ *(uint32_t*)&dst[0 + y*stride]+= *(uint32_t*)&block[0 + y*8]; *(uint32_t*)&dst[2 + y*stride]+= *(uint32_t*)&block[2 + y*8]; *(uint32_t*)&dst[4 + y*stride]+= *(uint32_t*)&block[4 + y*8]; *(uint32_t*)&dst[6 + y*stride]+= *(uint32_t*)&block[6 + y*8]; } } static void store_slice_c(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale){ int y, x; #define STORE(pos) \ temp= ((src[x + y*src_stride + pos]<<log2_scale) + d[pos])>>6;\ if(temp & 0x100) temp= ~(temp>>31);\ dst[x + y*dst_stride + pos]= temp; for(y=0; y<height; y++){ const uint8_t *d= dither[y]; for(x=0; x<width; x+=8){ int temp; STORE(0); STORE(1); STORE(2); STORE(3); STORE(4); STORE(5); STORE(6); STORE(7); } } } #ifdef HAVE_MMX static void store_slice_mmx(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale){ int y; for(y=0; y<height; y++){ uint8_t *dst1= dst; int16_t *src1= src; asm volatile( "movq (%3), %%mm3 \n\t" "movq (%3), %%mm4 \n\t" "movd %4, %%mm2 \n\t" "pxor %%mm0, %%mm0 \n\t" "punpcklbw %%mm0, %%mm3 \n\t" "punpckhbw %%mm0, %%mm4 \n\t" "psraw %%mm2, %%mm3 \n\t" "psraw %%mm2, %%mm4 \n\t" "movd %5, %%mm2 \n\t" "1: \n\t" "movq (%0), %%mm0 \n\t" "movq 8(%0), %%mm1 \n\t" "paddw %%mm3, %%mm0 \n\t" "paddw %%mm4, %%mm1 \n\t" "psraw %%mm2, %%mm0 \n\t" "psraw %%mm2, %%mm1 \n\t" "packuswb %%mm1, %%mm0 \n\t" "movq %%mm0, (%1) \n\t" "add $16, %0 \n\t" "add $8, %1 \n\t" "cmp %2, %1 \n\t" " jb 1b \n\t" : "+r" (src1), "+r"(dst1) : "r"(dst + width), "r"(dither[y]), "g"(log2_scale), "g"(6-log2_scale) ); src += src_stride; dst += dst_stride; } // if(width != mmxw) // store_slice_c(dst + mmxw, src + mmxw, dst_stride, src_stride, width - mmxw, log2_scale); } #endif static void (*store_slice)(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale)= store_slice_c; static void (*requantize)(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation)= hardthresh_c; static void filter(struct vf_priv_s *p, uint8_t *dst, uint8_t *src, int dst_stride, int src_stride, int width, int height, uint8_t *qp_store, int qp_stride, int is_luma){ int x, y, i; const int count= 1<<p->log2_count; const int stride= is_luma ? p->temp_stride : ((width+16+15)&(~15)); uint64_t __attribute__((aligned(16))) block_align[32]; DCTELEM *block = (DCTELEM *)block_align; DCTELEM *block2= (DCTELEM *)(block_align+16); if (!src || !dst) return; // HACK avoid crash for Y8 colourspace for(y=0; y<height; y++){ int index= 8 + 8*stride + y*stride; fast_memcpy(p->src + index, src + y*src_stride, width); for(x=0; x<8; x++){ p->src[index - x - 1]= p->src[index + x ]; p->src[index + width + x ]= p->src[index + width - x - 1]; } } for(y=0; y<8; y++){ fast_memcpy(p->src + ( 7-y)*stride, p->src + ( y+8)*stride, stride); fast_memcpy(p->src + (height+8+y)*stride, p->src + (height-y+7)*stride, stride); } //FIXME (try edge emu) for(y=0; y<height+8; y+=8){ memset(p->temp + (8+y)*stride, 0, 8*stride*sizeof(int16_t)); for(x=0; x<width+8; x+=8){ const int qps= 3 + is_luma; int qp; if(p->qp) qp= p->qp; else{ qp= qp_store[ (XMIN(x, width-1)>>qps) + (XMIN(y, height-1)>>qps) * qp_stride]; if(p->mpeg2) qp>>=1; } for(i=0; i<count; i++){ const int x1= x + offset[i+count-1][0]; const int y1= y + offset[i+count-1][1]; const int index= x1 + y1*stride; p->dsp.get_pixels(block, p->src + index, stride); p->dsp.fdct(block); requantize(block2, block, qp, p->dsp.idct_permutation); p->dsp.idct(block2); add_block(p->temp + index, stride, block2); } } if(y) store_slice(dst + (y-8)*dst_stride, p->temp + 8 + y*stride, dst_stride, stride, width, XMIN(8, height+8-y), 6-p->log2_count); } #if 0 for(y=0; y<height; y++){ for(x=0; x<width; x++){ if((((x>>6) ^ (y>>6)) & 1) == 0) dst[x + y*dst_stride]= p->src[8 + 8*stride + x + y*stride]; if((x&63) == 0 || (y&63)==0) dst[x + y*dst_stride] += 128; } } #endif //FIXME reorder for better caching } static int config(struct vf_instance_s* vf, int width, int height, int d_width, int d_height, unsigned int flags, unsigned int outfmt){ int h= (height+16+15)&(~15); vf->priv->temp_stride= (width+16+15)&(~15); vf->priv->temp= malloc(vf->priv->temp_stride*h*sizeof(int16_t)); vf->priv->src = malloc(vf->priv->temp_stride*h*sizeof(uint8_t)); return vf_next_config(vf,width,height,d_width,d_height,flags,outfmt); } static void get_image(struct vf_instance_s* vf, mp_image_t *mpi){ if(mpi->flags&MP_IMGFLAG_PRESERVE) return; // don't change // ok, we can do pp in-place (or pp disabled): vf->dmpi=vf_get_image(vf->next,mpi->imgfmt, mpi->type, mpi->flags | MP_IMGFLAG_READABLE, mpi->width, mpi->height); mpi->planes[0]=vf->dmpi->planes[0]; mpi->stride[0]=vf->dmpi->stride[0]; mpi->width=vf->dmpi->width; if(mpi->flags&MP_IMGFLAG_PLANAR){ mpi->planes[1]=vf->dmpi->planes[1]; mpi->planes[2]=vf->dmpi->planes[2]; mpi->stride[1]=vf->dmpi->stride[1]; mpi->stride[2]=vf->dmpi->stride[2]; } mpi->flags|=MP_IMGFLAG_DIRECT; } static int put_image(struct vf_instance_s* vf, mp_image_t *mpi, double pts){ mp_image_t *dmpi; if(!(mpi->flags&MP_IMGFLAG_DIRECT)){ // no DR, so get a new image! hope we'll get DR buffer: dmpi=vf_get_image(vf->next,mpi->imgfmt, MP_IMGTYPE_TEMP, MP_IMGFLAG_ACCEPT_STRIDE|MP_IMGFLAG_PREFER_ALIGNED_STRIDE, mpi->width,mpi->height); vf_clone_mpi_attributes(dmpi, mpi); }else{ dmpi=vf->dmpi; } vf->priv->mpeg2= mpi->qscale_type; if(mpi->pict_type != 3 && mpi->qscale && !vf->priv->qp){ if(!vf->priv->non_b_qp) vf->priv->non_b_qp= malloc(mpi->qstride * ((mpi->h + 15) >> 4)); fast_memcpy(vf->priv->non_b_qp, mpi->qscale, mpi->qstride * ((mpi->h + 15) >> 4)); } if(vf->priv->log2_count || !(mpi->flags&MP_IMGFLAG_DIRECT)){ char *qp_tab= vf->priv->non_b_qp; if((vf->priv->mode&4) || !qp_tab) qp_tab= mpi->qscale; if(qp_tab || vf->priv->qp){ filter(vf->priv, dmpi->planes[0], mpi->planes[0], dmpi->stride[0], mpi->stride[0], mpi->w, mpi->h, qp_tab, mpi->qstride, 1); filter(vf->priv, dmpi->planes[1], mpi->planes[1], dmpi->stride[1], mpi->stride[1], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, qp_tab, mpi->qstride, 0); filter(vf->priv, dmpi->planes[2], mpi->planes[2], dmpi->stride[2], mpi->stride[2], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, qp_tab, mpi->qstride, 0); }else{ memcpy_pic(dmpi->planes[0], mpi->planes[0], mpi->w, mpi->h, dmpi->stride[0], mpi->stride[0]); memcpy_pic(dmpi->planes[1], mpi->planes[1], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, dmpi->stride[1], mpi->stride[1]); memcpy_pic(dmpi->planes[2], mpi->planes[2], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, dmpi->stride[2], mpi->stride[2]); } } #ifdef HAVE_MMX if(gCpuCaps.hasMMX) asm volatile ("emms\n\t"); #endif #ifdef HAVE_MMX2 if(gCpuCaps.hasMMX2) asm volatile ("sfence\n\t"); #endif return vf_next_put_image(vf,dmpi, pts); } static void uninit(struct vf_instance_s* vf){ if(!vf->priv) return; if(vf->priv->temp) free(vf->priv->temp); vf->priv->temp= NULL; if(vf->priv->src) free(vf->priv->src); vf->priv->src= NULL; if(vf->priv->avctx) free(vf->priv->avctx); vf->priv->avctx= NULL; if(vf->priv->non_b_qp) free(vf->priv->non_b_qp); vf->priv->non_b_qp= NULL; free(vf->priv); vf->priv=NULL; } //===========================================================================// static int query_format(struct vf_instance_s* vf, unsigned int fmt){ switch(fmt){ case IMGFMT_YVU9: case IMGFMT_IF09: case IMGFMT_YV12: case IMGFMT_I420: case IMGFMT_IYUV: case IMGFMT_CLPL: case IMGFMT_Y800: case IMGFMT_Y8: case IMGFMT_444P: case IMGFMT_422P: case IMGFMT_411P: return vf_next_query_format(vf,fmt); } return 0; } static unsigned int fmt_list[]={ IMGFMT_YVU9, IMGFMT_IF09, IMGFMT_YV12, IMGFMT_I420, IMGFMT_IYUV, IMGFMT_CLPL, IMGFMT_Y800, IMGFMT_Y8, IMGFMT_444P, IMGFMT_422P, IMGFMT_411P, 0 }; static int control(struct vf_instance_s* vf, int request, void* data){ switch(request){ case VFCTRL_QUERY_MAX_PP_LEVEL: return 6; case VFCTRL_SET_PP_LEVEL: vf->priv->log2_count= *((unsigned int*)data); return CONTROL_TRUE; } return vf_next_control(vf,request,data); } static int open(vf_instance_t *vf, char* args){ int log2c=-1; vf->config=config; vf->put_image=put_image; vf->get_image=get_image; vf->query_format=query_format; vf->uninit=uninit; vf->control= control; vf->priv=malloc(sizeof(struct vf_priv_s)); memset(vf->priv, 0, sizeof(struct vf_priv_s)); avcodec_init(); vf->priv->avctx= avcodec_alloc_context(); dsputil_init(&vf->priv->dsp, vf->priv->avctx); vf->priv->log2_count= 3; if (args) sscanf(args, "%d:%d:%d", &log2c, &vf->priv->qp, &vf->priv->mode); if( log2c >=0 && log2c <=6 ) vf->priv->log2_count = log2c; if(vf->priv->qp < 0) vf->priv->qp = 0; switch(vf->priv->mode&3){ default: case 0: requantize= hardthresh_c; break; case 1: requantize= softthresh_c; break; } #ifdef HAVE_MMX if(gCpuCaps.hasMMX){ store_slice= store_slice_mmx; switch(vf->priv->mode&3){ case 0: requantize= hardthresh_mmx; break; case 1: requantize= softthresh_mmx; break; } } #endif return 1; } vf_info_t vf_info_spp = { "simple postprocess", "spp", "Michael Niedermayer", "", open, NULL };