view libmpcodecs/vf_spp.c @ 26273:8c92e7e4ccd9

command.h: Remove unnecessary includes Remove #include of "mp_core.h" and "input/input.h". Their only use was that functions declared in command.h took pointers to structs defined in those headers. Declare the structs directly as incomplete types instead.
author uau
date Mon, 31 Mar 2008 04:10:54 +0000
parents 00fff9a3b735
children 82601a38e2a7
line wrap: on
line source

/*
    Copyright (C) 2003 Michael Niedermayer <michaelni@gmx.at>

    This program is free software; you can redistribute it and/or modify
    it under the terms of the GNU General Public License as published by
    the Free Software Foundation; either version 2 of the License, or
    (at your option) any later version.

    This program is distributed in the hope that it will be useful,
    but WITHOUT ANY WARRANTY; without even the implied warranty of
    MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
    GNU General Public License for more details.

    You should have received a copy of the GNU General Public License
    along with this program; if not, write to the Free Software
    Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
*/

/*
 * This implementation is based on an algorithm described in
 * "Aria Nosratinia Embedded Post-Processing for 
 * Enhancement of Compressed Images (1999)"
 * (http://citeseer.nj.nec.com/nosratinia99embedded.html)
 */

 
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <inttypes.h>
#include <math.h>

#include "config.h"

#include "mp_msg.h"
#include "cpudetect.h"

#include "libavutil/intreadwrite.h"
#include "libavcodec/avcodec.h"
#include "libavcodec/dsputil.h"

#ifdef HAVE_MALLOC_H
#include <malloc.h>
#endif

#include "img_format.h"
#include "mp_image.h"
#include "vf.h"
#include "libvo/fastmemcpy.h"

#define XMIN(a,b) ((a) < (b) ? (a) : (b))

//===========================================================================//
static const uint8_t  __attribute__((aligned(8))) dither[8][8]={
{  0,  48,  12,  60,   3,  51,  15,  63, },
{ 32,  16,  44,  28,  35,  19,  47,  31, },
{  8,  56,   4,  52,  11,  59,   7,  55, },
{ 40,  24,  36,  20,  43,  27,  39,  23, },
{  2,  50,  14,  62,   1,  49,  13,  61, },
{ 34,  18,  46,  30,  33,  17,  45,  29, },
{ 10,  58,   6,  54,   9,  57,   5,  53, },
{ 42,  26,  38,  22,  41,  25,  37,  21, },
};

static const uint8_t offset[127][2]= {
{0,0},
{0,0}, {4,4},
{0,0}, {2,2}, {6,4}, {4,6},
{0,0}, {5,1}, {2,2}, {7,3}, {4,4}, {1,5}, {6,6}, {3,7}, 

{0,0}, {4,0}, {1,1}, {5,1}, {3,2}, {7,2}, {2,3}, {6,3}, 
{0,4}, {4,4}, {1,5}, {5,5}, {3,6}, {7,6}, {2,7}, {6,7}, 

{0,0}, {0,2}, {0,4}, {0,6}, {1,1}, {1,3}, {1,5}, {1,7}, 
{2,0}, {2,2}, {2,4}, {2,6}, {3,1}, {3,3}, {3,5}, {3,7}, 
{4,0}, {4,2}, {4,4}, {4,6}, {5,1}, {5,3}, {5,5}, {5,7}, 
{6,0}, {6,2}, {6,4}, {6,6}, {7,1}, {7,3}, {7,5}, {7,7}, 

{0,0}, {4,4}, {0,4}, {4,0}, {2,2}, {6,6}, {2,6}, {6,2}, 
{0,2}, {4,6}, {0,6}, {4,2}, {2,0}, {6,4}, {2,4}, {6,0}, 
{1,1}, {5,5}, {1,5}, {5,1}, {3,3}, {7,7}, {3,7}, {7,3}, 
{1,3}, {5,7}, {1,7}, {5,3}, {3,1}, {7,5}, {3,5}, {7,1}, 
{0,1}, {4,5}, {0,5}, {4,1}, {2,3}, {6,7}, {2,7}, {6,3}, 
{0,3}, {4,7}, {0,7}, {4,3}, {2,1}, {6,5}, {2,5}, {6,1}, 
{1,0}, {5,4}, {1,4}, {5,0}, {3,2}, {7,6}, {3,6}, {7,2}, 
{1,2}, {5,6}, {1,6}, {5,2}, {3,0}, {7,4}, {3,4}, {7,0},
};

struct vf_priv_s {
	int log2_count;
	int qp;
	int mode;
	int mpeg2;
	int temp_stride;
	uint8_t *src;
	int16_t *temp;
	AVCodecContext *avctx;
	DSPContext dsp;
        char *non_b_qp;
};

#define SHIFT 22

static void hardthresh_c(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){
	int i; 
	int bias= 0; //FIXME
	unsigned int threshold1, threshold2;
	
	threshold1= qp*((1<<4) - bias) - 1;
	threshold2= (threshold1<<1);
        
	memset(dst, 0, 64*sizeof(DCTELEM));
	dst[0]= (src[0] + 4)>>3;

        for(i=1; i<64; i++){
		int level= src[i];
		if(((unsigned)(level+threshold1))>threshold2){
			const int j= permutation[i];
			dst[j]= (level + 4)>>3;
		}
	}
}

static void softthresh_c(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){
	int i; 
	int bias= 0; //FIXME
	unsigned int threshold1, threshold2;
	
	threshold1= qp*((1<<4) - bias) - 1;
	threshold2= (threshold1<<1);
        
	memset(dst, 0, 64*sizeof(DCTELEM));
	dst[0]= (src[0] + 4)>>3;

        for(i=1; i<64; i++){
		int level= src[i];
		if(((unsigned)(level+threshold1))>threshold2){
			const int j= permutation[i];
			if(level>0)
				dst[j]= (level - threshold1 + 4)>>3;
			else
				dst[j]= (level + threshold1 + 4)>>3;
		}
	}
}

#ifdef HAVE_MMX
static void hardthresh_mmx(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){
	int bias= 0; //FIXME
	unsigned int threshold1;
	
	threshold1= qp*((1<<4) - bias) - 1;
	
        asm volatile(
#define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \
		"movq " #src0 ", %%mm0	\n\t"\
		"movq " #src1 ", %%mm1	\n\t"\
		"movq " #src2 ", %%mm2	\n\t"\
		"movq " #src3 ", %%mm3	\n\t"\
		"psubw %%mm4, %%mm0	\n\t"\
		"psubw %%mm4, %%mm1	\n\t"\
		"psubw %%mm4, %%mm2	\n\t"\
		"psubw %%mm4, %%mm3	\n\t"\
		"paddusw %%mm5, %%mm0	\n\t"\
		"paddusw %%mm5, %%mm1	\n\t"\
		"paddusw %%mm5, %%mm2	\n\t"\
		"paddusw %%mm5, %%mm3	\n\t"\
		"paddw %%mm6, %%mm0	\n\t"\
		"paddw %%mm6, %%mm1	\n\t"\
		"paddw %%mm6, %%mm2	\n\t"\
		"paddw %%mm6, %%mm3	\n\t"\
		"psubusw %%mm6, %%mm0	\n\t"\
		"psubusw %%mm6, %%mm1	\n\t"\
		"psubusw %%mm6, %%mm2	\n\t"\
		"psubusw %%mm6, %%mm3	\n\t"\
		"psraw $3, %%mm0	\n\t"\
		"psraw $3, %%mm1	\n\t"\
		"psraw $3, %%mm2	\n\t"\
		"psraw $3, %%mm3	\n\t"\
\
		"movq %%mm0, %%mm7	\n\t"\
		"punpcklwd %%mm2, %%mm0	\n\t" /*A*/\
		"punpckhwd %%mm2, %%mm7	\n\t" /*C*/\
		"movq %%mm1, %%mm2	\n\t"\
		"punpcklwd %%mm3, %%mm1	\n\t" /*B*/\
		"punpckhwd %%mm3, %%mm2	\n\t" /*D*/\
		"movq %%mm0, %%mm3	\n\t"\
		"punpcklwd %%mm1, %%mm0	\n\t" /*A*/\
		"punpckhwd %%mm7, %%mm3	\n\t" /*C*/\
		"punpcklwd %%mm2, %%mm7	\n\t" /*B*/\
		"punpckhwd %%mm2, %%mm1	\n\t" /*D*/\
\
		"movq %%mm0, " #dst0 "	\n\t"\
		"movq %%mm7, " #dst1 "	\n\t"\
		"movq %%mm3, " #dst2 "	\n\t"\
		"movq %%mm1, " #dst3 "	\n\t"
                
		"movd %2, %%mm4		\n\t"
		"movd %3, %%mm5		\n\t"
		"movd %4, %%mm6		\n\t"
		"packssdw %%mm4, %%mm4	\n\t"
		"packssdw %%mm5, %%mm5	\n\t"
		"packssdw %%mm6, %%mm6	\n\t"
		"packssdw %%mm4, %%mm4	\n\t"
		"packssdw %%mm5, %%mm5	\n\t"
		"packssdw %%mm6, %%mm6	\n\t"
		REQUANT_CORE(  (%1),  8(%1), 16(%1), 24(%1),  (%0), 8(%0), 64(%0), 72(%0))
		REQUANT_CORE(32(%1), 40(%1), 48(%1), 56(%1),16(%0),24(%0), 48(%0), 56(%0))
		REQUANT_CORE(64(%1), 72(%1), 80(%1), 88(%1),32(%0),40(%0), 96(%0),104(%0))
		REQUANT_CORE(96(%1),104(%1),112(%1),120(%1),80(%0),88(%0),112(%0),120(%0))
		: : "r" (src), "r" (dst), "g" (threshold1+1), "g" (threshold1+5), "g" (threshold1-4) //FIXME maybe more accurate then needed?
	);
	dst[0]= (src[0] + 4)>>3;
}

static void softthresh_mmx(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation){
	int bias= 0; //FIXME
	unsigned int threshold1;

	threshold1= qp*((1<<4) - bias) - 1;
	
        asm volatile(
#undef REQUANT_CORE
#define REQUANT_CORE(dst0, dst1, dst2, dst3, src0, src1, src2, src3) \
		"movq " #src0 ", %%mm0	\n\t"\
		"movq " #src1 ", %%mm1	\n\t"\
		"pxor %%mm6, %%mm6	\n\t"\
		"pxor %%mm7, %%mm7	\n\t"\
		"pcmpgtw %%mm0, %%mm6	\n\t"\
		"pcmpgtw %%mm1, %%mm7	\n\t"\
		"pxor %%mm6, %%mm0	\n\t"\
		"pxor %%mm7, %%mm1	\n\t"\
		"psubusw %%mm4, %%mm0	\n\t"\
		"psubusw %%mm4, %%mm1	\n\t"\
		"pxor %%mm6, %%mm0	\n\t"\
		"pxor %%mm7, %%mm1	\n\t"\
		"movq " #src2 ", %%mm2	\n\t"\
		"movq " #src3 ", %%mm3	\n\t"\
		"pxor %%mm6, %%mm6	\n\t"\
		"pxor %%mm7, %%mm7	\n\t"\
		"pcmpgtw %%mm2, %%mm6	\n\t"\
		"pcmpgtw %%mm3, %%mm7	\n\t"\
		"pxor %%mm6, %%mm2	\n\t"\
		"pxor %%mm7, %%mm3	\n\t"\
		"psubusw %%mm4, %%mm2	\n\t"\
		"psubusw %%mm4, %%mm3	\n\t"\
		"pxor %%mm6, %%mm2	\n\t"\
		"pxor %%mm7, %%mm3	\n\t"\
\
		"paddsw %%mm5, %%mm0	\n\t"\
		"paddsw %%mm5, %%mm1	\n\t"\
		"paddsw %%mm5, %%mm2	\n\t"\
		"paddsw %%mm5, %%mm3	\n\t"\
		"psraw $3, %%mm0	\n\t"\
		"psraw $3, %%mm1	\n\t"\
		"psraw $3, %%mm2	\n\t"\
		"psraw $3, %%mm3	\n\t"\
\
		"movq %%mm0, %%mm7	\n\t"\
		"punpcklwd %%mm2, %%mm0	\n\t" /*A*/\
		"punpckhwd %%mm2, %%mm7	\n\t" /*C*/\
		"movq %%mm1, %%mm2	\n\t"\
		"punpcklwd %%mm3, %%mm1	\n\t" /*B*/\
		"punpckhwd %%mm3, %%mm2	\n\t" /*D*/\
		"movq %%mm0, %%mm3	\n\t"\
		"punpcklwd %%mm1, %%mm0	\n\t" /*A*/\
		"punpckhwd %%mm7, %%mm3	\n\t" /*C*/\
		"punpcklwd %%mm2, %%mm7	\n\t" /*B*/\
		"punpckhwd %%mm2, %%mm1	\n\t" /*D*/\
\
		"movq %%mm0, " #dst0 "	\n\t"\
		"movq %%mm7, " #dst1 "	\n\t"\
		"movq %%mm3, " #dst2 "	\n\t"\
		"movq %%mm1, " #dst3 "	\n\t"
                
		"movd %2, %%mm4		\n\t"
		"movd %3, %%mm5		\n\t"
		"packssdw %%mm4, %%mm4	\n\t"
		"packssdw %%mm5, %%mm5	\n\t"
		"packssdw %%mm4, %%mm4	\n\t"
		"packssdw %%mm5, %%mm5	\n\t"
		REQUANT_CORE(  (%1),  8(%1), 16(%1), 24(%1),  (%0), 8(%0), 64(%0), 72(%0))
		REQUANT_CORE(32(%1), 40(%1), 48(%1), 56(%1),16(%0),24(%0), 48(%0), 56(%0))
		REQUANT_CORE(64(%1), 72(%1), 80(%1), 88(%1),32(%0),40(%0), 96(%0),104(%0))
		REQUANT_CORE(96(%1),104(%1),112(%1),120(%1),80(%0),88(%0),112(%0),120(%0))
		: : "r" (src), "r" (dst), "g" (threshold1), "rm" (4) //FIXME maybe more accurate then needed?
	);

	dst[0]= (src[0] + 4)>>3;
}
#endif

static inline void add_block(int16_t *dst, int stride, DCTELEM block[64]){
	int y;
	
	for(y=0; y<8; y++){
		*(uint32_t*)&dst[0 + y*stride]+= *(uint32_t*)&block[0 + y*8];
		*(uint32_t*)&dst[2 + y*stride]+= *(uint32_t*)&block[2 + y*8];
		*(uint32_t*)&dst[4 + y*stride]+= *(uint32_t*)&block[4 + y*8];
		*(uint32_t*)&dst[6 + y*stride]+= *(uint32_t*)&block[6 + y*8];
	}
}

static void store_slice_c(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale){
	int y, x;

#define STORE(pos) \
	temp= ((src[x + y*src_stride + pos]<<log2_scale) + d[pos])>>6;\
	if(temp & 0x100) temp= ~(temp>>31);\
	dst[x + y*dst_stride + pos]= temp;

	for(y=0; y<height; y++){
		const uint8_t *d= dither[y];
		for(x=0; x<width; x+=8){
			int temp;
			STORE(0);
			STORE(1);
			STORE(2);
			STORE(3);
			STORE(4);
			STORE(5);
			STORE(6);
			STORE(7);
		}
	}
}

#ifdef HAVE_MMX
static void store_slice_mmx(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale){
	int y;

	for(y=0; y<height; y++){
		uint8_t *dst1= dst;
		int16_t *src1= src;
		asm volatile(
			"movq (%3), %%mm3	\n\t"
			"movq (%3), %%mm4	\n\t"
			"movd %4, %%mm2		\n\t"
			"pxor %%mm0, %%mm0	\n\t"
			"punpcklbw %%mm0, %%mm3	\n\t"
			"punpckhbw %%mm0, %%mm4	\n\t"
			"psraw %%mm2, %%mm3	\n\t"
			"psraw %%mm2, %%mm4	\n\t"
			"movd %5, %%mm2		\n\t"
			"1:			\n\t"
			"movq (%0), %%mm0	\n\t"
			"movq 8(%0), %%mm1	\n\t"
			"paddw %%mm3, %%mm0	\n\t"
			"paddw %%mm4, %%mm1	\n\t"
			"psraw %%mm2, %%mm0	\n\t"
			"psraw %%mm2, %%mm1	\n\t"
			"packuswb %%mm1, %%mm0	\n\t"
			"movq %%mm0, (%1) 	\n\t"
			"add $16, %0		\n\t"
			"add $8, %1		\n\t"
			"cmp %2, %1		\n\t"
			" jb 1b			\n\t"
			: "+r" (src1), "+r"(dst1)
			: "r"(dst + width), "r"(dither[y]), "g"(log2_scale), "g"(6-log2_scale)
		);
		src += src_stride;
		dst += dst_stride;
	}
//	if(width != mmxw)
//		store_slice_c(dst + mmxw, src + mmxw, dst_stride, src_stride, width - mmxw, log2_scale);
}
#endif

static void (*store_slice)(uint8_t *dst, int16_t *src, int dst_stride, int src_stride, int width, int height, int log2_scale)= store_slice_c;

static void (*requantize)(DCTELEM dst[64], DCTELEM src[64], int qp, uint8_t *permutation)= hardthresh_c;

static void filter(struct vf_priv_s *p, uint8_t *dst, uint8_t *src, int dst_stride, int src_stride, int width, int height, uint8_t *qp_store, int qp_stride, int is_luma){
	int x, y, i;
	const int count= 1<<p->log2_count;
	const int stride= is_luma ? p->temp_stride : ((width+16+15)&(~15));
	uint64_t __attribute__((aligned(16))) block_align[32];
	DCTELEM *block = (DCTELEM *)block_align;
	DCTELEM *block2= (DCTELEM *)(block_align+16);

	if (!src || !dst) return; // HACK avoid crash for Y8 colourspace
	for(y=0; y<height; y++){
		int index= 8 + 8*stride + y*stride;
		fast_memcpy(p->src + index, src + y*src_stride, width);
		for(x=0; x<8; x++){ 
			p->src[index         - x - 1]= p->src[index +         x    ];
			p->src[index + width + x    ]= p->src[index + width - x - 1];
		}
	}
	for(y=0; y<8; y++){
		fast_memcpy(p->src + (      7-y)*stride, p->src + (      y+8)*stride, stride);
		fast_memcpy(p->src + (height+8+y)*stride, p->src + (height-y+7)*stride, stride);
	}
	//FIXME (try edge emu)

	for(y=0; y<height+8; y+=8){
		memset(p->temp + (8+y)*stride, 0, 8*stride*sizeof(int16_t));
		for(x=0; x<width+8; x+=8){
			const int qps= 3 + is_luma;
			int qp;
                        
			if(p->qp)
				qp= p->qp;
			else{
				qp= qp_store[ (XMIN(x, width-1)>>qps) + (XMIN(y, height-1)>>qps) * qp_stride];
				if(p->mpeg2) qp = FFMAX(1, qp>>1);
			}
			for(i=0; i<count; i++){
				const int x1= x + offset[i+count-1][0];
				const int y1= y + offset[i+count-1][1];
				const int index= x1 + y1*stride;
				p->dsp.get_pixels(block, p->src + index, stride);
				p->dsp.fdct(block);
				requantize(block2, block, qp, p->dsp.idct_permutation);
				p->dsp.idct(block2);
				add_block(p->temp + index, stride, block2);
			}
		}
		if(y)
			store_slice(dst + (y-8)*dst_stride, p->temp + 8 + y*stride, dst_stride, stride, width, XMIN(8, height+8-y), 6-p->log2_count);
	}
#if 0
	for(y=0; y<height; y++){
		for(x=0; x<width; x++){
			if((((x>>6) ^ (y>>6)) & 1) == 0)
				dst[x + y*dst_stride]= p->src[8 + 8*stride  + x + y*stride];
			if((x&63) == 0 || (y&63)==0)
				dst[x + y*dst_stride] += 128;
                }
	}
#endif
	//FIXME reorder for better caching
}

static int config(struct vf_instance_s* vf,
        int width, int height, int d_width, int d_height,
	unsigned int flags, unsigned int outfmt){
	int h= (height+16+15)&(~15);

	vf->priv->temp_stride= (width+16+15)&(~15);
        vf->priv->temp= malloc(vf->priv->temp_stride*h*sizeof(int16_t));
        vf->priv->src = malloc(vf->priv->temp_stride*h*sizeof(uint8_t));
        
	return vf_next_config(vf,width,height,d_width,d_height,flags,outfmt);
}

static void get_image(struct vf_instance_s* vf, mp_image_t *mpi){
    if(mpi->flags&MP_IMGFLAG_PRESERVE) return; // don't change
    // ok, we can do pp in-place (or pp disabled):
    vf->dmpi=vf_get_image(vf->next,mpi->imgfmt,
        mpi->type, mpi->flags | MP_IMGFLAG_READABLE, mpi->width, mpi->height);
    mpi->planes[0]=vf->dmpi->planes[0];
    mpi->stride[0]=vf->dmpi->stride[0];
    mpi->width=vf->dmpi->width;
    if(mpi->flags&MP_IMGFLAG_PLANAR){
        mpi->planes[1]=vf->dmpi->planes[1];
        mpi->planes[2]=vf->dmpi->planes[2];
	mpi->stride[1]=vf->dmpi->stride[1];
	mpi->stride[2]=vf->dmpi->stride[2];
    }
    mpi->flags|=MP_IMGFLAG_DIRECT;
}

static int put_image(struct vf_instance_s* vf, mp_image_t *mpi, double pts){
	mp_image_t *dmpi;

	if(!(mpi->flags&MP_IMGFLAG_DIRECT)){
		// no DR, so get a new image! hope we'll get DR buffer:
                dmpi=vf_get_image(vf->next,mpi->imgfmt,
                    MP_IMGTYPE_TEMP,
                    MP_IMGFLAG_ACCEPT_STRIDE|MP_IMGFLAG_PREFER_ALIGNED_STRIDE,
                    mpi->width,mpi->height);
                vf_clone_mpi_attributes(dmpi, mpi);
        }else{
           dmpi=vf->dmpi;
        }

        vf->priv->mpeg2= mpi->qscale_type;
        if(mpi->pict_type != 3 && mpi->qscale && !vf->priv->qp){
            if(!vf->priv->non_b_qp)
                vf->priv->non_b_qp= malloc(mpi->qstride * ((mpi->h + 15) >> 4));
            fast_memcpy(vf->priv->non_b_qp, mpi->qscale, mpi->qstride * ((mpi->h + 15) >> 4));
        }
	if(vf->priv->log2_count || !(mpi->flags&MP_IMGFLAG_DIRECT)){
            char *qp_tab= vf->priv->non_b_qp;
            if((vf->priv->mode&4) || !qp_tab)
                qp_tab= mpi->qscale;

	    if(qp_tab || vf->priv->qp){
		filter(vf->priv, dmpi->planes[0], mpi->planes[0], dmpi->stride[0], mpi->stride[0], mpi->w, mpi->h, qp_tab, mpi->qstride, 1);
		filter(vf->priv, dmpi->planes[1], mpi->planes[1], dmpi->stride[1], mpi->stride[1], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, qp_tab, mpi->qstride, 0);
		filter(vf->priv, dmpi->planes[2], mpi->planes[2], dmpi->stride[2], mpi->stride[2], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, qp_tab, mpi->qstride, 0);
	    }else{
		memcpy_pic(dmpi->planes[0], mpi->planes[0], mpi->w, mpi->h, dmpi->stride[0], mpi->stride[0]);
		memcpy_pic(dmpi->planes[1], mpi->planes[1], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, dmpi->stride[1], mpi->stride[1]);
		memcpy_pic(dmpi->planes[2], mpi->planes[2], mpi->w>>mpi->chroma_x_shift, mpi->h>>mpi->chroma_y_shift, dmpi->stride[2], mpi->stride[2]);
	    }
	}

#ifdef HAVE_MMX
	if(gCpuCaps.hasMMX) asm volatile ("emms\n\t");
#endif
#ifdef HAVE_MMX2
	if(gCpuCaps.hasMMX2) asm volatile ("sfence\n\t");
#endif

	return vf_next_put_image(vf,dmpi, pts);
}

static void uninit(struct vf_instance_s* vf){
	if(!vf->priv) return;

	if(vf->priv->temp) free(vf->priv->temp);
	vf->priv->temp= NULL;
	if(vf->priv->src) free(vf->priv->src);
	vf->priv->src= NULL;
        if(vf->priv->avctx) free(vf->priv->avctx);
        vf->priv->avctx= NULL;
        if(vf->priv->non_b_qp) free(vf->priv->non_b_qp);
        vf->priv->non_b_qp= NULL;
	
	free(vf->priv);
	vf->priv=NULL;
}

//===========================================================================//
static int query_format(struct vf_instance_s* vf, unsigned int fmt){
    switch(fmt){
	case IMGFMT_YVU9:
	case IMGFMT_IF09:
	case IMGFMT_YV12:
	case IMGFMT_I420:
	case IMGFMT_IYUV:
	case IMGFMT_CLPL:
	case IMGFMT_Y800:
	case IMGFMT_Y8:
	case IMGFMT_444P:
	case IMGFMT_422P:
	case IMGFMT_411P:
	    return vf_next_query_format(vf,fmt);
    }
    return 0;
}

static int control(struct vf_instance_s* vf, int request, void* data){
    switch(request){
    case VFCTRL_QUERY_MAX_PP_LEVEL:
	return 6;
    case VFCTRL_SET_PP_LEVEL:
	vf->priv->log2_count= *((unsigned int*)data);
	return CONTROL_TRUE;
    }
    return vf_next_control(vf,request,data);
}

static int open(vf_instance_t *vf, char* args){

    int log2c=-1;
    
    vf->config=config;
    vf->put_image=put_image;
    vf->get_image=get_image;
    vf->query_format=query_format;
    vf->uninit=uninit;
    vf->control= control;
    vf->priv=malloc(sizeof(struct vf_priv_s));
    memset(vf->priv, 0, sizeof(struct vf_priv_s));
    
    avcodec_init();

    vf->priv->avctx= avcodec_alloc_context();
    dsputil_init(&vf->priv->dsp, vf->priv->avctx);
    
    vf->priv->log2_count= 3;
    
    if (args) sscanf(args, "%d:%d:%d", &log2c, &vf->priv->qp, &vf->priv->mode);

    if( log2c >=0 && log2c <=6 )
        vf->priv->log2_count = log2c;

    if(vf->priv->qp < 0)
        vf->priv->qp = 0;

    switch(vf->priv->mode&3){
        default:
	case 0: requantize= hardthresh_c; break;
	case 1: requantize= softthresh_c; break;
    }

#ifdef HAVE_MMX
    if(gCpuCaps.hasMMX){
	store_slice= store_slice_mmx;
	switch(vf->priv->mode&3){
	    case 0: requantize= hardthresh_mmx; break;
	    case 1: requantize= softthresh_mmx; break;
	}
    }
#endif
    
    return 1;
}

const vf_info_t vf_info_spp = {
    "simple postprocess",
    "spp",
    "Michael Niedermayer",
    "",
    open,
    NULL
};