Mercurial > libavformat.hg
view grab.c @ 2357:e9d7db467cf0 libavformat
merge some REGISTER_MUXER/DEMUXER into REGISTER_MUXDEMUX
author | aurel |
---|---|
date | Fri, 10 Aug 2007 01:44:50 +0000 |
parents | b21c2af60bc9 |
children | 7d80b6e8adec |
line wrap: on
line source
/* * Linux video grab interface * Copyright (c) 2000,2001 Fabrice Bellard. * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "avformat.h" #include "dsputil.h" #include <unistd.h> #include <fcntl.h> #include <sys/ioctl.h> #include <sys/mman.h> #include <sys/time.h> #define _LINUX_TIME_H 1 #include <linux/videodev.h> #include <time.h> typedef struct { int fd; int frame_format; /* see VIDEO_PALETTE_xxx */ int use_mmap; int width, height; int frame_rate; int frame_rate_base; int64_t time_frame; int frame_size; struct video_capability video_cap; struct video_audio audio_saved; uint8_t *video_buf; struct video_mbuf gb_buffers; struct video_mmap gb_buf; int gb_frame; /* ATI All In Wonder specific stuff */ /* XXX: remove and merge in libavcodec/imgconvert.c */ int aiw_enabled; int deint; int halfw; uint8_t *src_mem; uint8_t *lum_m4_mem; } VideoData; struct { int palette; int depth; enum PixelFormat pix_fmt; } video_formats [] = { {.palette = VIDEO_PALETTE_YUV420P, .depth = 12, .pix_fmt = PIX_FMT_YUV420P }, {.palette = VIDEO_PALETTE_YUV422, .depth = 16, .pix_fmt = PIX_FMT_YUYV422 }, {.palette = VIDEO_PALETTE_UYVY, .depth = 16, .pix_fmt = PIX_FMT_UYVY422 }, {.palette = VIDEO_PALETTE_YUYV, .depth = 16, .pix_fmt = PIX_FMT_YUYV422 }, /* NOTE: v4l uses BGR24, not RGB24 */ {.palette = VIDEO_PALETTE_RGB24, .depth = 24, .pix_fmt = PIX_FMT_BGR24 }, {.palette = VIDEO_PALETTE_RGB565, .depth = 16, .pix_fmt = PIX_FMT_BGR565 }, {.palette = VIDEO_PALETTE_GREY, .depth = 8, .pix_fmt = PIX_FMT_GRAY8 }, }; static int aiw_init(VideoData *s); static int aiw_read_picture(VideoData *s, uint8_t *data); static int aiw_close(VideoData *s); static int grab_read_header(AVFormatContext *s1, AVFormatParameters *ap) { VideoData *s = s1->priv_data; AVStream *st; int width, height; int video_fd, frame_size; int ret, frame_rate, frame_rate_base; int desired_palette, desired_depth; struct video_tuner tuner; struct video_audio audio; struct video_picture pict; int j; int vformat_num = sizeof(video_formats) / sizeof(video_formats[0]); if (ap->width <= 0 || ap->height <= 0 || ap->time_base.den <= 0) { av_log(s1, AV_LOG_ERROR, "Bad capture size (%dx%d) or wrong time base (%d)\n", ap->width, ap->height, ap->time_base.den); return -1; } width = ap->width; height = ap->height; frame_rate = ap->time_base.den; frame_rate_base = ap->time_base.num; if((unsigned)width > 32767 || (unsigned)height > 32767) { av_log(s1, AV_LOG_ERROR, "Capture size is out of range: %dx%d\n", width, height); return -1; } st = av_new_stream(s1, 0); if (!st) return AVERROR(ENOMEM); av_set_pts_info(st, 64, 1, 1000000); /* 64 bits pts in us */ s->width = width; s->height = height; s->frame_rate = frame_rate; s->frame_rate_base = frame_rate_base; video_fd = open(s1->filename, O_RDWR); if (video_fd < 0) { perror(s1->filename); goto fail; } if (ioctl(video_fd,VIDIOCGCAP, &s->video_cap) < 0) { perror("VIDIOCGCAP"); goto fail; } if (!(s->video_cap.type & VID_TYPE_CAPTURE)) { av_log(s1, AV_LOG_ERROR, "Fatal: grab device does not handle capture\n"); goto fail; } desired_palette = -1; desired_depth = -1; for (j = 0; j < vformat_num; j++) { if (ap->pix_fmt == video_formats[j].pix_fmt) { desired_palette = video_formats[j].palette; desired_depth = video_formats[j].depth; break; } } /* set tv standard */ if (ap->standard && !ioctl(video_fd, VIDIOCGTUNER, &tuner)) { if (!strcasecmp(ap->standard, "pal")) tuner.mode = VIDEO_MODE_PAL; else if (!strcasecmp(ap->standard, "secam")) tuner.mode = VIDEO_MODE_SECAM; else tuner.mode = VIDEO_MODE_NTSC; ioctl(video_fd, VIDIOCSTUNER, &tuner); } /* unmute audio */ audio.audio = 0; ioctl(video_fd, VIDIOCGAUDIO, &audio); memcpy(&s->audio_saved, &audio, sizeof(audio)); audio.flags &= ~VIDEO_AUDIO_MUTE; ioctl(video_fd, VIDIOCSAUDIO, &audio); ioctl(video_fd, VIDIOCGPICT, &pict); #if 0 printf("v4l: colour=%d hue=%d brightness=%d constrast=%d whiteness=%d\n", pict.colour, pict.hue, pict.brightness, pict.contrast, pict.whiteness); #endif /* try to choose a suitable video format */ pict.palette = desired_palette; pict.depth= desired_depth; if (desired_palette == -1 || (ret = ioctl(video_fd, VIDIOCSPICT, &pict)) < 0) { for (j = 0; j < vformat_num; j++) { pict.palette = video_formats[j].palette; pict.depth = video_formats[j].depth; if (-1 != ioctl(video_fd, VIDIOCSPICT, &pict)) break; } if (j >= vformat_num) goto fail1; } ret = ioctl(video_fd,VIDIOCGMBUF,&s->gb_buffers); if (ret < 0) { /* try to use read based access */ struct video_window win; int val; win.x = 0; win.y = 0; win.width = width; win.height = height; win.chromakey = -1; win.flags = 0; ioctl(video_fd, VIDIOCSWIN, &win); s->frame_format = pict.palette; val = 1; ioctl(video_fd, VIDIOCCAPTURE, &val); s->time_frame = av_gettime() * s->frame_rate / s->frame_rate_base; s->use_mmap = 0; /* ATI All In Wonder automatic activation */ if (!strcmp(s->video_cap.name, "Km")) { if (aiw_init(s) < 0) goto fail; s->aiw_enabled = 1; /* force 420P format because conversion from YUV422 to YUV420P is done in this driver (ugly) */ s->frame_format = VIDEO_PALETTE_YUV420P; } } else { s->video_buf = mmap(0,s->gb_buffers.size,PROT_READ|PROT_WRITE,MAP_SHARED,video_fd,0); if ((unsigned char*)-1 == s->video_buf) { s->video_buf = mmap(0,s->gb_buffers.size,PROT_READ|PROT_WRITE,MAP_PRIVATE,video_fd,0); if ((unsigned char*)-1 == s->video_buf) { perror("mmap"); goto fail; } } s->gb_frame = 0; s->time_frame = av_gettime() * s->frame_rate / s->frame_rate_base; /* start to grab the first frame */ s->gb_buf.frame = s->gb_frame % s->gb_buffers.frames; s->gb_buf.height = height; s->gb_buf.width = width; s->gb_buf.format = pict.palette; ret = ioctl(video_fd, VIDIOCMCAPTURE, &s->gb_buf); if (ret < 0) { if (errno != EAGAIN) { fail1: av_log(s1, AV_LOG_ERROR, "Fatal: grab device does not support suitable format\n"); } else { av_log(s1, AV_LOG_ERROR,"Fatal: grab device does not receive any video signal\n"); } goto fail; } for (j = 1; j < s->gb_buffers.frames; j++) { s->gb_buf.frame = j; ioctl(video_fd, VIDIOCMCAPTURE, &s->gb_buf); } s->frame_format = s->gb_buf.format; s->use_mmap = 1; } for (j = 0; j < vformat_num; j++) { if (s->frame_format == video_formats[j].palette) { frame_size = width * height * video_formats[j].depth / 8; st->codec->pix_fmt = video_formats[j].pix_fmt; break; } } if (j >= vformat_num) goto fail; s->fd = video_fd; s->frame_size = frame_size; st->codec->codec_type = CODEC_TYPE_VIDEO; st->codec->codec_id = CODEC_ID_RAWVIDEO; st->codec->width = width; st->codec->height = height; st->codec->time_base.den = frame_rate; st->codec->time_base.num = frame_rate_base; st->codec->bit_rate = frame_size * 1/av_q2d(st->codec->time_base) * 8; return 0; fail: if (video_fd >= 0) close(video_fd); av_free(st); return AVERROR(EIO); } static int v4l_mm_read_picture(VideoData *s, uint8_t *buf) { uint8_t *ptr; while (ioctl(s->fd, VIDIOCSYNC, &s->gb_frame) < 0 && (errno == EAGAIN || errno == EINTR)); ptr = s->video_buf + s->gb_buffers.offsets[s->gb_frame]; memcpy(buf, ptr, s->frame_size); /* Setup to capture the next frame */ s->gb_buf.frame = s->gb_frame; if (ioctl(s->fd, VIDIOCMCAPTURE, &s->gb_buf) < 0) { if (errno == EAGAIN) av_log(NULL, AV_LOG_ERROR, "Cannot Sync\n"); else perror("VIDIOCMCAPTURE"); return AVERROR(EIO); } /* This is now the grabbing frame */ s->gb_frame = (s->gb_frame + 1) % s->gb_buffers.frames; return s->frame_size; } static int grab_read_packet(AVFormatContext *s1, AVPacket *pkt) { VideoData *s = s1->priv_data; int64_t curtime, delay; struct timespec ts; /* Calculate the time of the next frame */ s->time_frame += INT64_C(1000000); /* wait based on the frame rate */ for(;;) { curtime = av_gettime(); delay = s->time_frame * s->frame_rate_base / s->frame_rate - curtime; if (delay <= 0) { if (delay < INT64_C(-1000000) * s->frame_rate_base / s->frame_rate) { /* printf("grabbing is %d frames late (dropping)\n", (int) -(delay / 16666)); */ s->time_frame += INT64_C(1000000); } break; } ts.tv_sec = delay / 1000000; ts.tv_nsec = (delay % 1000000) * 1000; nanosleep(&ts, NULL); } if (av_new_packet(pkt, s->frame_size) < 0) return AVERROR(EIO); pkt->pts = curtime; /* read one frame */ if (s->aiw_enabled) { return aiw_read_picture(s, pkt->data); } else if (s->use_mmap) { return v4l_mm_read_picture(s, pkt->data); } else { if (read(s->fd, pkt->data, pkt->size) != pkt->size) return AVERROR(EIO); return s->frame_size; } } static int grab_read_close(AVFormatContext *s1) { VideoData *s = s1->priv_data; if (s->aiw_enabled) aiw_close(s); if (s->use_mmap) munmap(s->video_buf, s->gb_buffers.size); /* mute audio. we must force it because the BTTV driver does not return its state correctly */ s->audio_saved.flags |= VIDEO_AUDIO_MUTE; ioctl(s->fd, VIDIOCSAUDIO, &s->audio_saved); close(s->fd); return 0; } AVInputFormat video_grab_v4l_demuxer = { "video4linux", "video grab", sizeof(VideoData), NULL, grab_read_header, grab_read_packet, grab_read_close, .flags = AVFMT_NOFILE, }; /* All in Wonder specific stuff */ /* XXX: remove and merge in libavcodec/imgconvert.c */ static int aiw_init(VideoData *s) { int width, height; width = s->width; height = s->height; if ((width == s->video_cap.maxwidth && height == s->video_cap.maxheight) || (width == s->video_cap.maxwidth && height == s->video_cap.maxheight*2) || (width == s->video_cap.maxwidth/2 && height == s->video_cap.maxheight)) { s->deint=0; s->halfw=0; if (height == s->video_cap.maxheight*2) s->deint=1; if (width == s->video_cap.maxwidth/2) s->halfw=1; } else { av_log(NULL, AV_LOG_ERROR, "\nIncorrect Grab Size Supplied - Supported Sizes Are:\n"); av_log(NULL, AV_LOG_ERROR, " %dx%d %dx%d %dx%d\n\n", s->video_cap.maxwidth,s->video_cap.maxheight, s->video_cap.maxwidth,s->video_cap.maxheight*2, s->video_cap.maxwidth/2,s->video_cap.maxheight); goto fail; } if (s->halfw == 0) { s->src_mem = av_malloc(s->width*2); } else { s->src_mem = av_malloc(s->width*4); } if (!s->src_mem) goto fail; s->lum_m4_mem = av_malloc(s->width); if (!s->lum_m4_mem) goto fail; return 0; fail: av_freep(&s->src_mem); av_freep(&s->lum_m4_mem); return -1; } #ifdef HAVE_MMX #include "i386/mmx.h" #define LINE_WITH_UV \ movq_m2r(ptr[0],mm0); \ movq_m2r(ptr[8],mm1); \ movq_r2r(mm0, mm4); \ punpcklbw_r2r(mm1,mm0); \ punpckhbw_r2r(mm1,mm4); \ movq_r2r(mm0,mm5); \ punpcklbw_r2r(mm4,mm0); \ punpckhbw_r2r(mm4,mm5); \ movq_r2r(mm0,mm1); \ punpcklbw_r2r(mm5,mm1); \ movq_r2m(mm1,lum[0]); \ movq_m2r(ptr[16],mm2); \ movq_m2r(ptr[24],mm1); \ movq_r2r(mm2,mm4); \ punpcklbw_r2r(mm1,mm2); \ punpckhbw_r2r(mm1,mm4); \ movq_r2r(mm2,mm3); \ punpcklbw_r2r(mm4,mm2); \ punpckhbw_r2r(mm4,mm3); \ movq_r2r(mm2,mm1); \ punpcklbw_r2r(mm3,mm1); \ movq_r2m(mm1,lum[8]); \ punpckhdq_r2r(mm2,mm0); \ punpckhdq_r2r(mm3,mm5); \ movq_r2m(mm0,cb[0]); \ movq_r2m(mm5,cr[0]); #define LINE_NO_UV \ movq_m2r(ptr[0],mm0);\ movq_m2r(ptr[8],mm1);\ movq_r2r(mm0, mm4);\ punpcklbw_r2r(mm1,mm0); \ punpckhbw_r2r(mm1,mm4);\ movq_r2r(mm0,mm5);\ punpcklbw_r2r(mm4,mm0);\ punpckhbw_r2r(mm4,mm5);\ movq_r2r(mm0,mm1);\ punpcklbw_r2r(mm5,mm1);\ movq_r2m(mm1,lum[0]);\ movq_m2r(ptr[16],mm2);\ movq_m2r(ptr[24],mm1);\ movq_r2r(mm2,mm4);\ punpcklbw_r2r(mm1,mm2);\ punpckhbw_r2r(mm1,mm4);\ movq_r2r(mm2,mm3);\ punpcklbw_r2r(mm4,mm2);\ punpckhbw_r2r(mm4,mm3);\ movq_r2r(mm2,mm1);\ punpcklbw_r2r(mm3,mm1);\ movq_r2m(mm1,lum[8]); #define LINE_WITHUV_AVG \ movq_m2r(ptr[0], mm0);\ movq_m2r(ptr[8], mm1);\ movq_r2r(mm0, mm4);\ punpcklbw_r2r(mm1,mm0);\ punpckhbw_r2r(mm1,mm4);\ movq_r2r(mm0,mm5);\ punpcklbw_r2r(mm4,mm0);\ punpckhbw_r2r(mm4,mm5);\ movq_r2r(mm0,mm1);\ movq_r2r(mm5,mm2);\ punpcklbw_r2r(mm7,mm1);\ punpcklbw_r2r(mm7,mm2);\ paddw_r2r(mm6,mm1);\ paddw_r2r(mm2,mm1);\ psraw_i2r(1,mm1);\ packuswb_r2r(mm7,mm1);\ movd_r2m(mm1,lum[0]);\ movq_m2r(ptr[16],mm2);\ movq_m2r(ptr[24],mm1);\ movq_r2r(mm2,mm4);\ punpcklbw_r2r(mm1,mm2);\ punpckhbw_r2r(mm1,mm4);\ movq_r2r(mm2,mm3);\ punpcklbw_r2r(mm4,mm2);\ punpckhbw_r2r(mm4,mm3);\ movq_r2r(mm2,mm1);\ movq_r2r(mm3,mm4);\ punpcklbw_r2r(mm7,mm1);\ punpcklbw_r2r(mm7,mm4);\ paddw_r2r(mm6,mm1);\ paddw_r2r(mm4,mm1);\ psraw_i2r(1,mm1);\ packuswb_r2r(mm7,mm1);\ movd_r2m(mm1,lum[4]);\ punpckhbw_r2r(mm7,mm0);\ punpckhbw_r2r(mm7,mm2);\ paddw_r2r(mm6,mm0);\ paddw_r2r(mm2,mm0);\ psraw_i2r(1,mm0);\ packuswb_r2r(mm7,mm0);\ punpckhbw_r2r(mm7,mm5);\ punpckhbw_r2r(mm7,mm3);\ paddw_r2r(mm6,mm5);\ paddw_r2r(mm3,mm5);\ psraw_i2r(1,mm5);\ packuswb_r2r(mm7,mm5);\ movd_r2m(mm0,cb[0]);\ movd_r2m(mm5,cr[0]); #define LINE_NOUV_AVG \ movq_m2r(ptr[0],mm0);\ movq_m2r(ptr[8],mm1);\ pand_r2r(mm5,mm0);\ pand_r2r(mm5,mm1);\ pmaddwd_r2r(mm6,mm0);\ pmaddwd_r2r(mm6,mm1);\ packssdw_r2r(mm1,mm0);\ paddw_r2r(mm6,mm0);\ psraw_i2r(1,mm0);\ movq_m2r(ptr[16],mm2);\ movq_m2r(ptr[24],mm3);\ pand_r2r(mm5,mm2);\ pand_r2r(mm5,mm3);\ pmaddwd_r2r(mm6,mm2);\ pmaddwd_r2r(mm6,mm3);\ packssdw_r2r(mm3,mm2);\ paddw_r2r(mm6,mm2);\ psraw_i2r(1,mm2);\ packuswb_r2r(mm2,mm0);\ movq_r2m(mm0,lum[0]); #define DEINT_LINE_LUM(ptroff) \ movd_m2r(lum_m4[(ptroff)],mm0);\ movd_m2r(lum_m3[(ptroff)],mm1);\ movd_m2r(lum_m2[(ptroff)],mm2);\ movd_m2r(lum_m1[(ptroff)],mm3);\ movd_m2r(lum[(ptroff)],mm4);\ punpcklbw_r2r(mm7,mm0);\ movd_r2m(mm2,lum_m4[(ptroff)]);\ punpcklbw_r2r(mm7,mm1);\ punpcklbw_r2r(mm7,mm2);\ punpcklbw_r2r(mm7,mm3);\ punpcklbw_r2r(mm7,mm4);\ psllw_i2r(2,mm1);\ psllw_i2r(1,mm2);\ paddw_r2r(mm6,mm1);\ psllw_i2r(2,mm3);\ paddw_r2r(mm2,mm1);\ paddw_r2r(mm4,mm0);\ paddw_r2r(mm3,mm1);\ psubusw_r2r(mm0,mm1);\ psrlw_i2r(3,mm1);\ packuswb_r2r(mm7,mm1);\ movd_r2m(mm1,lum_m2[(ptroff)]); #else #include "dsputil.h" #define LINE_WITH_UV \ lum[0]=ptr[0];lum[1]=ptr[2];lum[2]=ptr[4];lum[3]=ptr[6];\ cb[0]=ptr[1];cb[1]=ptr[5];\ cr[0]=ptr[3];cr[1]=ptr[7];\ lum[4]=ptr[8];lum[5]=ptr[10];lum[6]=ptr[12];lum[7]=ptr[14];\ cb[2]=ptr[9];cb[3]=ptr[13];\ cr[2]=ptr[11];cr[3]=ptr[15];\ lum[8]=ptr[16];lum[9]=ptr[18];lum[10]=ptr[20];lum[11]=ptr[22];\ cb[4]=ptr[17];cb[5]=ptr[21];\ cr[4]=ptr[19];cr[5]=ptr[23];\ lum[12]=ptr[24];lum[13]=ptr[26];lum[14]=ptr[28];lum[15]=ptr[30];\ cb[6]=ptr[25];cb[7]=ptr[29];\ cr[6]=ptr[27];cr[7]=ptr[31]; #define LINE_NO_UV \ lum[0]=ptr[0];lum[1]=ptr[2];lum[2]=ptr[4];lum[3]=ptr[6];\ lum[4]=ptr[8];lum[5]=ptr[10];lum[6]=ptr[12];lum[7]=ptr[14];\ lum[8]=ptr[16];lum[9]=ptr[18];lum[10]=ptr[20];lum[11]=ptr[22];\ lum[12]=ptr[24];lum[13]=ptr[26];lum[14]=ptr[28];lum[15]=ptr[30]; #define LINE_WITHUV_AVG \ sum=(ptr[0]+ptr[2]+1) >> 1;lum[0]=sum; \ sum=(ptr[4]+ptr[6]+1) >> 1;lum[1]=sum; \ sum=(ptr[1]+ptr[5]+1) >> 1;cb[0]=sum; \ sum=(ptr[3]+ptr[7]+1) >> 1;cr[0]=sum; \ sum=(ptr[8]+ptr[10]+1) >> 1;lum[2]=sum; \ sum=(ptr[12]+ptr[14]+1) >> 1;lum[3]=sum; \ sum=(ptr[9]+ptr[13]+1) >> 1;cb[1]=sum; \ sum=(ptr[11]+ptr[15]+1) >> 1;cr[1]=sum; \ sum=(ptr[16]+ptr[18]+1) >> 1;lum[4]=sum; \ sum=(ptr[20]+ptr[22]+1) >> 1;lum[5]=sum; \ sum=(ptr[17]+ptr[21]+1) >> 1;cb[2]=sum; \ sum=(ptr[19]+ptr[23]+1) >> 1;cr[2]=sum; \ sum=(ptr[24]+ptr[26]+1) >> 1;lum[6]=sum; \ sum=(ptr[28]+ptr[30]+1) >> 1;lum[7]=sum; \ sum=(ptr[25]+ptr[29]+1) >> 1;cb[3]=sum; \ sum=(ptr[27]+ptr[31]+1) >> 1;cr[3]=sum; #define LINE_NOUV_AVG \ sum=(ptr[0]+ptr[2]+1) >> 1;lum[0]=sum; \ sum=(ptr[4]+ptr[6]+1) >> 1;lum[1]=sum; \ sum=(ptr[8]+ptr[10]+1) >> 1;lum[2]=sum; \ sum=(ptr[12]+ptr[14]+1) >> 1;lum[3]=sum; \ sum=(ptr[16]+ptr[18]+1) >> 1;lum[4]=sum; \ sum=(ptr[20]+ptr[22]+1) >> 1;lum[5]=sum; \ sum=(ptr[24]+ptr[26]+1) >> 1;lum[6]=sum; \ sum=(ptr[28]+ptr[30]+1) >> 1;lum[7]=sum; #define DEINT_LINE_LUM(ptroff) \ sum=(-lum_m4[(ptroff)]+(lum_m3[(ptroff)]<<2)+(lum_m2[(ptroff)]<<1)+(lum_m1[(ptroff)]<<2)-lum[(ptroff)]); \ lum_m4[(ptroff)]=lum_m2[(ptroff)];\ lum_m2[(ptroff)]=cm[(sum+4)>>3];\ sum=(-lum_m4[(ptroff)+1]+(lum_m3[(ptroff)+1]<<2)+(lum_m2[(ptroff)+1]<<1)+(lum_m1[(ptroff)+1]<<2)-lum[(ptroff)+1]); \ lum_m4[(ptroff)+1]=lum_m2[(ptroff)+1];\ lum_m2[(ptroff)+1]=cm[(sum+4)>>3];\ sum=(-lum_m4[(ptroff)+2]+(lum_m3[(ptroff)+2]<<2)+(lum_m2[(ptroff)+2]<<1)+(lum_m1[(ptroff)+2]<<2)-lum[(ptroff)+2]); \ lum_m4[(ptroff)+2]=lum_m2[(ptroff)+2];\ lum_m2[(ptroff)+2]=cm[(sum+4)>>3];\ sum=(-lum_m4[(ptroff)+3]+(lum_m3[(ptroff)+3]<<2)+(lum_m2[(ptroff)+3]<<1)+(lum_m1[(ptroff)+3]<<2)-lum[(ptroff)+3]); \ lum_m4[(ptroff)+3]=lum_m2[(ptroff)+3];\ lum_m2[(ptroff)+3]=cm[(sum+4)>>3]; #endif /* Read two fields separately. */ static int aiw_read_picture(VideoData *s, uint8_t *data) { uint8_t *ptr, *lum, *cb, *cr; int h; #ifndef HAVE_MMX int sum; #endif uint8_t* src = s->src_mem; uint8_t *ptrend = &src[s->width*2]; lum=data; cb=&lum[s->width*s->height]; cr=&cb[(s->width*s->height)/4]; if (s->deint == 0 && s->halfw == 0) { while (read(s->fd,src,s->width*2) < 0) { usleep(100); } for (h = 0; h < s->height-2; h+=2) { for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16, cb+=8, cr+=8) { LINE_WITH_UV } read(s->fd,src,s->width*2); for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16) { LINE_NO_UV } read(s->fd,src,s->width*2); } /* * Do last two lines */ for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16, cb+=8, cr+=8) { LINE_WITH_UV } read(s->fd,src,s->width*2); for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16) { LINE_NO_UV } /* drop second field */ while (read(s->fd,src,s->width*2) < 0) { usleep(100); } for (h = 0; h < s->height - 1; h++) { read(s->fd,src,s->width*2); } } else if (s->halfw == 1) { #ifdef HAVE_MMX mmx_t rounder; mmx_t masker; rounder.uw[0]=1; rounder.uw[1]=1; rounder.uw[2]=1; rounder.uw[3]=1; masker.ub[0]=0xff; masker.ub[1]=0; masker.ub[2]=0xff; masker.ub[3]=0; masker.ub[4]=0xff; masker.ub[5]=0; masker.ub[6]=0xff; masker.ub[7]=0; pxor_r2r(mm7,mm7); movq_m2r(rounder,mm6); #endif while (read(s->fd,src,s->width*4) < 0) { usleep(100); } ptrend = &src[s->width*4]; for (h = 0; h < s->height-2; h+=2) { for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=8, cb+=4, cr+=4) { LINE_WITHUV_AVG } read(s->fd,src,s->width*4); #ifdef HAVE_MMX movq_m2r(masker,mm5); #endif for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=8) { LINE_NOUV_AVG } read(s->fd,src,s->width*4); } /* * Do last two lines */ for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=8, cb+=4, cr+=4) { LINE_WITHUV_AVG } read(s->fd,src,s->width*4); #ifdef HAVE_MMX movq_m2r(masker,mm5); #endif for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=8) { LINE_NOUV_AVG } /* drop second field */ while (read(s->fd,src,s->width*4) < 0) { usleep(100); } for (h = 0; h < s->height - 1; h++) { read(s->fd,src,s->width*4); } } else { uint8_t *lum_m1, *lum_m2, *lum_m3, *lum_m4; #ifdef HAVE_MMX mmx_t rounder; rounder.uw[0]=4; rounder.uw[1]=4; rounder.uw[2]=4; rounder.uw[3]=4; movq_m2r(rounder,mm6); pxor_r2r(mm7,mm7); #else uint8_t *cm = ff_cropTbl + MAX_NEG_CROP; #endif /* read two fields and deinterlace them */ while (read(s->fd,src,s->width*2) < 0) { usleep(100); } for (h = 0; h < (s->height/2)-2; h+=2) { for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16, cb+=8, cr+=8) { LINE_WITH_UV } read(s->fd,src,s->width*2); /* skip a luminance line - will be filled in later */ lum += s->width; for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16, cb+=8, cr+=8) { LINE_WITH_UV } /* skip a luminance line - will be filled in later */ lum += s->width; read(s->fd,src,s->width*2); } /* * Do last two lines */ for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16, cb+=8, cr+=8) { LINE_WITH_UV } /* skip a luminance line - will be filled in later */ lum += s->width; read(s->fd,src,s->width*2); for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16, cb+=8, cr+=8) { LINE_WITH_UV } /* * * SECOND FIELD * */ lum=&data[s->width]; while (read(s->fd,src,s->width*2) < 0) { usleep(10); } /* First (and last) two lines not interlaced */ for (h = 0; h < 2; h++) { for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16) { LINE_NO_UV } read(s->fd,src,s->width*2); /* skip a luminance line */ lum += s->width; } lum_m1=&lum[-s->width]; lum_m2=&lum_m1[-s->width]; lum_m3=&lum_m2[-s->width]; memmove(s->lum_m4_mem,&lum_m3[-s->width],s->width); for (; h < (s->height/2)-1; h++) { lum_m4=s->lum_m4_mem; for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16,lum_m1+=16,lum_m2+=16,lum_m3+=16,lum_m4+=16) { LINE_NO_UV DEINT_LINE_LUM(0) DEINT_LINE_LUM(4) DEINT_LINE_LUM(8) DEINT_LINE_LUM(12) } read(s->fd,src,s->width*2); /* skip a luminance line */ lum += s->width; lum_m1 += s->width; lum_m2 += s->width; lum_m3 += s->width; // lum_m4 += s->width; } /* * Do last line */ lum_m4=s->lum_m4_mem; for (ptr = &src[0]; ptr < ptrend; ptr+=32, lum+=16, lum_m1+=16, lum_m2+=16, lum_m3+=16, lum_m4+=16) { LINE_NO_UV DEINT_LINE_LUM(0) DEINT_LINE_LUM(4) DEINT_LINE_LUM(8) DEINT_LINE_LUM(12) } } emms_c(); return s->frame_size; } static int aiw_close(VideoData *s) { av_freep(&s->lum_m4_mem); av_freep(&s->src_mem); return 0; }