Mercurial > mplayer.hg
view libmpdemux/tvi_v4l.c @ 7598:48f8c731efb5
Adding function for estimating required buffer length
author | anders |
---|---|
date | Sat, 05 Oct 2002 11:02:39 +0000 |
parents | cfd6a99021ac |
children | 172ee2b4612f |
line wrap: on
line source
/* Video 4 Linux input (C) Alex Beregszaszi <alex@naxine.org> Some ideas are based on xawtv/libng's grab-v4l.c written by Gerd Knorr <kraxel@bytesex.org> Multithreading, a/v sync and native ALSA support by Jindrich Makovicka <makovick@kmlinux.fjfi.cvut.cz> CODE IS UNDER DEVELOPMENT, NO FEATURE REQUESTS PLEASE! */ #include "config.h" #if defined(USE_TV) && defined(HAVE_TV_V4L) #include <stdio.h> #include <errno.h> #include <fcntl.h> #include <signal.h> #include <sys/ioctl.h> #include <sys/types.h> #include <sys/time.h> #include <linux/videodev.h> #include <unistd.h> #include <sys/mman.h> #include <stdlib.h> #include <string.h> #include <pthread.h> #ifdef HAVE_SYS_SYSINFO_H #include <sys/sysinfo.h> #endif #include "mp_msg.h" #include "../libao2/afmt.h" #include "../libvo/img_format.h" #include "../libvo/fastmemcpy.h" #include "tv.h" #include "audio_in.h" static tvi_info_t info = { "Video 4 Linux input", "v4l", "Alex Beregszaszi <alex@naxine.org>", "under development" }; #define PAL_WIDTH 768 #define PAL_HEIGHT 576 #define PAL_FPS 25 #define NTSC_WIDTH 640 #define NTSC_HEIGHT 480 #define NTSC_FPS 30 #define MAX_AUDIO_CHANNELS 10 #define VID_BUF_SIZE_IMMEDIATE 2 typedef struct { /* general */ char *video_device; int video_fd; struct video_capability capability; struct video_channel *channels; int act_channel; struct video_tuner tuner; /* video */ struct video_picture picture; int format; /* output format */ int width; int height; int bytesperline; int fps; struct video_mbuf mbuf; unsigned char *mmap; struct video_mmap *buf; int nbuf; /* audio */ char *audio_device; audio_in_t audio_in; int audio_id; struct video_audio audio[MAX_AUDIO_CHANNELS]; int audio_channels[MAX_AUDIO_CHANNELS]; /* buffering stuff */ int immediate_mode; int audio_buffer_size; int aud_skew_cnt; unsigned char *audio_ringbuffer; double *audio_skew_buffer; volatile int audio_head; volatile int audio_tail; volatile int audio_cnt; volatile double audio_skew; volatile double audio_skew_factor; volatile double audio_skew_measure_time; volatile int audio_drop; int first; int video_buffer_size; unsigned char *video_ringbuffer; double *video_timebuffer; volatile int video_head; volatile int video_tail; volatile int video_cnt; volatile int shutdown; pthread_t audio_grabber_thread; pthread_t video_grabber_thread; pthread_mutex_t audio_starter; pthread_mutex_t skew_mutex; double starttime; double audio_secs_per_block; double audio_skew_total; long audio_recv_blocks_total; long audio_sent_blocks_total; } priv_t; #include "tvi_def.h" static const char *device_cap2name[] = { "capture", "tuner", "teletext", "overlay", "chromakey", "clipping", "frameram", "scales", "monochrome", "subcapture", "mpeg-decoder", "mpeg-encoder", "mjpeg-decoder", "mjpeg-encoder", NULL }; static const char *device_palette2name[] = { "-", "grey", "hi240", "rgb16", "rgb24", "rgb32", "rgb15", "yuv422", "yuyv", "uyvy", "yuv420", "yuv411", "raw", "yuv422p", "yuv411p", "yuv420p", "yuv410p" }; #define PALETTE(x) ((x < sizeof(device_palette2name)/sizeof(char*)) ? device_palette2name[x] : "UNKNOWN") static const char *norm2name(int mode) { switch (mode) { case VIDEO_MODE_PAL: return "pal"; case VIDEO_MODE_SECAM: return "secam"; case VIDEO_MODE_NTSC: return "ntsc"; case VIDEO_MODE_AUTO: return "auto"; default: return "unknown"; } }; static const char *audio_mode2name(int mode) { switch (mode) { case VIDEO_SOUND_MONO: return "mono"; case VIDEO_SOUND_STEREO: return "stereo"; case VIDEO_SOUND_LANG1: return "language1"; case VIDEO_SOUND_LANG2: return "language2"; default: return "unknown"; } }; static void *audio_grabber(void *data); static void *video_grabber(void *data); static int palette2depth(int palette) { switch(palette) { /* component */ case VIDEO_PALETTE_RGB555: return(15); case VIDEO_PALETTE_RGB565: return(16); case VIDEO_PALETTE_RGB24: return(24); case VIDEO_PALETTE_RGB32: return(32); /* planar */ case VIDEO_PALETTE_YUV411P: case VIDEO_PALETTE_YUV420P: case VIDEO_PALETTE_YUV410P: return(12); /* packed */ case VIDEO_PALETTE_YUV422P: case VIDEO_PALETTE_YUV422: case VIDEO_PALETTE_YUYV: case VIDEO_PALETTE_UYVY: case VIDEO_PALETTE_YUV420: case VIDEO_PALETTE_YUV411: return(16); } return(-1); } static int format2palette(int format) { switch(format) { case IMGFMT_BGR15: return(VIDEO_PALETTE_RGB555); case IMGFMT_BGR16: return(VIDEO_PALETTE_RGB565); case IMGFMT_BGR24: return(VIDEO_PALETTE_RGB24); case IMGFMT_BGR32: return(VIDEO_PALETTE_RGB32); case IMGFMT_YV12: case IMGFMT_I420: return(VIDEO_PALETTE_YUV420P); case IMGFMT_YUY2: return(VIDEO_PALETTE_YUV422); } return(-1); } // sets and sanitizes audio buffer/block sizes static void setup_audio_buffer_sizes(priv_t *priv) { int bytes_per_sample = priv->audio_in.bytes_per_sample; // make the audio buffer at least 5 seconds long priv->audio_buffer_size = 1 + 5*priv->audio_in.samplerate *priv->audio_in.channels *bytes_per_sample/priv->audio_in.blocksize; if (priv->audio_buffer_size < 256) priv->audio_buffer_size = 256; // make the skew buffer at least 1 second long priv->aud_skew_cnt = 1 + 1*priv->audio_in.samplerate *priv->audio_in.channels *bytes_per_sample/priv->audio_in.blocksize; if (priv->aud_skew_cnt < 16) priv->aud_skew_cnt = 16; mp_msg(MSGT_TV, MSGL_V, "Audio capture - buffer %d blocks of %d bytes, skew average from %d meas.\n", priv->audio_buffer_size, priv->audio_in.blocksize, priv->aud_skew_cnt); } tvi_handle_t *tvi_init_v4l(char *device, char *adevice) { tvi_handle_t *h; priv_t *priv; h = new_handle(); if (!h) return(NULL); priv = h->priv; /* set video device name */ if (!device) priv->video_device = strdup("/dev/video"); else priv->video_device = strdup(device); /* set video device name */ if (!adevice) priv->audio_device = NULL; else { priv->audio_device = strdup(adevice); } /* allocation failed */ if (!priv->video_device) { free_handle(h); return(NULL); } return(h); } /* retrieves info about audio channels from the BTTV */ static void init_v4l_audio(priv_t *priv) { int i; int reqmode; if (!priv->capability.audios) return; /* audio chanlist */ mp_msg(MSGT_TV, MSGL_V, " Audio devices: %d\n", priv->capability.audios); mp_msg(MSGT_TV, MSGL_V, "Video capture card reports the audio setup as follows:\n"); for (i = 0; i < priv->capability.audios; i++) { if (i >= MAX_AUDIO_CHANNELS) { mp_msg(MSGT_TV, MSGL_ERR, "no space for more audio channels (increase in source!) (%d > %d)\n", i, MAX_AUDIO_CHANNELS); i = priv->capability.audios; break; } priv->audio[i].audio = i; if (ioctl(priv->video_fd, VIDIOCGAUDIO, &priv->audio[i]) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get audio failed: %s\n", strerror(errno)); break; } /* mute all channels */ priv->audio[i].volume = 0; priv->audio[i].flags |= VIDEO_AUDIO_MUTE; reqmode = -1; if (tv_param_amode >= 0) { switch (tv_param_amode) { case 0: reqmode = VIDEO_SOUND_MONO; break; case 1: reqmode = VIDEO_SOUND_STEREO; break; case 2: reqmode = VIDEO_SOUND_LANG1; break; case 3: reqmode = VIDEO_SOUND_LANG2; break; default: mp_msg(MSGT_TV, MSGL_ERR, "Unknown audio mode requested.\n"); break; } if (reqmode >= 0) priv->audio[i].mode = reqmode; } ioctl(priv->video_fd, VIDIOCSAUDIO, &priv->audio[i]); // get the parameters back if (ioctl(priv->video_fd, VIDIOCGAUDIO, &priv->audio[i]) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get audio failed: %s\n", strerror(errno)); break; } switch(priv->audio[i].mode) { case VIDEO_SOUND_MONO: case VIDEO_SOUND_LANG1: case VIDEO_SOUND_LANG2: priv->audio_channels[i] = 1; break; case VIDEO_SOUND_STEREO: priv->audio_channels[i] = 2; break; } if (reqmode >= 0 && priv->audio[i].mode != reqmode) { mp_msg(MSGT_TV, MSGL_ERR, "Audio mode setup warning!\n"); mp_msg(MSGT_TV, MSGL_ERR, "Requested mode was %s, but v4l still reports %s.\n", audio_mode2name(reqmode), audio_mode2name(priv->audio[i].mode)); mp_msg(MSGT_TV, MSGL_ERR, "You may need \"forcechan\" option to force stereo/mono audio recording.\n"); } /* display stuff */ mp_msg(MSGT_TV, MSGL_V, " %d: %s: ", priv->audio[i].audio, priv->audio[i].name); if (priv->audio[i].flags & VIDEO_AUDIO_MUTABLE) { mp_msg(MSGT_TV, MSGL_V, "muted=%s ", (priv->audio[i].flags & VIDEO_AUDIO_MUTE) ? "yes" : "no"); } mp_msg(MSGT_TV, MSGL_V, "vol=%d bass=%d treble=%d balance=%d mode=%s", priv->audio[i].volume, priv->audio[i].bass, priv->audio[i].treble, priv->audio[i].balance, audio_mode2name(priv->audio[i].mode)); mp_msg(MSGT_TV, MSGL_V, " chan=%d\n", priv->audio_channels[i]); if (tv_param_forcechan >= 0) priv->audio_channels[i] = tv_param_forcechan; // we'll call VIDIOCSAUDIO again when starting capture // let's set audio mode to requested mode again for the case // when VIDIOCGAUDIO just cannot report the mode correctly if (reqmode >= 0) priv->audio[i].mode = reqmode; } } static int init(priv_t *priv) { int i; if (tv_param_immediate == 1) tv_param_noaudio = 1; priv->video_ringbuffer = NULL; priv->video_timebuffer = NULL; priv->audio_ringbuffer = NULL; priv->audio_skew_buffer = NULL; priv->video_fd = open(priv->video_device, O_RDWR); mp_msg(MSGT_TV, MSGL_DBG2, "Video fd: %d, %x\n", priv->video_fd, priv->video_device); if (priv->video_fd == -1) { mp_msg(MSGT_TV, MSGL_ERR, "unable to open '%s': %s\n", priv->video_device, strerror(errno)); goto err; } /* get capabilities (priv->capability is needed!) */ if (ioctl(priv->video_fd, VIDIOCGCAP, &priv->capability) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get capabilites failed: %s\n", strerror(errno)); goto err; } fcntl(priv->video_fd, F_SETFD, FD_CLOEXEC); mp_msg(MSGT_TV, MSGL_INFO, "Selected device: %s\n", priv->capability.name); mp_msg(MSGT_TV, MSGL_INFO, " Capabilites: "); for (i = 0; device_cap2name[i] != NULL; i++) if (priv->capability.type & (1 << i)) mp_msg(MSGT_TV, MSGL_INFO, "%s ", device_cap2name[i]); mp_msg(MSGT_TV, MSGL_INFO, "\n"); mp_msg(MSGT_TV, MSGL_INFO, " Device type: %d\n", priv->capability.type); mp_msg(MSGT_TV, MSGL_INFO, " Supported sizes: %dx%d => %dx%d\n", priv->capability.minwidth, priv->capability.minheight, priv->capability.maxwidth, priv->capability.maxheight); priv->width = priv->capability.minwidth; priv->height = priv->capability.minheight; mp_msg(MSGT_TV, MSGL_INFO, " Inputs: %d\n", priv->capability.channels); priv->channels = (struct video_channel *)malloc(sizeof(struct video_channel)*priv->capability.channels); if (!priv->channels) goto malloc_failed; memset(priv->channels, 0, sizeof(struct video_channel)*priv->capability.channels); for (i = 0; i < priv->capability.channels; i++) { priv->channels[i].channel = i; if (ioctl(priv->video_fd, VIDIOCGCHAN, &priv->channels[i]) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get channel failed: %s\n", strerror(errno)); break; } mp_msg(MSGT_TV, MSGL_INFO, " %d: %s: %s%s%s%s (tuner:%d, norm:%s)\n", i, priv->channels[i].name, (priv->channels[i].flags & VIDEO_VC_TUNER) ? "tuner " : "", (priv->channels[i].flags & VIDEO_VC_AUDIO) ? "audio " : "", (priv->channels[i].flags & VIDEO_TYPE_TV) ? "tv " : "", (priv->channels[i].flags & VIDEO_TYPE_CAMERA) ? "camera " : "", priv->channels[i].tuners, norm2name(priv->channels[i].norm)); } priv->act_channel = 0; if (!(priv->capability.type & VID_TYPE_CAPTURE)) { mp_msg(MSGT_TV, MSGL_ERR, "Only grabbing supported (for overlay use another program)\n"); goto err; } /* map grab buffer */ if (ioctl(priv->video_fd, VIDIOCGMBUF, &priv->mbuf) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get mbuf failed: %s\n", strerror(errno)); goto err; } mp_msg(MSGT_TV, MSGL_V, "mbuf: size=%d, frames=%d\n", priv->mbuf.size, priv->mbuf.frames); priv->mmap = mmap(0, priv->mbuf.size, PROT_READ|PROT_WRITE, MAP_SHARED, priv->video_fd, 0); if (priv->mmap == (unsigned char *)-1) { mp_msg(MSGT_TV, MSGL_ERR, "Unable to map memory for buffers: %s\n", strerror(errno)); goto err; } mp_msg(MSGT_TV, MSGL_DBG2, "our buffer: %p\n", priv->mmap); /* num of buffers */ priv->nbuf = priv->mbuf.frames; /* video buffers */ priv->buf = (struct video_mmap *)malloc(priv->nbuf * sizeof(struct video_mmap)); if (!priv->buf) goto malloc_failed; memset(priv->buf, 0, priv->nbuf * sizeof(struct video_mmap)); /* init v4l audio even when we don't capture */ init_v4l_audio(priv); if (!priv->capability.audios) tv_param_noaudio = 1; /* audio init */ if (!tv_param_noaudio) { #ifdef HAVE_ALSA9 if (tv_param_alsa) audio_in_init(&priv->audio_in, AUDIO_IN_ALSA); else audio_in_init(&priv->audio_in, AUDIO_IN_OSS); #else audio_in_init(&priv->audio_in, AUDIO_IN_OSS); #endif if (priv->audio_device) { audio_in_set_device(&priv->audio_in, priv->audio_device); } if (tv_param_audio_id < priv->capability.audios) priv->audio_id = tv_param_audio_id; else priv->audio_id = 0; audio_in_set_samplerate(&priv->audio_in, 44100); audio_in_set_channels(&priv->audio_in, priv->audio_channels[priv->audio_id]); if (audio_in_setup(&priv->audio_in) < 0) return 0; setup_audio_buffer_sizes(priv); } return(1); malloc_failed: if (priv->channels) free(priv->channels); if (priv->buf) free(priv->buf); err: if (priv->video_fd != -1) close(priv->video_fd); return(0); } static int uninit(priv_t *priv) { priv->shutdown = 1; mp_msg(MSGT_TV, MSGL_V, "Waiting for threads to finish... "); if (!tv_param_noaudio) { pthread_join(priv->audio_grabber_thread, NULL); pthread_mutex_destroy(&priv->audio_starter); pthread_mutex_destroy(&priv->skew_mutex); } pthread_join(priv->video_grabber_thread, NULL); mp_msg(MSGT_TV, MSGL_V, "done\n"); priv->audio[priv->audio_id].volume = 0; priv->audio[priv->audio_id].flags |= VIDEO_AUDIO_MUTE; ioctl(priv->video_fd, VIDIOCSAUDIO, &priv->audio[priv->audio_id]); close(priv->video_fd); audio_in_uninit(&priv->audio_in); if (priv->video_ringbuffer) free(priv->video_ringbuffer); if (priv->video_timebuffer) free(priv->video_timebuffer); if (!tv_param_noaudio) { if (priv->audio_ringbuffer) free(priv->audio_ringbuffer); if (priv->audio_skew_buffer) free(priv->audio_skew_buffer); } return(1); } static int get_capture_buffer_size(priv_t *priv) { int bufsize, cnt; #ifdef HAVE_SYS_SYSINFO_H struct sysinfo si; sysinfo(&si); if (si.totalram<2*1024*1024) { bufsize = 1024*1024; } else { bufsize = si.totalram/2; } #else bufsize = 16*1024*1024; #endif cnt = bufsize/(priv->height*priv->bytesperline); if (cnt < 2) cnt = 2; return cnt; } static int start(priv_t *priv) { int i; int bytes_per_sample; if (ioctl(priv->video_fd, VIDIOCGPICT, &priv->picture) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get picture failed: %s\n", strerror(errno)); return(0); } priv->picture.palette = format2palette(priv->format); priv->picture.depth = palette2depth(priv->picture.palette); if (priv->format != IMGFMT_BGR15) { priv->bytesperline = priv->width * priv->picture.depth / 8; } else { priv->bytesperline = priv->width * 2; } mp_msg(MSGT_TV, MSGL_V, "Picture values:\n"); mp_msg(MSGT_TV, MSGL_V, " Depth: %d, Palette: %s (Format: %s)\n", priv->picture.depth, PALETTE(priv->picture.palette), vo_format_name(priv->format)); mp_msg(MSGT_TV, MSGL_V, " Brightness: %d, Hue: %d, Colour: %d, Contrast: %d\n", priv->picture.brightness, priv->picture.hue, priv->picture.colour, priv->picture.contrast); if (ioctl(priv->video_fd, VIDIOCSPICT, &priv->picture) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl set picture failed: %s\n", strerror(errno)); return(0); } priv->nbuf = priv->mbuf.frames; for (i=0; i < priv->nbuf; i++) { priv->buf[i].format = priv->picture.palette; priv->buf[i].frame = i; priv->buf[i].width = priv->width; priv->buf[i].height = priv->height; mp_msg(MSGT_TV, MSGL_DBG2, "buffer: %d => %p\n", i, &priv->buf[i]); } #if 0 { struct video_play_mode pmode; pmode.mode = VID_PLAY_NORMAL; pmode.p1 = 1; pmode.p2 = 0; if (ioctl(priv->video_fd, VIDIOCSPLAYMODE, &pmode) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl set play mode failed: %s\n", strerror(errno)); // return(0); } } #endif #if 0 { struct video_window win; win.x = 0; win.y = 0; win.width = priv->width; win.height = priv->height; win.chromakey = -1; win.flags = 0; //win.clipcount = 0; ioctl(priv->video_fd, VIDIOCSWIN, &win); } // initialize video capture if (ioctl(priv->video_fd, VIDIOCCAPTURE, &one) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "FATAL: ioctl ccapture failed: %s\n", strerror(errno)); return(0); } #endif /* setup audio parameters */ if (!tv_param_noaudio) { setup_audio_buffer_sizes(priv); bytes_per_sample = priv->audio_in.bytes_per_sample; priv->audio_skew_buffer = (double*)malloc(sizeof(double)*priv->aud_skew_cnt); if (!priv->audio_skew_buffer) { mp_msg(MSGT_TV, MSGL_ERR, "cannot allocate skew buffer: %s\n", strerror(errno)); return 0; } priv->audio_ringbuffer = (unsigned char*)malloc(priv->audio_in.blocksize*priv->audio_buffer_size); if (!priv->audio_ringbuffer) { mp_msg(MSGT_TV, MSGL_ERR, "cannot allocate audio buffer: %s\n", strerror(errno)); return 0; } priv->audio_secs_per_block = (double)priv->audio_in.blocksize/(priv->audio_in.samplerate *priv->audio_in.channels *bytes_per_sample); priv->audio_head = 0; priv->audio_tail = 0; priv->audio_cnt = 0; priv->audio_drop = 0; priv->audio_skew = 0.0; priv->audio_skew_total = 0.0; priv->audio_recv_blocks_total = 0; priv->audio_sent_blocks_total = 0; } /* setup video parameters */ if (priv->immediate_mode) { priv->video_buffer_size = VID_BUF_SIZE_IMMEDIATE; } else { priv->video_buffer_size = get_capture_buffer_size(priv); } if (!tv_param_noaudio) { if (priv->video_buffer_size < 3.0*priv->fps*priv->audio_secs_per_block) { mp_msg(MSGT_TV, MSGL_ERR, "Video buffer shorter than 3 times audio frame duration.\n" "You will probably experience heavy framedrops.\n"); } } mp_msg(MSGT_TV, MSGL_V, "Allocating a ring buffer for %d frames, %d MB total size.\n", priv->video_buffer_size, priv->video_buffer_size*priv->height*priv->bytesperline/(1024*1024)); priv->video_ringbuffer = (unsigned char*)malloc(priv->bytesperline * priv->height * priv->video_buffer_size); if (!priv->video_ringbuffer) { mp_msg(MSGT_TV, MSGL_ERR, "cannot allocate video buffer: %s\n", strerror(errno)); return 0; } priv->video_timebuffer = (double*)malloc(sizeof(double) * priv->video_buffer_size); if (!priv->video_timebuffer) { mp_msg(MSGT_TV, MSGL_ERR, "cannot allocate time buffer: %s\n", strerror(errno)); return 0; } priv->video_head = 0; priv->video_tail = 0; priv->video_cnt = 0; priv->first = 1; if (priv->capability.audios) { /* enable audio */ if (tv_param_volume >= 0) priv->audio[priv->audio_id].volume = tv_param_volume; if (tv_param_bass >= 0) priv->audio[priv->audio_id].bass = tv_param_bass; if (tv_param_treble >= 0) priv->audio[priv->audio_id].treble = tv_param_treble; if (tv_param_balance >= 0) priv->audio[priv->audio_id].balance = tv_param_balance; priv->audio[priv->audio_id].flags &= ~VIDEO_AUDIO_MUTE; mp_msg(MSGT_TV, MSGL_V, "Enabling tv audio. Requested setup is:\n"); mp_msg(MSGT_TV, MSGL_V, "id=%d vol=%d bass=%d treble=%d balance=%d mode=%s", priv->audio_id, priv->audio[priv->audio_id].volume, priv->audio[priv->audio_id].bass, priv->audio[priv->audio_id].treble, priv->audio[priv->audio_id].balance, audio_mode2name(priv->audio[priv->audio_id].mode)); mp_msg(MSGT_TV, MSGL_V, " chan=%d\n", priv->audio_channels[priv->audio_id]); ioctl(priv->video_fd, VIDIOCSAUDIO, &priv->audio[priv->audio_id]); } /* launch capture threads */ priv->shutdown = 0; if (!tv_param_noaudio) { pthread_mutex_init(&priv->audio_starter, NULL); pthread_mutex_init(&priv->skew_mutex, NULL); pthread_mutex_lock(&priv->audio_starter); pthread_create(&priv->audio_grabber_thread, NULL, audio_grabber, priv); } /* we'll launch the video capture later, when a first request for a frame arrives */ return(1); } static int control(priv_t *priv, int cmd, void *arg) { mp_msg(MSGT_TV, MSGL_DBG2, "\ndebug: control(priv=%p, cmd=%d, arg=%p)\n", priv, cmd, arg); switch(cmd) { /* ========== GENERIC controls =========== */ case TVI_CONTROL_IS_VIDEO: { if (priv->capability.type & VID_TYPE_CAPTURE) return(TVI_CONTROL_TRUE); return(TVI_CONTROL_FALSE); } case TVI_CONTROL_IS_AUDIO: if (priv->channels[priv->act_channel].flags & VIDEO_VC_AUDIO) { return(TVI_CONTROL_TRUE); } return(TVI_CONTROL_TRUE); case TVI_CONTROL_IS_TUNER: { // if (priv->capability.type & VID_TYPE_TUNER) if (priv->channels[priv->act_channel].flags & VIDEO_VC_TUNER) return(TVI_CONTROL_TRUE); return(TVI_CONTROL_FALSE); } /* ========== VIDEO controls =========== */ case TVI_CONTROL_VID_GET_FORMAT: { int output_fmt = -1; output_fmt = priv->format; (int)*(void **)arg = output_fmt; mp_msg(MSGT_TV, MSGL_V, "Output format: %s\n", vo_format_name(output_fmt)); return(TVI_CONTROL_TRUE); } case TVI_CONTROL_VID_SET_FORMAT: priv->format = (int)*(void **)arg; // !HACK! v4l uses BGR format instead of RGB // and we have to correct this. Fortunately, // tv.c reads later the format back so we // can persuade it to use what we want. if (IMGFMT_IS_RGB(priv->format)) { priv->format &= ~IMGFMT_RGB_MASK; priv->format |= IMGFMT_BGR; } return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_GET_PLANES: (int)*(void **)arg = 1; /* FIXME, also not needed at this time */ return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_GET_BITS: (int)*(void **)arg = palette2depth(format2palette(priv->format)); return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_GET_WIDTH: (int)*(void **)arg = priv->width; return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_CHK_WIDTH: { int req_width = (int)*(void **)arg; mp_msg(MSGT_TV, MSGL_V, "Requested width: %d\n", req_width); if ((req_width >= priv->capability.minwidth) && (req_width <= priv->capability.maxwidth)) return(TVI_CONTROL_TRUE); return(TVI_CONTROL_FALSE); } case TVI_CONTROL_VID_SET_WIDTH: priv->width = (int)*(void **)arg; return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_GET_HEIGHT: (int)*(void **)arg = priv->height; return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_CHK_HEIGHT: { int req_height = (int)*(void **)arg; mp_msg(MSGT_TV, MSGL_V, "Requested height: %d\n", req_height); if ((req_height >= priv->capability.minheight) && (req_height <= priv->capability.maxheight)) return(TVI_CONTROL_TRUE); return(TVI_CONTROL_FALSE); } case TVI_CONTROL_VID_SET_HEIGHT: priv->height = (int)*(void **)arg; return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_GET_PICTURE: if (ioctl(priv->video_fd, VIDIOCGPICT, &priv->picture) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get picture failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_SET_PICTURE: if (ioctl(priv->video_fd, VIDIOCSPICT, &priv->picture) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get picture failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_SET_BRIGHTNESS: priv->picture.brightness = (int)*(void **)arg; control(priv, TVI_CONTROL_VID_SET_PICTURE, 0); return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_SET_HUE: priv->picture.hue = (int)*(void **)arg; control(priv, TVI_CONTROL_VID_SET_PICTURE, 0); return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_SET_SATURATION: priv->picture.colour = (int)*(void **)arg; control(priv, TVI_CONTROL_VID_SET_PICTURE, 0); return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_SET_CONTRAST: priv->picture.contrast = (int)*(void **)arg; control(priv, TVI_CONTROL_VID_SET_PICTURE, 0); return(TVI_CONTROL_TRUE); case TVI_CONTROL_VID_GET_FPS: (int)*(void **)arg=priv->fps; return(TVI_CONTROL_TRUE); /* ========== TUNER controls =========== */ case TVI_CONTROL_TUN_GET_FREQ: { unsigned long freq; if (ioctl(priv->video_fd, VIDIOCGFREQ, &freq) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get freq failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } /* tuner uses khz not mhz ! */ // if (priv->tuner.flags & VIDEO_TUNER_LOW) // freq /= 1000; (unsigned long)*(void **)arg = freq; return(TVI_CONTROL_TRUE); } case TVI_CONTROL_TUN_SET_FREQ: { /* argument is in MHz ! */ unsigned long freq = (unsigned long)*(void **)arg; mp_msg(MSGT_TV, MSGL_V, "requested frequency: %.3f\n", (float)freq/16); /* tuner uses khz not mhz ! */ // if (priv->tuner.flags & VIDEO_TUNER_LOW) // freq *= 1000; // mp_msg(MSGT_TV, MSGL_V, " requesting from driver: freq=%.3f\n", (float)freq/16); if (ioctl(priv->video_fd, VIDIOCSFREQ, &freq) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl set freq failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } usleep(100000); // wait to supress noise during switching return(TVI_CONTROL_TRUE); } case TVI_CONTROL_TUN_GET_TUNER: { if (ioctl(priv->video_fd, VIDIOCGTUNER, &priv->tuner) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get tuner failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } mp_msg(MSGT_TV, MSGL_V, "Tuner (%s) range: %lu -> %lu\n", priv->tuner.name, priv->tuner.rangelow, priv->tuner.rangehigh); return(TVI_CONTROL_TRUE); } case TVI_CONTROL_TUN_SET_TUNER: { if (ioctl(priv->video_fd, VIDIOCSTUNER, &priv->tuner) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl set tuner failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } return(TVI_CONTROL_TRUE); } case TVI_CONTROL_TUN_SET_NORM: { int req_mode = (int)*(void **)arg; if ((req_mode != TV_NORM_PAL) && (req_mode != TV_NORM_NTSC) && (req_mode != TV_NORM_SECAM)) { mp_msg(MSGT_TV, MSGL_ERR, "Unknown norm!\n"); return(TVI_CONTROL_FALSE); } if (priv->channels[priv->act_channel].flags & VIDEO_VC_TUNER) { control(priv, TVI_CONTROL_TUN_GET_TUNER, 0); if (((req_mode == TV_NORM_PAL) && !(priv->tuner.flags & VIDEO_TUNER_PAL)) || ((req_mode == TV_NORM_NTSC) && !(priv->tuner.flags & VIDEO_TUNER_NTSC)) || ((req_mode == TV_NORM_SECAM) && !(priv->tuner.flags & VIDEO_TUNER_SECAM))) { mp_msg(MSGT_TV, MSGL_ERR, "Tuner isn't capable to set norm!\n"); return(TVI_CONTROL_FALSE); } switch(req_mode) { case TV_NORM_PAL: priv->tuner.mode = VIDEO_MODE_PAL; break; case TV_NORM_NTSC: priv->tuner.mode = VIDEO_MODE_NTSC; break; case TV_NORM_SECAM: priv->tuner.mode = VIDEO_MODE_SECAM; break; } if (control(priv, TVI_CONTROL_TUN_SET_TUNER, &priv->tuner) != TVI_CONTROL_TRUE) { return(TVI_CONTROL_FALSE); } } switch(req_mode) { case TV_NORM_PAL: priv->channels[priv->act_channel].norm = VIDEO_MODE_PAL; break; case TV_NORM_NTSC: priv->channels[priv->act_channel].norm = VIDEO_MODE_NTSC; break; case TV_NORM_SECAM: priv->channels[priv->act_channel].norm = VIDEO_MODE_SECAM; break; } if (ioctl(priv->video_fd, VIDIOCSCHAN, &priv->channels[priv->act_channel]) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl set chan failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } if (ioctl(priv->video_fd, VIDIOCGCAP, &priv->capability) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get capabilites failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } if(req_mode == TV_NORM_PAL || req_mode == TV_NORM_SECAM) { priv->fps = PAL_FPS; } if(req_mode == TV_NORM_NTSC) { priv->fps = NTSC_FPS; } if(priv->height > priv->capability.maxheight) { priv->height = priv->capability.maxheight; } if(priv->width > priv->capability.maxwidth) { priv->width = priv->capability.maxwidth; } return(TVI_CONTROL_TRUE); } case TVI_CONTROL_TUN_GET_NORM: { (int)*(void **)arg = priv->tuner.mode; return(TVI_CONTROL_TRUE); } /* ========== AUDIO controls =========== */ case TVI_CONTROL_AUD_GET_FORMAT: { (int)*(void **)arg = AFMT_S16_LE; return(TVI_CONTROL_TRUE); } case TVI_CONTROL_AUD_GET_CHANNELS: { (int)*(void **)arg = priv->audio_in.channels; return(TVI_CONTROL_TRUE); } case TVI_CONTROL_AUD_GET_SAMPLERATE: { (int)*(void **)arg = priv->audio_in.samplerate; return(TVI_CONTROL_TRUE); } case TVI_CONTROL_AUD_GET_SAMPLESIZE: { (int)*(void **)arg = priv->audio_in.bytes_per_sample; return(TVI_CONTROL_TRUE); } case TVI_CONTROL_AUD_SET_SAMPLERATE: { if (audio_in_set_samplerate(&priv->audio_in, (int)*(void **)arg) < 0) return TVI_CONTROL_FALSE; setup_audio_buffer_sizes(priv); return(TVI_CONTROL_TRUE); } /* ========== SPECIFIC controls =========== */ case TVI_CONTROL_SPC_GET_INPUT: { int req_chan = (int)*(void **)arg; int i; for (i = 0; i < priv->capability.channels; i++) { if (priv->channels[i].channel == req_chan) break; } priv->act_channel = i; if (ioctl(priv->video_fd, VIDIOCGCHAN, &priv->channels[i]) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl get channel failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } return(TVI_CONTROL_TRUE); } case TVI_CONTROL_SPC_SET_INPUT: { struct video_channel chan; int req_chan = (int)*(void **)arg; int i; if (req_chan >= priv->capability.channels) { mp_msg(MSGT_TV, MSGL_ERR, "Invalid input requested: %d, valid: 0-%d\n", req_chan, priv->capability.channels); return(TVI_CONTROL_FALSE); } for (i = 0; i < priv->capability.channels; i++) { if (priv->channels[i].channel == req_chan) chan = priv->channels[i]; } if (ioctl(priv->video_fd, VIDIOCSCHAN, &chan) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "ioctl set chan failed: %s\n", strerror(errno)); return(TVI_CONTROL_FALSE); } mp_msg(MSGT_TV, MSGL_INFO, "Using input '%s'\n", chan.name); priv->act_channel = i; /* update tuner state */ // if (priv->capability.type & VID_TYPE_TUNER) if (priv->channels[priv->act_channel].flags & VIDEO_VC_TUNER) control(priv, TVI_CONTROL_TUN_GET_TUNER, 0); /* update local channel list */ control(priv, TVI_CONTROL_SPC_GET_INPUT, &req_chan); return(TVI_CONTROL_TRUE); case TVI_CONTROL_IMMEDIATE: priv->immediate_mode = 1; return(TVI_CONTROL_TRUE); } } return(TVI_CONTROL_UNKNOWN); } // copies a video frame // for RGB (i.e. BGR in mplayer) flips the image upside down // for YV12 swaps the 2nd and 3rd plane static inline void copy_frame(priv_t *priv, unsigned char *dest, unsigned char *source) { int i; unsigned char *sptr; // YV12 uses VIDEO_PALETTE_YUV420P, but the planes are swapped if (priv->format == IMGFMT_YV12) { memcpy(dest, source, priv->width * priv->height); memcpy(dest+priv->width * priv->height*5/4, source+priv->width * priv->height, priv->width * priv->height/4); memcpy(dest+priv->width * priv->height, source+priv->width * priv->height*5/4, priv->width * priv->height/4); return; } switch (priv->picture.palette) { case VIDEO_PALETTE_RGB24: case VIDEO_PALETTE_RGB32: case VIDEO_PALETTE_RGB555: case VIDEO_PALETTE_RGB565: sptr = source + (priv->height-1)*priv->bytesperline; for (i = 0; i < priv->height; i++) { memcpy(dest, sptr, priv->bytesperline); dest += priv->bytesperline; sptr -= priv->bytesperline; } break; case VIDEO_PALETTE_UYVY: case VIDEO_PALETTE_YUV420P: default: memcpy(dest, source, priv->bytesperline * priv->height); } } // maximum skew change, in frames #define MAX_SKEW_DELTA 0.6 static void *video_grabber(void *data) { priv_t *priv = (priv_t*)data; struct timeval curtime; double skew, prev_skew, xskew, interval, prev_interval; int frame, nextframe; int fsize = priv->bytesperline * priv->height; int i; int first = 1; int dropped = 0; double dropped_time = 0.0; double drop_delta = 0.0; /* start the capture process */ if (ioctl(priv->video_fd, VIDIOCMCAPTURE, &priv->buf[0]) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "\nioctl mcapture failed: %s\n", strerror(errno)); } while (ioctl(priv->video_fd, VIDIOCSYNC, &priv->buf[1].frame) < 0 && (errno == EAGAIN || errno == EINTR)); mp_dbg(MSGT_TV, MSGL_DBG3, "\npicture sync failed\n"); prev_interval = 0.0; prev_skew = 0.0; for (;!priv->shutdown;) { for (i = 0; i < priv->nbuf && !priv->shutdown; i++) { frame = i; nextframe = (i+1)%priv->nbuf; if (ioctl(priv->video_fd, VIDIOCMCAPTURE, &priv->buf[nextframe]) == -1) { mp_msg(MSGT_TV, MSGL_ERR, "\nioctl mcapture failed: %s\n", strerror(errno)); continue; } while (ioctl(priv->video_fd, VIDIOCSYNC, &priv->buf[frame].frame) < 0 && (errno == EAGAIN || errno == EINTR)); mp_dbg(MSGT_TV, MSGL_DBG3, "\npicture sync failed\n"); gettimeofday(&curtime, NULL); if (first) { // this was a first frame - let's launch the audio capture thread immediately // before that, just initialize some variables priv->starttime = curtime.tv_sec + curtime.tv_usec*.000001; priv->audio_skew_measure_time = 0; pthread_mutex_unlock(&priv->audio_starter); // first frame must always have timestamp of zero xskew = 0.0; skew = 0.0; interval = 0.0; first = 0; } else { interval = curtime.tv_sec + curtime.tv_usec*.000001 - priv->starttime; if (!priv->immediate_mode && ( (interval - prev_interval < 1.0/priv->fps*0.85) || (interval - prev_interval > 1.0/priv->fps*1.15) ) ) { mp_msg(MSGT_TV, MSGL_V, "\nvideo capture thread: frame delta ~ %.1lf fps\n", 1.0/(interval - prev_interval)); } // interpolate the skew in time pthread_mutex_lock(&priv->skew_mutex); xskew = priv->audio_skew + (interval - priv->audio_skew_measure_time)*priv->audio_skew_factor; pthread_mutex_unlock(&priv->skew_mutex); // correct extreme skew changes to avoid (especially) moving backwards in time if (xskew - prev_skew > (interval - prev_interval)*MAX_SKEW_DELTA) { skew = prev_skew + (interval - prev_interval)*MAX_SKEW_DELTA; } else if (xskew - prev_skew < -(interval - prev_interval)*MAX_SKEW_DELTA) { skew = prev_skew - (interval - prev_interval)*MAX_SKEW_DELTA; } else { skew = xskew; } } mp_msg(MSGT_TV, MSGL_DBG3, "\nfps = %lf, v_interval = %lf, a_skew = %f, corr_skew = %f\n", 1.0/(interval - prev_interval), interval/2, xskew, skew); mp_msg(MSGT_TV, MSGL_DBG3, "vcnt = %d, acnt = %d\n", priv->video_cnt, priv->audio_cnt); prev_skew = skew; prev_interval = interval; if ((priv->video_tail+1) % priv->video_buffer_size == priv->video_head) { if (!priv->immediate_mode) { mp_msg(MSGT_TV, MSGL_ERR, "\nvideo buffer full - dropping frame\n"); } else { if (!dropped) { dropped = 1; dropped_time = interval - skew; } } } else { if (priv->immediate_mode) { if (dropped) { drop_delta += interval - skew - dropped_time; dropped = 0; } // compensate for audio skew // negative skew => there are more audio samples, increase interval // positive skew => less samples, shorten the interval // for TV, we pretend that dropped frames never existed // without this, mplayer gets confused after pressing pause priv->video_timebuffer[priv->video_tail] = interval - skew - drop_delta; } else { priv->video_timebuffer[priv->video_tail] = interval - skew; } copy_frame(priv, priv->video_ringbuffer+(priv->video_tail)*fsize, priv->mmap+priv->mbuf.offsets[frame]); priv->video_tail = (priv->video_tail+1)%priv->video_buffer_size; priv->video_cnt++; } } } return NULL; } static double grab_video_frame(priv_t *priv, char *buffer, int len) { double interval; if (priv->first) { pthread_create(&priv->video_grabber_thread, NULL, video_grabber, priv); priv->first = 0; } while (priv->video_head == priv->video_tail) { usleep(10000); } interval = priv->video_timebuffer[priv->video_head]; memcpy(buffer, priv->video_ringbuffer+priv->video_head*priv->bytesperline * priv->height, len); priv->video_cnt--; priv->video_head = (++priv->video_head)%priv->video_buffer_size; return interval; } static int get_video_framesize(priv_t *priv) { return(priv->bytesperline * priv->height); } static void *audio_grabber(void *data) { priv_t *priv = (priv_t*)data; struct timeval tv; int i, audio_skew_ptr = 0; double tmp, current_time, prev_skew = 0.0; pthread_mutex_lock(&priv->audio_starter); audio_in_start_capture(&priv->audio_in); for (i = 0; i < priv->aud_skew_cnt; i++) priv->audio_skew_buffer[i] = 0.0; for (; !priv->shutdown;) { if (audio_in_read_chunk(&priv->audio_in, priv->audio_ringbuffer+priv->audio_tail*priv->audio_in.blocksize) < 0) continue; gettimeofday(&tv, NULL); priv->audio_recv_blocks_total++; current_time = tv.tv_sec + tv.tv_usec*.000001 - priv->starttime; // compute the moving sum of the skews if (priv->audio_recv_blocks_total % 1024 == 0) { // recompute the moving sum to avoid truncation errors priv->audio_skew_buffer[audio_skew_ptr] = current_time - priv->audio_secs_per_block*priv->audio_recv_blocks_total; audio_skew_ptr = (audio_skew_ptr+1) % priv->aud_skew_cnt; for (i = 0, tmp = 0.0; i < priv->aud_skew_cnt; i++) tmp += priv->audio_skew_buffer[i]; priv->audio_skew_total = tmp; } else { priv->audio_skew_total -= priv->audio_skew_buffer[audio_skew_ptr]; priv->audio_skew_buffer[audio_skew_ptr] = current_time - priv->audio_secs_per_block*priv->audio_recv_blocks_total; priv->audio_skew_total += priv->audio_skew_buffer[audio_skew_ptr]; audio_skew_ptr = (audio_skew_ptr+1) % priv->aud_skew_cnt; } pthread_mutex_lock(&priv->skew_mutex); // linear interpolation - here we interpolate current skew value // from the moving average, which we expect to be in the middle // of the interval if (priv->audio_recv_blocks_total > priv->aud_skew_cnt) { priv->audio_skew = priv->audio_skew_total/priv->aud_skew_cnt; priv->audio_skew += (priv->audio_skew*priv->aud_skew_cnt)/(2*priv->audio_recv_blocks_total-priv->aud_skew_cnt); } else { // priv->audio_skew = 2*priv->audio_skew_total/priv->aud_skew_cnt; priv->audio_skew = (priv->aud_skew_cnt+priv->audio_recv_blocks_total)/priv->aud_skew_cnt*priv->audio_skew_total/priv->audio_recv_blocks_total; // priv->audio_skew = current_time - priv->audio_secs_per_block*priv->audio_recv_blocks_total; } // current skew factor (assuming linearity) // used for further interpolation in video_grabber // probably overkill but seems to be necessary for // stress testing by dropping half of the audio frames ;) // especially when using ALSA with large block sizes // where audio_skew remains a long while behind priv->audio_skew_factor = (priv->audio_skew-prev_skew)/(current_time - priv->audio_skew_measure_time); priv->audio_skew_measure_time = current_time; prev_skew = priv->audio_skew; pthread_mutex_unlock(&priv->skew_mutex); if ((priv->audio_tail+1) % priv->audio_buffer_size == priv->audio_head) { mp_msg(MSGT_TV, MSGL_ERR, "\ntoo bad - dropping audio frame !\n"); priv->audio_drop++; } else { priv->audio_tail = (++priv->audio_tail) % priv->audio_buffer_size; priv->audio_cnt++; } } return NULL; } static double grab_audio_frame(priv_t *priv, char *buffer, int len) { mp_dbg(MSGT_TV, MSGL_DBG2, "grab_audio_frame(priv=%p, buffer=%p, len=%d)\n", priv, buffer, len); // compensate for dropped audio frames if (priv->audio_drop && (priv->audio_head == priv->audio_tail)) { priv->audio_drop--; priv->audio_sent_blocks_total++; memset(buffer, 0, len); return (double)priv->audio_sent_blocks_total*priv->audio_secs_per_block; } while (priv->audio_head == priv->audio_tail) { usleep(10000); } memcpy(buffer, priv->audio_ringbuffer+priv->audio_head*priv->audio_in.blocksize, len); priv->audio_head = (++priv->audio_head) % priv->audio_buffer_size; priv->audio_cnt--; priv->audio_sent_blocks_total++; return (double)priv->audio_sent_blocks_total*priv->audio_secs_per_block; } static int get_audio_framesize(priv_t *priv) { return(priv->audio_in.blocksize); } #endif /* USE_TV */