Mercurial > libavcodec.hg
view pnmenc.c @ 9950:a23ecc9d7976 libavcodec
Make VAAPI/VDPAU variant of decoder foo depend on decoder foo.
The VAAPI/VDPAU variants of foo require almost all of the code of foo.
Thus it makes little sense to enable just the accelerated variant.
This allows refactoring some object dependencies in the Makefiles.
author | diego |
---|---|
date | Sun, 12 Jul 2009 13:22:01 +0000 |
parents | 7f15f8f14e97 |
children | 38cfe222e1a4 |
line wrap: on
line source
/* * PNM image format * Copyright (c) 2002, 2003 Fabrice Bellard * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "avcodec.h" #include "bytestream.h" #include "pnm.h" static av_cold int common_init(AVCodecContext *avctx){ PNMContext *s = avctx->priv_data; avcodec_get_frame_defaults((AVFrame*)&s->picture); avctx->coded_frame= (AVFrame*)&s->picture; return 0; } static int pnm_decode_frame(AVCodecContext *avctx, void *data, int *data_size, AVPacket *avpkt) { const uint8_t *buf = avpkt->data; int buf_size = avpkt->size; PNMContext * const s = avctx->priv_data; AVFrame *picture = data; AVFrame * const p= (AVFrame*)&s->picture; int i, n, linesize, h, upgrade = 0; unsigned char *ptr; s->bytestream_start= s->bytestream= buf; s->bytestream_end= buf + buf_size; if(ff_pnm_decode_header(avctx, s) < 0) return -1; if(p->data[0]) avctx->release_buffer(avctx, p); p->reference= 0; if(avctx->get_buffer(avctx, p) < 0){ av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); return -1; } p->pict_type= FF_I_TYPE; p->key_frame= 1; switch(avctx->pix_fmt) { default: return -1; case PIX_FMT_RGB48BE: n = avctx->width * 6; goto do_read; case PIX_FMT_RGB24: n = avctx->width * 3; goto do_read; case PIX_FMT_GRAY8: n = avctx->width; if (s->maxval < 255) upgrade = 1; goto do_read; case PIX_FMT_GRAY16BE: case PIX_FMT_GRAY16LE: n = avctx->width * 2; if (s->maxval < 65535) upgrade = 2; goto do_read; case PIX_FMT_MONOWHITE: case PIX_FMT_MONOBLACK: n = (avctx->width + 7) >> 3; do_read: ptr = p->data[0]; linesize = p->linesize[0]; if(s->bytestream + n*avctx->height > s->bytestream_end) return -1; for(i = 0; i < avctx->height; i++) { if (!upgrade) memcpy(ptr, s->bytestream, n); else if (upgrade == 1) { unsigned int j, f = (255*128 + s->maxval/2) / s->maxval; for (j=0; j<n; j++) ptr[j] = (s->bytestream[j] * f + 64) >> 7; } else if (upgrade == 2) { unsigned int j, v, f = (65535*32768 + s->maxval/2) / s->maxval; for (j=0; j<n/2; j++) { v = be2me_16(((uint16_t *)s->bytestream)[j]); ((uint16_t *)ptr)[j] = (v * f + 16384) >> 15; } } s->bytestream += n; ptr += linesize; } break; case PIX_FMT_YUV420P: { unsigned char *ptr1, *ptr2; n = avctx->width; ptr = p->data[0]; linesize = p->linesize[0]; if(s->bytestream + n*avctx->height*3/2 > s->bytestream_end) return -1; for(i = 0; i < avctx->height; i++) { memcpy(ptr, s->bytestream, n); s->bytestream += n; ptr += linesize; } ptr1 = p->data[1]; ptr2 = p->data[2]; n >>= 1; h = avctx->height >> 1; for(i = 0; i < h; i++) { memcpy(ptr1, s->bytestream, n); s->bytestream += n; memcpy(ptr2, s->bytestream, n); s->bytestream += n; ptr1 += p->linesize[1]; ptr2 += p->linesize[2]; } } break; case PIX_FMT_RGB32: ptr = p->data[0]; linesize = p->linesize[0]; if(s->bytestream + avctx->width*avctx->height*4 > s->bytestream_end) return -1; for(i = 0; i < avctx->height; i++) { int j, r, g, b, a; for(j = 0;j < avctx->width; j++) { r = *s->bytestream++; g = *s->bytestream++; b = *s->bytestream++; a = *s->bytestream++; ((uint32_t *)ptr)[j] = (a << 24) | (r << 16) | (g << 8) | b; } ptr += linesize; } break; } *picture= *(AVFrame*)&s->picture; *data_size = sizeof(AVPicture); return s->bytestream - s->bytestream_start; } static int pnm_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){ PNMContext *s = avctx->priv_data; AVFrame *pict = data; AVFrame * const p= (AVFrame*)&s->picture; int i, h, h1, c, n, linesize; uint8_t *ptr, *ptr1, *ptr2; if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){ av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n"); return -1; } *p = *pict; p->pict_type= FF_I_TYPE; p->key_frame= 1; s->bytestream_start= s->bytestream= outbuf; s->bytestream_end= outbuf+buf_size; h = avctx->height; h1 = h; switch(avctx->pix_fmt) { case PIX_FMT_MONOWHITE: c = '4'; n = (avctx->width + 7) >> 3; break; case PIX_FMT_GRAY8: c = '5'; n = avctx->width; break; case PIX_FMT_GRAY16BE: c = '5'; n = avctx->width * 2; break; case PIX_FMT_RGB24: c = '6'; n = avctx->width * 3; break; case PIX_FMT_RGB48BE: c = '6'; n = avctx->width * 6; break; case PIX_FMT_YUV420P: c = '5'; n = avctx->width; h1 = (h * 3) / 2; break; default: return -1; } snprintf(s->bytestream, s->bytestream_end - s->bytestream, "P%c\n%d %d\n", c, avctx->width, h1); s->bytestream += strlen(s->bytestream); if (avctx->pix_fmt != PIX_FMT_MONOWHITE) { snprintf(s->bytestream, s->bytestream_end - s->bytestream, "%d\n", (avctx->pix_fmt != PIX_FMT_GRAY16BE && avctx->pix_fmt != PIX_FMT_RGB48BE) ? 255 : 65535); s->bytestream += strlen(s->bytestream); } ptr = p->data[0]; linesize = p->linesize[0]; for(i=0;i<h;i++) { memcpy(s->bytestream, ptr, n); s->bytestream += n; ptr += linesize; } if (avctx->pix_fmt == PIX_FMT_YUV420P) { h >>= 1; n >>= 1; ptr1 = p->data[1]; ptr2 = p->data[2]; for(i=0;i<h;i++) { memcpy(s->bytestream, ptr1, n); s->bytestream += n; memcpy(s->bytestream, ptr2, n); s->bytestream += n; ptr1 += p->linesize[1]; ptr2 += p->linesize[2]; } } return s->bytestream - s->bytestream_start; } static int pam_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){ PNMContext *s = avctx->priv_data; AVFrame *pict = data; AVFrame * const p= (AVFrame*)&s->picture; int i, h, w, n, linesize, depth, maxval; const char *tuple_type; uint8_t *ptr; if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){ av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n"); return -1; } *p = *pict; p->pict_type= FF_I_TYPE; p->key_frame= 1; s->bytestream_start= s->bytestream= outbuf; s->bytestream_end= outbuf+buf_size; h = avctx->height; w = avctx->width; switch(avctx->pix_fmt) { case PIX_FMT_MONOWHITE: n = (w + 7) >> 3; depth = 1; maxval = 1; tuple_type = "BLACKANDWHITE"; break; case PIX_FMT_GRAY8: n = w; depth = 1; maxval = 255; tuple_type = "GRAYSCALE"; break; case PIX_FMT_RGB24: n = w * 3; depth = 3; maxval = 255; tuple_type = "RGB"; break; case PIX_FMT_RGB32: n = w * 4; depth = 4; maxval = 255; tuple_type = "RGB_ALPHA"; break; default: return -1; } snprintf(s->bytestream, s->bytestream_end - s->bytestream, "P7\nWIDTH %d\nHEIGHT %d\nDEPTH %d\nMAXVAL %d\nTUPLETYPE %s\nENDHDR\n", w, h, depth, maxval, tuple_type); s->bytestream += strlen(s->bytestream); ptr = p->data[0]; linesize = p->linesize[0]; if (avctx->pix_fmt == PIX_FMT_RGB32) { int j; unsigned int v; for(i=0;i<h;i++) { for(j=0;j<w;j++) { v = ((uint32_t *)ptr)[j]; bytestream_put_be24(&s->bytestream, v); *s->bytestream++ = v >> 24; } ptr += linesize; } } else { for(i=0;i<h;i++) { memcpy(s->bytestream, ptr, n); s->bytestream += n; ptr += linesize; } } return s->bytestream - s->bytestream_start; } #if 0 static int pnm_probe(AVProbeData *pd) { const char *p = pd->buf; if (pd->buf_size >= 8 && p[0] == 'P' && p[1] >= '4' && p[1] <= '6' && pnm_space(p[2]) ) return AVPROBE_SCORE_MAX - 1; /* to permit pgmyuv probe */ else return 0; } static int pgmyuv_probe(AVProbeData *pd) { if (match_ext(pd->filename, "pgmyuv")) return AVPROBE_SCORE_MAX; else return 0; } static int pam_probe(AVProbeData *pd) { const char *p = pd->buf; if (pd->buf_size >= 8 && p[0] == 'P' && p[1] == '7' && p[2] == '\n') return AVPROBE_SCORE_MAX; else return 0; } #endif #if CONFIG_PGM_DECODER AVCodec pgm_decoder = { "pgm", CODEC_TYPE_VIDEO, CODEC_ID_PGM, sizeof(PNMContext), common_init, NULL, NULL, pnm_decode_frame, CODEC_CAP_DR1, .pix_fmts= (enum PixelFormat[]){PIX_FMT_GRAY8, PIX_FMT_GRAY16BE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PGM (Portable GrayMap) image"), }; #endif #if CONFIG_PGM_ENCODER AVCodec pgm_encoder = { "pgm", CODEC_TYPE_VIDEO, CODEC_ID_PGM, sizeof(PNMContext), common_init, pnm_encode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_GRAY8, PIX_FMT_GRAY16BE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PGM (Portable GrayMap) image"), }; #endif // CONFIG_PGM_ENCODER #if CONFIG_PGMYUV_DECODER AVCodec pgmyuv_decoder = { "pgmyuv", CODEC_TYPE_VIDEO, CODEC_ID_PGMYUV, sizeof(PNMContext), common_init, NULL, NULL, pnm_decode_frame, CODEC_CAP_DR1, .pix_fmts= (enum PixelFormat[]){PIX_FMT_YUV420P, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PGMYUV (Portable GrayMap YUV) image"), }; #endif #if CONFIG_PGMYUV_ENCODER AVCodec pgmyuv_encoder = { "pgmyuv", CODEC_TYPE_VIDEO, CODEC_ID_PGMYUV, sizeof(PNMContext), common_init, pnm_encode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_YUV420P, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PGMYUV (Portable GrayMap YUV) image"), }; #endif // CONFIG_PGMYUV_ENCODER #if CONFIG_PPM_DECODER AVCodec ppm_decoder = { "ppm", CODEC_TYPE_VIDEO, CODEC_ID_PPM, sizeof(PNMContext), common_init, NULL, NULL, pnm_decode_frame, CODEC_CAP_DR1, .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB48BE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PPM (Portable PixelMap) image"), }; #endif #if CONFIG_PPM_ENCODER AVCodec ppm_encoder = { "ppm", CODEC_TYPE_VIDEO, CODEC_ID_PPM, sizeof(PNMContext), common_init, pnm_encode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB48BE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PPM (Portable PixelMap) image"), }; #endif // CONFIG_PPM_ENCODER #if CONFIG_PBM_DECODER AVCodec pbm_decoder = { "pbm", CODEC_TYPE_VIDEO, CODEC_ID_PBM, sizeof(PNMContext), common_init, NULL, NULL, pnm_decode_frame, CODEC_CAP_DR1, .pix_fmts= (enum PixelFormat[]){PIX_FMT_MONOWHITE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PBM (Portable BitMap) image"), }; #endif #if CONFIG_PBM_ENCODER AVCodec pbm_encoder = { "pbm", CODEC_TYPE_VIDEO, CODEC_ID_PBM, sizeof(PNMContext), common_init, pnm_encode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_MONOWHITE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PBM (Portable BitMap) image"), }; #endif // CONFIG_PBM_ENCODER #if CONFIG_PAM_DECODER AVCodec pam_decoder = { "pam", CODEC_TYPE_VIDEO, CODEC_ID_PAM, sizeof(PNMContext), common_init, NULL, NULL, pnm_decode_frame, CODEC_CAP_DR1, .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB32, PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PAM (Portable AnyMap) image"), }; #endif #if CONFIG_PAM_ENCODER AVCodec pam_encoder = { "pam", CODEC_TYPE_VIDEO, CODEC_ID_PAM, sizeof(PNMContext), common_init, pam_encode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB32, PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PAM (Portable AnyMap) image"), }; #endif // CONFIG_PAM_ENCODER