Mercurial > libavcodec.hg
view pnmenc.c @ 8457:fa07932f2c89 libavcodec
flatten an array, since gcc fails at optimizing multidimensional arrays
h264_idct8_add_c: 780 -> 735 cycles on conroe
author | lorenm |
---|---|
date | Thu, 25 Dec 2008 01:20:37 +0000 |
parents | e943e1409077 |
children | 7a463923ecd1 |
line wrap: on
line source
/* * PNM image format * Copyright (c) 2002, 2003 Fabrice Bellard. * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "avcodec.h" #include "bytestream.h" #include "pnm.h" static av_cold int common_init(AVCodecContext *avctx){ PNMContext *s = avctx->priv_data; avcodec_get_frame_defaults((AVFrame*)&s->picture); avctx->coded_frame= (AVFrame*)&s->picture; return 0; } static int pnm_decode_frame(AVCodecContext *avctx, void *data, int *data_size, const uint8_t *buf, int buf_size) { PNMContext * const s = avctx->priv_data; AVFrame *picture = data; AVFrame * const p= (AVFrame*)&s->picture; int i, n, linesize, h, upgrade = 0; unsigned char *ptr; s->bytestream_start= s->bytestream= buf; s->bytestream_end= buf + buf_size; if(ff_pnm_decode_header(avctx, s) < 0) return -1; if(p->data[0]) avctx->release_buffer(avctx, p); p->reference= 0; if(avctx->get_buffer(avctx, p) < 0){ av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); return -1; } p->pict_type= FF_I_TYPE; p->key_frame= 1; switch(avctx->pix_fmt) { default: return -1; case PIX_FMT_RGB24: n = avctx->width * 3; goto do_read; case PIX_FMT_GRAY8: n = avctx->width; if (s->maxval < 255) upgrade = 1; goto do_read; case PIX_FMT_GRAY16BE: case PIX_FMT_GRAY16LE: n = avctx->width * 2; if (s->maxval < 65535) upgrade = 2; goto do_read; case PIX_FMT_MONOWHITE: case PIX_FMT_MONOBLACK: n = (avctx->width + 7) >> 3; do_read: ptr = p->data[0]; linesize = p->linesize[0]; if(s->bytestream + n*avctx->height > s->bytestream_end) return -1; for(i = 0; i < avctx->height; i++) { if (!upgrade) memcpy(ptr, s->bytestream, n); else if (upgrade == 1) { unsigned int j, f = (255*128 + s->maxval/2) / s->maxval; for (j=0; j<n; j++) ptr[j] = (s->bytestream[j] * f + 64) >> 7; } else if (upgrade == 2) { unsigned int j, v, f = (65535*32768 + s->maxval/2) / s->maxval; for (j=0; j<n/2; j++) { v = be2me_16(((uint16_t *)s->bytestream)[j]); ((uint16_t *)ptr)[j] = (v * f + 16384) >> 15; } } s->bytestream += n; ptr += linesize; } break; case PIX_FMT_YUV420P: { unsigned char *ptr1, *ptr2; n = avctx->width; ptr = p->data[0]; linesize = p->linesize[0]; if(s->bytestream + n*avctx->height*3/2 > s->bytestream_end) return -1; for(i = 0; i < avctx->height; i++) { memcpy(ptr, s->bytestream, n); s->bytestream += n; ptr += linesize; } ptr1 = p->data[1]; ptr2 = p->data[2]; n >>= 1; h = avctx->height >> 1; for(i = 0; i < h; i++) { memcpy(ptr1, s->bytestream, n); s->bytestream += n; memcpy(ptr2, s->bytestream, n); s->bytestream += n; ptr1 += p->linesize[1]; ptr2 += p->linesize[2]; } } break; case PIX_FMT_RGB32: ptr = p->data[0]; linesize = p->linesize[0]; if(s->bytestream + avctx->width*avctx->height*4 > s->bytestream_end) return -1; for(i = 0; i < avctx->height; i++) { int j, r, g, b, a; for(j = 0;j < avctx->width; j++) { r = *s->bytestream++; g = *s->bytestream++; b = *s->bytestream++; a = *s->bytestream++; ((uint32_t *)ptr)[j] = (a << 24) | (r << 16) | (g << 8) | b; } ptr += linesize; } break; } *picture= *(AVFrame*)&s->picture; *data_size = sizeof(AVPicture); return s->bytestream - s->bytestream_start; } static int pnm_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){ PNMContext *s = avctx->priv_data; AVFrame *pict = data; AVFrame * const p= (AVFrame*)&s->picture; int i, h, h1, c, n, linesize; uint8_t *ptr, *ptr1, *ptr2; if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){ av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n"); return -1; } *p = *pict; p->pict_type= FF_I_TYPE; p->key_frame= 1; s->bytestream_start= s->bytestream= outbuf; s->bytestream_end= outbuf+buf_size; h = avctx->height; h1 = h; switch(avctx->pix_fmt) { case PIX_FMT_MONOWHITE: c = '4'; n = (avctx->width + 7) >> 3; break; case PIX_FMT_GRAY8: c = '5'; n = avctx->width; break; case PIX_FMT_GRAY16BE: c = '5'; n = avctx->width * 2; break; case PIX_FMT_RGB24: c = '6'; n = avctx->width * 3; break; case PIX_FMT_YUV420P: c = '5'; n = avctx->width; h1 = (h * 3) / 2; break; default: return -1; } snprintf(s->bytestream, s->bytestream_end - s->bytestream, "P%c\n%d %d\n", c, avctx->width, h1); s->bytestream += strlen(s->bytestream); if (avctx->pix_fmt != PIX_FMT_MONOWHITE) { snprintf(s->bytestream, s->bytestream_end - s->bytestream, "%d\n", (avctx->pix_fmt != PIX_FMT_GRAY16BE) ? 255 : 65535); s->bytestream += strlen(s->bytestream); } ptr = p->data[0]; linesize = p->linesize[0]; for(i=0;i<h;i++) { memcpy(s->bytestream, ptr, n); s->bytestream += n; ptr += linesize; } if (avctx->pix_fmt == PIX_FMT_YUV420P) { h >>= 1; n >>= 1; ptr1 = p->data[1]; ptr2 = p->data[2]; for(i=0;i<h;i++) { memcpy(s->bytestream, ptr1, n); s->bytestream += n; memcpy(s->bytestream, ptr2, n); s->bytestream += n; ptr1 += p->linesize[1]; ptr2 += p->linesize[2]; } } return s->bytestream - s->bytestream_start; } static int pam_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){ PNMContext *s = avctx->priv_data; AVFrame *pict = data; AVFrame * const p= (AVFrame*)&s->picture; int i, h, w, n, linesize, depth, maxval; const char *tuple_type; uint8_t *ptr; if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){ av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n"); return -1; } *p = *pict; p->pict_type= FF_I_TYPE; p->key_frame= 1; s->bytestream_start= s->bytestream= outbuf; s->bytestream_end= outbuf+buf_size; h = avctx->height; w = avctx->width; switch(avctx->pix_fmt) { case PIX_FMT_MONOWHITE: n = (w + 7) >> 3; depth = 1; maxval = 1; tuple_type = "BLACKANDWHITE"; break; case PIX_FMT_GRAY8: n = w; depth = 1; maxval = 255; tuple_type = "GRAYSCALE"; break; case PIX_FMT_RGB24: n = w * 3; depth = 3; maxval = 255; tuple_type = "RGB"; break; case PIX_FMT_RGB32: n = w * 4; depth = 4; maxval = 255; tuple_type = "RGB_ALPHA"; break; default: return -1; } snprintf(s->bytestream, s->bytestream_end - s->bytestream, "P7\nWIDTH %d\nHEIGHT %d\nDEPTH %d\nMAXVAL %d\nTUPLETYPE %s\nENDHDR\n", w, h, depth, maxval, tuple_type); s->bytestream += strlen(s->bytestream); ptr = p->data[0]; linesize = p->linesize[0]; if (avctx->pix_fmt == PIX_FMT_RGB32) { int j; unsigned int v; for(i=0;i<h;i++) { for(j=0;j<w;j++) { v = ((uint32_t *)ptr)[j]; bytestream_put_be24(&s->bytestream, v); *s->bytestream++ = v >> 24; } ptr += linesize; } } else { for(i=0;i<h;i++) { memcpy(s->bytestream, ptr, n); s->bytestream += n; ptr += linesize; } } return s->bytestream - s->bytestream_start; } #if 0 static int pnm_probe(AVProbeData *pd) { const char *p = pd->buf; if (pd->buf_size >= 8 && p[0] == 'P' && p[1] >= '4' && p[1] <= '6' && pnm_space(p[2]) ) return AVPROBE_SCORE_MAX - 1; /* to permit pgmyuv probe */ else return 0; } static int pgmyuv_probe(AVProbeData *pd) { if (match_ext(pd->filename, "pgmyuv")) return AVPROBE_SCORE_MAX; else return 0; } static int pam_probe(AVProbeData *pd) { const char *p = pd->buf; if (pd->buf_size >= 8 && p[0] == 'P' && p[1] == '7' && p[2] == '\n') return AVPROBE_SCORE_MAX; else return 0; } #endif #ifdef CONFIG_PGM_ENCODER AVCodec pgm_encoder = { "pgm", CODEC_TYPE_VIDEO, CODEC_ID_PGM, sizeof(PNMContext), common_init, pnm_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_GRAY8, PIX_FMT_GRAY16BE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PGM (Portable GrayMap) image"), }; #endif // CONFIG_PGM_ENCODER #ifdef CONFIG_PGMYUV_ENCODER AVCodec pgmyuv_encoder = { "pgmyuv", CODEC_TYPE_VIDEO, CODEC_ID_PGMYUV, sizeof(PNMContext), common_init, pnm_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_YUV420P, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PGMYUV (Portable GrayMap YUV) image"), }; #endif // CONFIG_PGMYUV_ENCODER #ifdef CONFIG_PPM_ENCODER AVCodec ppm_encoder = { "ppm", CODEC_TYPE_VIDEO, CODEC_ID_PPM, sizeof(PNMContext), common_init, pnm_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PPM (Portable PixelMap) image"), }; #endif // CONFIG_PPM_ENCODER #ifdef CONFIG_PBM_ENCODER AVCodec pbm_encoder = { "pbm", CODEC_TYPE_VIDEO, CODEC_ID_PBM, sizeof(PNMContext), common_init, pnm_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_MONOWHITE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PBM (Portable BitMap) image"), }; #endif // CONFIG_PBM_ENCODER #ifdef CONFIG_PAM_ENCODER AVCodec pam_encoder = { "pam", CODEC_TYPE_VIDEO, CODEC_ID_PAM, sizeof(PNMContext), common_init, pam_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGB32, PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("PAM (Portable AnyMap) image"), }; #endif // CONFIG_PAM_ENCODER