Mercurial > libavcodec.hg
view pnm.c @ 4166:eced83504436 libavcodec
mp3 header (de)compression bitstream filter
this will make mp3 frames 4 bytes smaller, it will not give you binary identical mp3 files, but it will give you mp3 files which decode to binary identical output
this will only work in containers providing at least packet size, sample_rate and number of channels
bugreports about mp3 files for which this fails are welcome
and this is experimental (dont expect compatibility and dont even expect to be able to decompress what you compressed, hell dont even expect this to work without editing the source a little)
author | michael |
---|---|
date | Fri, 10 Nov 2006 01:41:53 +0000 |
parents | e893c1438b36 |
children | ce643a22f049 |
line wrap: on
line source
/* * PNM image format * Copyright (c) 2002, 2003 Fabrice Bellard. * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "avcodec.h" #include "parser.h" //for ParseContext typedef struct PNMContext { uint8_t *bytestream; uint8_t *bytestream_start; uint8_t *bytestream_end; AVFrame picture; } PNMContext; static inline int pnm_space(int c) { return (c == ' ' || c == '\n' || c == '\r' || c == '\t'); } static void pnm_get(PNMContext *sc, char *str, int buf_size) { char *s; int c; /* skip spaces and comments */ for(;;) { c = *sc->bytestream++; if (c == '#') { do { c = *sc->bytestream++; } while (c != '\n' && sc->bytestream < sc->bytestream_end); } else if (!pnm_space(c)) { break; } } s = str; while (sc->bytestream < sc->bytestream_end && !pnm_space(c)) { if ((s - str) < buf_size - 1) *s++ = c; c = *sc->bytestream++; } *s = '\0'; } static int common_init(AVCodecContext *avctx){ PNMContext *s = avctx->priv_data; avcodec_get_frame_defaults((AVFrame*)&s->picture); avctx->coded_frame= (AVFrame*)&s->picture; return 0; } static int pnm_decode_header(AVCodecContext *avctx, PNMContext * const s){ char buf1[32], tuple_type[32]; int h, w, depth, maxval; pnm_get(s, buf1, sizeof(buf1)); if (!strcmp(buf1, "P4")) { avctx->pix_fmt = PIX_FMT_MONOWHITE; } else if (!strcmp(buf1, "P5")) { if (avctx->codec_id == CODEC_ID_PGMYUV) avctx->pix_fmt = PIX_FMT_YUV420P; else avctx->pix_fmt = PIX_FMT_GRAY8; } else if (!strcmp(buf1, "P6")) { avctx->pix_fmt = PIX_FMT_RGB24; } else if (!strcmp(buf1, "P7")) { w = -1; h = -1; maxval = -1; depth = -1; tuple_type[0] = '\0'; for(;;) { pnm_get(s, buf1, sizeof(buf1)); if (!strcmp(buf1, "WIDTH")) { pnm_get(s, buf1, sizeof(buf1)); w = strtol(buf1, NULL, 10); } else if (!strcmp(buf1, "HEIGHT")) { pnm_get(s, buf1, sizeof(buf1)); h = strtol(buf1, NULL, 10); } else if (!strcmp(buf1, "DEPTH")) { pnm_get(s, buf1, sizeof(buf1)); depth = strtol(buf1, NULL, 10); } else if (!strcmp(buf1, "MAXVAL")) { pnm_get(s, buf1, sizeof(buf1)); maxval = strtol(buf1, NULL, 10); } else if (!strcmp(buf1, "TUPLETYPE")) { pnm_get(s, tuple_type, sizeof(tuple_type)); } else if (!strcmp(buf1, "ENDHDR")) { break; } else { return -1; } } /* check that all tags are present */ if (w <= 0 || h <= 0 || maxval <= 0 || depth <= 0 || tuple_type[0] == '\0' || avcodec_check_dimensions(avctx, w, h)) return -1; avctx->width = w; avctx->height = h; if (depth == 1) { if (maxval == 1) avctx->pix_fmt = PIX_FMT_MONOWHITE; else avctx->pix_fmt = PIX_FMT_GRAY8; } else if (depth == 3) { avctx->pix_fmt = PIX_FMT_RGB24; } else if (depth == 4) { avctx->pix_fmt = PIX_FMT_RGBA32; } else { return -1; } return 0; } else { return -1; } pnm_get(s, buf1, sizeof(buf1)); avctx->width = atoi(buf1); if (avctx->width <= 0) return -1; pnm_get(s, buf1, sizeof(buf1)); avctx->height = atoi(buf1); if(avcodec_check_dimensions(avctx, avctx->width, avctx->height)) return -1; if (avctx->pix_fmt != PIX_FMT_MONOWHITE) { pnm_get(s, buf1, sizeof(buf1)); if(atoi(buf1) == 65535 && avctx->pix_fmt == PIX_FMT_GRAY8) avctx->pix_fmt = PIX_FMT_GRAY16BE; } /* more check if YUV420 */ if (avctx->pix_fmt == PIX_FMT_YUV420P) { if ((avctx->width & 1) != 0) return -1; h = (avctx->height * 2); if ((h % 3) != 0) return -1; h /= 3; avctx->height = h; } return 0; } static int pnm_decode_frame(AVCodecContext *avctx, void *data, int *data_size, uint8_t *buf, int buf_size) { PNMContext * const s = avctx->priv_data; AVFrame *picture = data; AVFrame * const p= (AVFrame*)&s->picture; int i, n, linesize, h; unsigned char *ptr; s->bytestream_start= s->bytestream= buf; s->bytestream_end= buf + buf_size; if(pnm_decode_header(avctx, s) < 0) return -1; if(p->data[0]) avctx->release_buffer(avctx, p); p->reference= 0; if(avctx->get_buffer(avctx, p) < 0){ av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); return -1; } p->pict_type= FF_I_TYPE; p->key_frame= 1; switch(avctx->pix_fmt) { default: return -1; case PIX_FMT_RGB24: n = avctx->width * 3; goto do_read; case PIX_FMT_GRAY8: n = avctx->width; goto do_read; case PIX_FMT_GRAY16BE: n = avctx->width * 2; goto do_read; case PIX_FMT_MONOWHITE: case PIX_FMT_MONOBLACK: n = (avctx->width + 7) >> 3; do_read: ptr = p->data[0]; linesize = p->linesize[0]; if(s->bytestream + n*avctx->height > s->bytestream_end) return -1; for(i = 0; i < avctx->height; i++) { memcpy(ptr, s->bytestream, n); s->bytestream += n; ptr += linesize; } break; case PIX_FMT_YUV420P: { unsigned char *ptr1, *ptr2; n = avctx->width; ptr = p->data[0]; linesize = p->linesize[0]; if(s->bytestream + n*avctx->height*3/2 > s->bytestream_end) return -1; for(i = 0; i < avctx->height; i++) { memcpy(ptr, s->bytestream, n); s->bytestream += n; ptr += linesize; } ptr1 = p->data[1]; ptr2 = p->data[2]; n >>= 1; h = avctx->height >> 1; for(i = 0; i < h; i++) { memcpy(ptr1, s->bytestream, n); s->bytestream += n; memcpy(ptr2, s->bytestream, n); s->bytestream += n; ptr1 += p->linesize[1]; ptr2 += p->linesize[2]; } } break; case PIX_FMT_RGBA32: ptr = p->data[0]; linesize = p->linesize[0]; if(s->bytestream + avctx->width*avctx->height*4 > s->bytestream_end) return -1; for(i = 0; i < avctx->height; i++) { int j, r, g, b, a; for(j = 0;j < avctx->width; j++) { r = *s->bytestream++; g = *s->bytestream++; b = *s->bytestream++; a = *s->bytestream++; ((uint32_t *)ptr)[j] = (a << 24) | (r << 16) | (g << 8) | b; } ptr += linesize; } break; } *picture= *(AVFrame*)&s->picture; *data_size = sizeof(AVPicture); return s->bytestream - s->bytestream_start; } static int pnm_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){ PNMContext *s = avctx->priv_data; AVFrame *pict = data; AVFrame * const p= (AVFrame*)&s->picture; int i, h, h1, c, n, linesize; uint8_t *ptr, *ptr1, *ptr2; if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){ av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n"); return -1; } *p = *pict; p->pict_type= FF_I_TYPE; p->key_frame= 1; s->bytestream_start= s->bytestream= outbuf; s->bytestream_end= outbuf+buf_size; h = avctx->height; h1 = h; switch(avctx->pix_fmt) { case PIX_FMT_MONOWHITE: c = '4'; n = (avctx->width + 7) >> 3; break; case PIX_FMT_GRAY8: c = '5'; n = avctx->width; break; case PIX_FMT_GRAY16BE: c = '5'; n = avctx->width * 2; break; case PIX_FMT_RGB24: c = '6'; n = avctx->width * 3; break; case PIX_FMT_YUV420P: c = '5'; n = avctx->width; h1 = (h * 3) / 2; break; default: return -1; } snprintf(s->bytestream, s->bytestream_end - s->bytestream, "P%c\n%d %d\n", c, avctx->width, h1); s->bytestream += strlen(s->bytestream); if (avctx->pix_fmt != PIX_FMT_MONOWHITE) { snprintf(s->bytestream, s->bytestream_end - s->bytestream, "%d\n", (avctx->pix_fmt != PIX_FMT_GRAY16BE) ? 255 : 65535); s->bytestream += strlen(s->bytestream); } ptr = p->data[0]; linesize = p->linesize[0]; for(i=0;i<h;i++) { memcpy(s->bytestream, ptr, n); s->bytestream += n; ptr += linesize; } if (avctx->pix_fmt == PIX_FMT_YUV420P) { h >>= 1; n >>= 1; ptr1 = p->data[1]; ptr2 = p->data[2]; for(i=0;i<h;i++) { memcpy(s->bytestream, ptr1, n); s->bytestream += n; memcpy(s->bytestream, ptr2, n); s->bytestream += n; ptr1 += p->linesize[1]; ptr2 += p->linesize[2]; } } return s->bytestream - s->bytestream_start; } static int pam_encode_frame(AVCodecContext *avctx, unsigned char *outbuf, int buf_size, void *data){ PNMContext *s = avctx->priv_data; AVFrame *pict = data; AVFrame * const p= (AVFrame*)&s->picture; int i, h, w, n, linesize, depth, maxval; const char *tuple_type; uint8_t *ptr; if(buf_size < avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height) + 200){ av_log(avctx, AV_LOG_ERROR, "encoded frame too large\n"); return -1; } *p = *pict; p->pict_type= FF_I_TYPE; p->key_frame= 1; s->bytestream_start= s->bytestream= outbuf; s->bytestream_end= outbuf+buf_size; h = avctx->height; w = avctx->width; switch(avctx->pix_fmt) { case PIX_FMT_MONOWHITE: n = (w + 7) >> 3; depth = 1; maxval = 1; tuple_type = "BLACKANDWHITE"; break; case PIX_FMT_GRAY8: n = w; depth = 1; maxval = 255; tuple_type = "GRAYSCALE"; break; case PIX_FMT_RGB24: n = w * 3; depth = 3; maxval = 255; tuple_type = "RGB"; break; case PIX_FMT_RGBA32: n = w * 4; depth = 4; maxval = 255; tuple_type = "RGB_ALPHA"; break; default: return -1; } snprintf(s->bytestream, s->bytestream_end - s->bytestream, "P7\nWIDTH %d\nHEIGHT %d\nDEPTH %d\nMAXVAL %d\nTUPLETYPE %s\nENDHDR\n", w, h, depth, maxval, tuple_type); s->bytestream += strlen(s->bytestream); ptr = p->data[0]; linesize = p->linesize[0]; if (avctx->pix_fmt == PIX_FMT_RGBA32) { int j; unsigned int v; for(i=0;i<h;i++) { for(j=0;j<w;j++) { v = ((uint32_t *)ptr)[j]; *s->bytestream++ = v >> 16; *s->bytestream++ = v >> 8; *s->bytestream++ = v; *s->bytestream++ = v >> 24; } ptr += linesize; } } else { for(i=0;i<h;i++) { memcpy(s->bytestream, ptr, n); s->bytestream += n; ptr += linesize; } } return s->bytestream - s->bytestream_start; } #if 0 static int pnm_probe(AVProbeData *pd) { const char *p = pd->buf; if (pd->buf_size >= 8 && p[0] == 'P' && p[1] >= '4' && p[1] <= '6' && pnm_space(p[2]) ) return AVPROBE_SCORE_MAX - 1; /* to permit pgmyuv probe */ else return 0; } static int pgmyuv_probe(AVProbeData *pd) { if (match_ext(pd->filename, "pgmyuv")) return AVPROBE_SCORE_MAX; else return 0; } static int pam_probe(AVProbeData *pd) { const char *p = pd->buf; if (pd->buf_size >= 8 && p[0] == 'P' && p[1] == '7' && p[2] == '\n') return AVPROBE_SCORE_MAX; else return 0; } #endif #ifdef CONFIG_PNM_PARSER static int pnm_parse(AVCodecParserContext *s, AVCodecContext *avctx, uint8_t **poutbuf, int *poutbuf_size, const uint8_t *buf, int buf_size) { ParseContext *pc = s->priv_data; PNMContext pnmctx; int next; for(; pc->overread>0; pc->overread--){ pc->buffer[pc->index++]= pc->buffer[pc->overread_index++]; } retry: if(pc->index){ pnmctx.bytestream_start= pnmctx.bytestream= pc->buffer; pnmctx.bytestream_end= pc->buffer + pc->index; }else{ pnmctx.bytestream_start= pnmctx.bytestream= (uint8_t *) buf; /* casts avoid warnings */ pnmctx.bytestream_end= (uint8_t *) buf + buf_size; } if(pnm_decode_header(avctx, &pnmctx) < 0){ if(pnmctx.bytestream < pnmctx.bytestream_end){ if(pc->index){ pc->index=0; }else{ buf++; buf_size--; } goto retry; } #if 0 if(pc->index && pc->index*2 + FF_INPUT_BUFFER_PADDING_SIZE < pc->buffer_size && buf_size > pc->index){ memcpy(pc->buffer + pc->index, buf, pc->index); pc->index += pc->index; buf += pc->index; buf_size -= pc->index; goto retry; } #endif next= END_NOT_FOUND; }else{ next= pnmctx.bytestream - pnmctx.bytestream_start + avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height); if(pnmctx.bytestream_start!=buf) next-= pc->index; if(next > buf_size) next= END_NOT_FOUND; } if(ff_combine_frame(pc, next, (uint8_t **)&buf, &buf_size)<0){ *poutbuf = NULL; *poutbuf_size = 0; return buf_size; } *poutbuf = (uint8_t *)buf; *poutbuf_size = buf_size; return next; } AVCodecParser pnm_parser = { { CODEC_ID_PGM, CODEC_ID_PGMYUV, CODEC_ID_PPM, CODEC_ID_PBM, CODEC_ID_PAM}, sizeof(ParseContext), NULL, pnm_parse, ff_parse_close, }; #endif /* CONFIG_PNM_PARSER */ #ifdef CONFIG_PGM_ENCODER AVCodec pgm_encoder = { "pgm", CODEC_TYPE_VIDEO, CODEC_ID_PGM, sizeof(PNMContext), common_init, pnm_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_GRAY8, PIX_FMT_GRAY16BE, -1}, }; #endif // CONFIG_PGM_ENCODER #ifdef CONFIG_PGMYUV_ENCODER AVCodec pgmyuv_encoder = { "pgmyuv", CODEC_TYPE_VIDEO, CODEC_ID_PGMYUV, sizeof(PNMContext), common_init, pnm_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_YUV420P, -1}, }; #endif // CONFIG_PGMYUV_ENCODER #ifdef CONFIG_PPM_ENCODER AVCodec ppm_encoder = { "ppm", CODEC_TYPE_VIDEO, CODEC_ID_PPM, sizeof(PNMContext), common_init, pnm_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, -1}, }; #endif // CONFIG_PPM_ENCODER #ifdef CONFIG_PBM_ENCODER AVCodec pbm_encoder = { "pbm", CODEC_TYPE_VIDEO, CODEC_ID_PBM, sizeof(PNMContext), common_init, pnm_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_MONOWHITE, -1}, }; #endif // CONFIG_PBM_ENCODER #ifdef CONFIG_PAM_ENCODER AVCodec pam_encoder = { "pam", CODEC_TYPE_VIDEO, CODEC_ID_PAM, sizeof(PNMContext), common_init, pam_encode_frame, NULL, //encode_end, pnm_decode_frame, .pix_fmts= (enum PixelFormat[]){PIX_FMT_RGB24, PIX_FMT_RGBA32, PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, -1}, }; #endif // CONFIG_PAM_ENCODER