diff options
Diffstat (limited to 'libavcodec/huffyuvenc.c')
-rw-r--r-- | libavcodec/huffyuvenc.c | 147 |
1 files changed, 107 insertions, 40 deletions
diff --git a/libavcodec/huffyuvenc.c b/libavcodec/huffyuvenc.c index 13c0a79566..95dcb88061 100644 --- a/libavcodec/huffyuvenc.c +++ b/libavcodec/huffyuvenc.c @@ -4,20 +4,20 @@ * see http://www.pcisys.net/~melanson/codecs/huffyuv.txt for a description of * the algorithm used * - * This file is part of Libav. + * This file is part of FFmpeg. * - * Libav is free software; you can redistribute it and/or + * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * - * Libav is distributed in the hope that it will be useful, + * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public - * License along with Libav; if not, write to the Free Software + * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ @@ -29,10 +29,11 @@ #include "avcodec.h" #include "huffyuv.h" #include "huffman.h" +#include "internal.h" #include "put_bits.h" static inline int sub_left_prediction(HYuvContext *s, uint8_t *dst, - uint8_t *src, int w, int left) + const uint8_t *src, int w, int left) { int i; if (w < 32) { @@ -54,25 +55,28 @@ static inline int sub_left_prediction(HYuvContext *s, uint8_t *dst, } static inline void sub_left_prediction_bgr32(HYuvContext *s, uint8_t *dst, - uint8_t *src, int w, - int *red, int *green, int *blue) + const uint8_t *src, int w, + int *red, int *green, int *blue, int *alpha) { int i; - int r,g,b; + int r,g,b,a; r = *red; g = *green; b = *blue; - + a = *alpha; for (i = 0; i < FFMIN(w, 4); i++) { const int rt = src[i * 4 + R]; const int gt = src[i * 4 + G]; const int bt = src[i * 4 + B]; + const int at = src[i * 4 + A]; dst[i * 4 + R] = rt - r; dst[i * 4 + G] = gt - g; dst[i * 4 + B] = bt - b; + dst[i * 4 + A] = at - a; r = rt; g = gt; b = bt; + a = at; } s->dsp.diff_bytes(dst + 16, src + 16, src + 12, w * 4 - 16); @@ -80,6 +84,32 @@ static inline void sub_left_prediction_bgr32(HYuvContext *s, uint8_t *dst, *red = src[(w - 1) * 4 + R]; *green = src[(w - 1) * 4 + G]; *blue = src[(w - 1) * 4 + B]; + *alpha = src[(w - 1) * 4 + A]; +} + +static inline void sub_left_prediction_rgb24(HYuvContext *s, uint8_t *dst, const uint8_t *src, int w, int *red, int *green, int *blue){ + int i; + int r,g,b; + r = *red; + g = *green; + b = *blue; + for (i = 0; i < FFMIN(w,16); i++) { + const int rt = src[i*3 + 0]; + const int gt = src[i*3 + 1]; + const int bt = src[i*3 + 2]; + dst[i*3 + 0] = rt - r; + dst[i*3 + 1] = gt - g; + dst[i*3 + 2] = bt - b; + r = rt; + g = gt; + b = bt; + } + + s->dsp.diff_bytes(dst + 48, src + 48, src + 48 - 3, w*3 - 48); + + *red = src[(w - 1)*3 + 0]; + *green = src[(w - 1)*3 + 1]; + *blue = src[(w - 1)*3 + 2]; } static int store_table(HYuvContext *s, const uint8_t *len, uint8_t *buf) @@ -94,8 +124,8 @@ static int store_table(HYuvContext *s, const uint8_t *len, uint8_t *buf) for (; i < 256 && len[i] == val && repeat < 255; i++) repeat++; - assert(val < 32 && val >0 && repeat<256 && repeat>0); - if ( repeat > 7) { + av_assert0(val < 32 && val >0 && repeat<256 && repeat>0); + if (repeat > 7) { buf[index++] = val; buf[index++] = repeat; } else { @@ -115,23 +145,32 @@ static av_cold int encode_init(AVCodecContext *avctx) avctx->extradata = av_mallocz(1024*30); // 256*3+4 == 772 avctx->stats_out = av_mallocz(1024*30); // 21*256*3(%llu ) + 3(\n) + 1(0) = 16132 + if (!avctx->extradata || !avctx->stats_out) { + av_freep(&avctx->stats_out); + return AVERROR(ENOMEM); + } s->version = 2; avctx->coded_frame = &s->picture; switch (avctx->pix_fmt) { case AV_PIX_FMT_YUV420P: - s->bitstream_bpp = 12; - break; case AV_PIX_FMT_YUV422P: - s->bitstream_bpp = 16; + if (s->width & 1) { + av_log(avctx, AV_LOG_ERROR, "width must be even for this colorspace\n"); + return AVERROR(EINVAL); + } + s->bitstream_bpp = avctx->pix_fmt == AV_PIX_FMT_YUV420P ? 12 : 16; break; case AV_PIX_FMT_RGB32: + s->bitstream_bpp = 32; + break; + case AV_PIX_FMT_RGB24: s->bitstream_bpp = 24; break; default: av_log(avctx, AV_LOG_ERROR, "format not supported\n"); - return -1; + return AVERROR(EINVAL); } avctx->bits_per_coded_sample = s->bitstream_bpp; s->decorrelate = s->bitstream_bpp >= 24; @@ -143,7 +182,7 @@ static av_cold int encode_init(AVCodecContext *avctx) av_log(avctx, AV_LOG_ERROR, "context=1 is not compatible with " "2 pass huffyuv encoding\n"); - return -1; + return AVERROR(EINVAL); } }else s->context= 0; @@ -152,13 +191,13 @@ static av_cold int encode_init(AVCodecContext *avctx) av_log(avctx, AV_LOG_ERROR, "Error: YV12 is not supported by huffyuv; use " "vcodec=ffvhuff or format=422p\n"); - return -1; + return AVERROR(EINVAL); } if (avctx->context_model) { av_log(avctx, AV_LOG_ERROR, "Error: per-frame huffman tables are not supported " "by huffyuv; use vcodec=ffvhuff\n"); - return -1; + return AVERROR(EINVAL); } if (s->interlaced != ( s->height > 288 )) av_log(avctx, AV_LOG_INFO, @@ -168,7 +207,7 @@ static av_cold int encode_init(AVCodecContext *avctx) if (s->bitstream_bpp >= 24 && s->predictor == MEDIAN) { av_log(avctx, AV_LOG_ERROR, "Error: RGB is incompatible with median predictor\n"); - return -1; + return AVERROR(EINVAL); } ((uint8_t*)avctx->extradata)[0] = s->predictor | (s->decorrelate << 6); @@ -232,7 +271,10 @@ static av_cold int encode_init(AVCodecContext *avctx) s->stats[i][j]= 0; } - ff_huffyuv_alloc_temp(s); + if (ff_huffyuv_alloc_temp(s)) { + ff_huffyuv_common_end(s); + return AVERROR(ENOMEM); + } s->picture_number=0; @@ -338,27 +380,30 @@ static int encode_gray_bitstream(HYuvContext *s, int count) return 0; } -static int encode_bgr_bitstream(HYuvContext *s, int count) +static inline int encode_bgra_bitstream(HYuvContext *s, int count, int planes) { int i; - if (s->pb.buf_end - s->pb.buf - (put_bits_count(&s->pb) >> 3) < 3 * 4 * count) { + if (s->pb.buf_end - s->pb.buf - (put_bits_count(&s->pb)>>3) < 4*planes*count) { av_log(s->avctx, AV_LOG_ERROR, "encoded frame too large\n"); return -1; } #define LOAD3\ - int g = s->temp[0][4 * i + G];\ - int b = (s->temp[0][4 * i + B] - g) & 0xff;\ - int r = (s->temp[0][4 * i + R] - g) & 0xff; + int g = s->temp[0][planes==3 ? 3*i + 1 : 4*i + G];\ + int b = (s->temp[0][planes==3 ? 3*i + 2 : 4*i + B] - g) & 0xff;\ + int r = (s->temp[0][planes==3 ? 3*i + 0 : 4*i + R] - g) & 0xff;\ + int a = s->temp[0][planes*i + A]; #define STAT3\ s->stats[0][b]++;\ s->stats[1][g]++;\ - s->stats[2][r]++; + s->stats[2][r]++;\ + if(planes==4) s->stats[2][a]++; #define WRITE3\ put_bits(&s->pb, s->len[1][g], s->bits[1][g]);\ put_bits(&s->pb, s->len[0][b], s->bits[0][b]);\ - put_bits(&s->pb, s->len[2][r], s->bits[2][r]); + put_bits(&s->pb, s->len[2][r], s->bits[2][r]);\ + if(planes==4) put_bits(&s->pb, s->len[2][a], s->bits[2][a]); if ((s->flags & CODEC_FLAG_PASS1) && (s->avctx->flags2 & CODEC_FLAG2_NO_OUTPUT)) { @@ -394,11 +439,8 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, AVFrame * const p = &s->picture; int i, j, size = 0, ret; - if (!pkt->data && - (ret = av_new_packet(pkt, width * height * 3 * 4 + FF_MIN_BUFFER_SIZE)) < 0) { - av_log(avctx, AV_LOG_ERROR, "Error allocating output packet.\n"); + if ((ret = ff_alloc_packet2(avctx, pkt, width * height * 3 * 4 + FF_MIN_BUFFER_SIZE)) < 0) return ret; - } *p = *pict; p->pict_type = AV_PICTURE_TYPE_I; @@ -529,25 +571,50 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, const int stride = -p->linesize[0]; const int fake_stride = -fake_ystride; int y; - int leftr, leftg, leftb; + int leftr, leftg, leftb, lefta; + put_bits(&s->pb, 8, lefta = data[A]); put_bits(&s->pb, 8, leftr = data[R]); put_bits(&s->pb, 8, leftg = data[G]); put_bits(&s->pb, 8, leftb = data[B]); - put_bits(&s->pb, 8, 0); - sub_left_prediction_bgr32(s, s->temp[0], data + 4, width - 1, &leftr, &leftg, &leftb); - encode_bgr_bitstream(s, width - 1); + sub_left_prediction_bgr32(s, s->temp[0], data + 4, width - 1, &leftr, &leftg, &leftb, &lefta); + encode_bgra_bitstream(s, width - 1, 4); for (y = 1; y < s->height; y++) { uint8_t *dst = data + y*stride; if (s->predictor == PLANE && s->interlaced < y) { s->dsp.diff_bytes(s->temp[1], dst, dst - fake_stride, width * 4); - sub_left_prediction_bgr32(s, s->temp[0], s->temp[1], width, &leftr, &leftg, &leftb); + sub_left_prediction_bgr32(s, s->temp[0], s->temp[1], width, &leftr, &leftg, &leftb, &lefta); } else { - sub_left_prediction_bgr32(s, s->temp[0], dst, width, &leftr, &leftg, &leftb); + sub_left_prediction_bgr32(s, s->temp[0], dst, width, &leftr, &leftg, &leftb, &lefta); + } + encode_bgra_bitstream(s, width, 4); + } + }else if(avctx->pix_fmt == AV_PIX_FMT_RGB24){ + uint8_t *data = p->data[0] + (height-1)*p->linesize[0]; + const int stride = -p->linesize[0]; + const int fake_stride = -fake_ystride; + int y; + int leftr, leftg, leftb; + + put_bits(&s->pb, 8, leftr= data[0]); + put_bits(&s->pb, 8, leftg= data[1]); + put_bits(&s->pb, 8, leftb= data[2]); + put_bits(&s->pb, 8, 0); + + sub_left_prediction_rgb24(s, s->temp[0], data+3, width-1, &leftr, &leftg, &leftb); + encode_bgra_bitstream(s, width-1, 3); + + for(y=1; y<s->height; y++){ + uint8_t *dst = data + y*stride; + if(s->predictor == PLANE && s->interlaced < y){ + s->dsp.diff_bytes(s->temp[1], dst, dst - fake_stride, width*3); + sub_left_prediction_rgb24(s, s->temp[0], s->temp[1], width, &leftr, &leftg, &leftb); + }else{ + sub_left_prediction_rgb24(s, s->temp[0], dst, width, &leftr, &leftg, &leftb); } - encode_bgr_bitstream(s, width); + encode_bgra_bitstream(s, width, 3); } } else { av_log(avctx, AV_LOG_ERROR, "Format not supported!\n"); @@ -610,7 +677,7 @@ AVCodec ff_huffyuv_encoder = { .encode2 = encode_frame, .close = encode_end, .pix_fmts = (const enum AVPixelFormat[]){ - AV_PIX_FMT_YUV422P, AV_PIX_FMT_RGB32, AV_PIX_FMT_NONE + AV_PIX_FMT_YUV422P, AV_PIX_FMT_RGB24, AV_PIX_FMT_RGB32, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Huffyuv / HuffYUV"), }; @@ -626,7 +693,7 @@ AVCodec ff_ffvhuff_encoder = { .encode2 = encode_frame, .close = encode_end, .pix_fmts = (const enum AVPixelFormat[]){ - AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_RGB32, AV_PIX_FMT_NONE + AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_RGB24, AV_PIX_FMT_RGB32, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Huffyuv FFmpeg variant"), }; |