diff options
Diffstat (limited to 'libavcodec')
201 files changed, 1160 insertions, 1159 deletions
diff --git a/libavcodec/4xm.c b/libavcodec/4xm.c index 2b59c6b19d..38fc2f2e64 100644 --- a/libavcodec/4xm.c +++ b/libavcodec/4xm.c @@ -900,9 +900,9 @@ static av_cold int decode_init(AVCodecContext *avctx) init_vlcs(f); if (f->version > 2) - avctx->pix_fmt = PIX_FMT_RGB565; + avctx->pix_fmt = AV_PIX_FMT_RGB565; else - avctx->pix_fmt = PIX_FMT_BGR555; + avctx->pix_fmt = AV_PIX_FMT_BGR555; return 0; } diff --git a/libavcodec/8bps.c b/libavcodec/8bps.c index 12e9358ad1..95e811a43c 100644 --- a/libavcodec/8bps.c +++ b/libavcodec/8bps.c @@ -40,8 +40,8 @@ #include "avcodec.h" -static const enum PixelFormat pixfmt_rgb24[] = { - PIX_FMT_BGR24, PIX_FMT_RGB32, PIX_FMT_NONE }; +static const enum AVPixelFormat pixfmt_rgb24[] = { + AV_PIX_FMT_BGR24, AV_PIX_FMT_RGB32, AV_PIX_FMT_NONE }; /* * Decoder context @@ -95,7 +95,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, if (planes == 4) planes--; - px_inc = planes + (avctx->pix_fmt == PIX_FMT_RGB32); + px_inc = planes + (avctx->pix_fmt == AV_PIX_FMT_RGB32); for (p = 0; p < planes; p++) { /* Lines length pointer for this plane */ @@ -170,7 +170,7 @@ static av_cold int decode_init(AVCodecContext *avctx) switch (avctx->bits_per_coded_sample) { case 8: - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; c->planes = 1; c->planemap[0] = 0; // 1st plane is palette indexes break; @@ -182,7 +182,7 @@ static av_cold int decode_init(AVCodecContext *avctx) c->planemap[2] = 0; // 3rd plane is blue break; case 32: - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; c->planes = 4; #if HAVE_BIGENDIAN c->planemap[0] = 1; // 1st plane is red diff --git a/libavcodec/a64multienc.c b/libavcodec/a64multienc.c index c9d3b3c44b..beddf9f9c3 100644 --- a/libavcodec/a64multienc.c +++ b/libavcodec/a64multienc.c @@ -379,7 +379,7 @@ AVCodec ff_a64multi_encoder = { .init = a64multi_init_encoder, .encode2 = a64multi_encode_frame, .close = a64multi_close_encoder, - .pix_fmts = (const enum PixelFormat[]) {PIX_FMT_GRAY8, PIX_FMT_NONE}, + .pix_fmts = (const enum AVPixelFormat[]) {AV_PIX_FMT_GRAY8, AV_PIX_FMT_NONE}, .long_name = NULL_IF_CONFIG_SMALL("Multicolor charset for Commodore 64"), .capabilities = CODEC_CAP_DELAY, }; @@ -392,7 +392,7 @@ AVCodec ff_a64multi5_encoder = { .init = a64multi_init_encoder, .encode2 = a64multi_encode_frame, .close = a64multi_close_encoder, - .pix_fmts = (const enum PixelFormat[]) {PIX_FMT_GRAY8, PIX_FMT_NONE}, + .pix_fmts = (const enum AVPixelFormat[]) {AV_PIX_FMT_GRAY8, AV_PIX_FMT_NONE}, .long_name = NULL_IF_CONFIG_SMALL("Multicolor charset for Commodore 64, extended with 5th color (colram)"), .capabilities = CODEC_CAP_DELAY, }; diff --git a/libavcodec/aasc.c b/libavcodec/aasc.c index f27f44dd51..60bf7c626d 100644 --- a/libavcodec/aasc.c +++ b/libavcodec/aasc.c @@ -44,7 +44,7 @@ static av_cold int aasc_decode_init(AVCodecContext *avctx) s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; return 0; } diff --git a/libavcodec/anm.c b/libavcodec/anm.c index 4b6a5564d0..de563dda2b 100644 --- a/libavcodec/anm.c +++ b/libavcodec/anm.c @@ -39,7 +39,7 @@ static av_cold int decode_init(AVCodecContext *avctx) AnmContext *s = avctx->priv_data; int i; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; s->frame.reference = 1; bytestream2_init(&s->gb, avctx->extradata, avctx->extradata_size); diff --git a/libavcodec/ansi.c b/libavcodec/ansi.c index 04a1106605..964abc6d20 100644 --- a/libavcodec/ansi.c +++ b/libavcodec/ansi.c @@ -74,7 +74,7 @@ typedef struct { static av_cold int decode_init(AVCodecContext *avctx) { AnsiContext *s = avctx->priv_data; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; /* defaults */ s->font = ff_vga16_font; diff --git a/libavcodec/api-example.c b/libavcodec/api-example.c index 62750804ba..198a59d5f5 100644 --- a/libavcodec/api-example.c +++ b/libavcodec/api-example.c @@ -345,7 +345,7 @@ static void video_encode_example(const char *filename) c->time_base= (AVRational){1,25}; c->gop_size = 10; /* emit one intra frame every ten frames */ c->max_b_frames=1; - c->pix_fmt = PIX_FMT_YUV420P; + c->pix_fmt = AV_PIX_FMT_YUV420P; /* open it */ if (avcodec_open2(c, codec, NULL) < 0) { diff --git a/libavcodec/asv1.c b/libavcodec/asv1.c index a3106b2a54..d10513dc46 100644 --- a/libavcodec/asv1.c +++ b/libavcodec/asv1.c @@ -553,7 +553,7 @@ static av_cold int decode_init(AVCodecContext *avctx){ common_init(avctx); init_vlcs(a); ff_init_scantable(a->dsp.idct_permutation, &a->scantable, scantab); - avctx->pix_fmt= PIX_FMT_YUV420P; + avctx->pix_fmt= AV_PIX_FMT_YUV420P; a->inv_qscale= avctx->extradata[0]; if(a->inv_qscale == 0){ @@ -649,7 +649,7 @@ AVCodec ff_asv1_encoder = { .priv_data_size = sizeof(ASV1Context), .init = encode_init, .encode2 = encode_frame, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("ASUS V1"), }; #endif @@ -662,7 +662,7 @@ AVCodec ff_asv2_encoder = { .priv_data_size = sizeof(ASV1Context), .init = encode_init, .encode2 = encode_frame, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("ASUS V2"), }; #endif diff --git a/libavcodec/aura.c b/libavcodec/aura.c index 68e0c7a8a9..9f910212fb 100644 --- a/libavcodec/aura.c +++ b/libavcodec/aura.c @@ -39,7 +39,7 @@ static av_cold int aura_decode_init(AVCodecContext *avctx) /* width needs to be divisible by 4 for this codec to work */ if (avctx->width & 0x3) return -1; - avctx->pix_fmt = PIX_FMT_YUV422P; + avctx->pix_fmt = AV_PIX_FMT_YUV422P; return 0; } diff --git a/libavcodec/avcodec.h b/libavcodec/avcodec.h index 8e37b8373e..223db9ca75 100644 --- a/libavcodec/avcodec.h +++ b/libavcodec/avcodec.h @@ -1011,7 +1011,7 @@ typedef struct AVFrame { /** * format of the frame, -1 if unknown or unset - * Values correspond to enum PixelFormat for video frames, + * Values correspond to enum AVPixelFormat for video frames, * enum AVSampleFormat for audio) * - encoding: unused * - decoding: Read by user. @@ -1492,13 +1492,13 @@ typedef struct AVCodecContext { int gop_size; /** - * Pixel format, see PIX_FMT_xxx. + * Pixel format, see AV_PIX_FMT_xxx. * May be set by the demuxer if known from headers. * May be overriden by the decoder if it knows better. * - encoding: Set by user. * - decoding: Set by user if known, overridden by libavcodec if known */ - enum PixelFormat pix_fmt; + enum AVPixelFormat pix_fmt; /** * Motion estimation algorithm used for video coding. @@ -1545,7 +1545,7 @@ typedef struct AVCodecContext { * - encoding: unused * - decoding: Set by user, if not set the native format will be chosen. */ - enum PixelFormat (*get_format)(struct AVCodecContext *s, const enum PixelFormat * fmt); + enum AVPixelFormat (*get_format)(struct AVCodecContext *s, const enum AVPixelFormat * fmt); /** * maximum number of B-frames between non-B-frames @@ -2931,7 +2931,7 @@ typedef struct AVCodec { */ int capabilities; const AVRational *supported_framerates; ///< array of supported framerates, or NULL if any, array is terminated by {0,0} - const enum PixelFormat *pix_fmts; ///< array of supported pixel formats, or NULL if unknown, array is terminated by -1 + const enum AVPixelFormat *pix_fmts; ///< array of supported pixel formats, or NULL if unknown, array is terminated by -1 const int *supported_samplerates; ///< array of supported audio samplerates, or NULL if unknown, array is terminated by 0 const enum AVSampleFormat *sample_fmts; ///< array of supported sample formats, or NULL if unknown, array is terminated by -1 const uint64_t *channel_layouts; ///< array of support channel layouts, or NULL if unknown. array is terminated by 0 @@ -3032,7 +3032,7 @@ typedef struct AVHWAccel { * * Only hardware accelerated formats are supported here. */ - enum PixelFormat pix_fmt; + enum AVPixelFormat pix_fmt; /** * Hardware accelerated codec capabilities. @@ -4135,7 +4135,7 @@ void av_resample_close(struct AVResampleContext *c); * @param height the height of the picture * @return zero if successful, a negative value if not */ -int avpicture_alloc(AVPicture *picture, enum PixelFormat pix_fmt, int width, int height); +int avpicture_alloc(AVPicture *picture, enum AVPixelFormat pix_fmt, int width, int height); /** * Free a picture previously allocated by avpicture_alloc(). @@ -4167,7 +4167,7 @@ void avpicture_free(AVPicture *picture); * @return size of the image data in bytes */ int avpicture_fill(AVPicture *picture, uint8_t *ptr, - enum PixelFormat pix_fmt, int width, int height); + enum AVPixelFormat pix_fmt, int width, int height); /** * Copy pixel data from an AVPicture into a buffer. @@ -4184,7 +4184,8 @@ int avpicture_fill(AVPicture *picture, uint8_t *ptr, * @param[in] dest_size The size of 'dest'. * @return The number of bytes written to dest, or a negative value (error code) on error. */ -int avpicture_layout(const AVPicture* src, enum PixelFormat pix_fmt, int width, int height, +int avpicture_layout(const AVPicture* src, enum AVPixelFormat pix_fmt, + int width, int height, unsigned char *dest, int dest_size); /** @@ -4199,29 +4200,29 @@ int avpicture_layout(const AVPicture* src, enum PixelFormat pix_fmt, int width, * @param height the height of the image * @return Image data size in bytes or -1 on error (e.g. too large dimensions). */ -int avpicture_get_size(enum PixelFormat pix_fmt, int width, int height); +int avpicture_get_size(enum AVPixelFormat pix_fmt, int width, int height); /** * deinterlace - if not supported return -1 */ int avpicture_deinterlace(AVPicture *dst, const AVPicture *src, - enum PixelFormat pix_fmt, int width, int height); + enum AVPixelFormat pix_fmt, int width, int height); /** * Copy image src to dst. Wraps av_picture_data_copy() above. */ void av_picture_copy(AVPicture *dst, const AVPicture *src, - enum PixelFormat pix_fmt, int width, int height); + enum AVPixelFormat pix_fmt, int width, int height); /** * Crop image top and left side. */ int av_picture_crop(AVPicture *dst, const AVPicture *src, - enum PixelFormat pix_fmt, int top_band, int left_band); + enum AVPixelFormat pix_fmt, int top_band, int left_band); /** * Pad image. */ -int av_picture_pad(AVPicture *dst, const AVPicture *src, int height, int width, enum PixelFormat pix_fmt, +int av_picture_pad(AVPicture *dst, const AVPicture *src, int height, int width, enum AVPixelFormat pix_fmt, int padtop, int padbottom, int padleft, int padright, int *color); /** @@ -4244,14 +4245,14 @@ int av_picture_pad(AVPicture *dst, const AVPicture *src, int height, int width, * @{ */ -void avcodec_get_chroma_sub_sample(enum PixelFormat pix_fmt, int *h_shift, int *v_shift); +void avcodec_get_chroma_sub_sample(enum AVPixelFormat pix_fmt, int *h_shift, int *v_shift); /** * Return a value representing the fourCC code associated to the * pixel format pix_fmt, or 0 if no associated fourCC code can be * found. */ -unsigned int avcodec_pix_fmt_to_codec_tag(enum PixelFormat pix_fmt); +unsigned int avcodec_pix_fmt_to_codec_tag(enum AVPixelFormat pix_fmt); #define FF_LOSS_RESOLUTION 0x0001 /**< loss due to resolution change */ #define FF_LOSS_DEPTH 0x0002 /**< loss due to color depth change */ @@ -4277,7 +4278,7 @@ unsigned int avcodec_pix_fmt_to_codec_tag(enum PixelFormat pix_fmt); * @param[in] has_alpha Whether the source pixel format alpha channel is used. * @return Combination of flags informing you what kind of losses will occur. */ -int avcodec_get_pix_fmt_loss(enum PixelFormat dst_pix_fmt, enum PixelFormat src_pix_fmt, +int avcodec_get_pix_fmt_loss(enum AVPixelFormat dst_pix_fmt, enum AVPixelFormat src_pix_fmt, int has_alpha); #if FF_API_FIND_BEST_PIX_FMT @@ -4294,8 +4295,8 @@ int avcodec_get_pix_fmt_loss(enum PixelFormat dst_pix_fmt, enum PixelFormat src_ * pix_fmt_mask parameter. * * @code - * src_pix_fmt = PIX_FMT_YUV420P; - * pix_fmt_mask = (1 << PIX_FMT_YUV422P) || (1 << PIX_FMT_RGB24); + * src_pix_fmt = AV_PIX_FMT_YUV420P; + * pix_fmt_mask = (1 << AV_PIX_FMT_YUV422P) || (1 << AV_PIX_FMT_RGB24); * dst_pix_fmt = avcodec_find_best_pix_fmt(pix_fmt_mask, src_pix_fmt, alpha, &loss); * @endcode * @@ -4306,7 +4307,7 @@ int avcodec_get_pix_fmt_loss(enum PixelFormat dst_pix_fmt, enum PixelFormat src_ * @return The best pixel format to convert to or -1 if none was found. */ attribute_deprecated -enum PixelFormat avcodec_find_best_pix_fmt(int64_t pix_fmt_mask, enum PixelFormat src_pix_fmt, +enum AVPixelFormat avcodec_find_best_pix_fmt(int64_t pix_fmt_mask, enum AVPixelFormat src_pix_fmt, int has_alpha, int *loss_ptr); #endif /* FF_API_FIND_BEST_PIX_FMT */ @@ -4321,17 +4322,17 @@ enum PixelFormat avcodec_find_best_pix_fmt(int64_t pix_fmt_mask, enum PixelForma * pix_fmt_list parameter. * * - * @param[in] pix_fmt_list PIX_FMT_NONE terminated array of pixel formats to choose from + * @param[in] pix_fmt_list AV_PIX_FMT_NONE terminated array of pixel formats to choose from * @param[in] src_pix_fmt source pixel format * @param[in] has_alpha Whether the source pixel format alpha channel is used. * @param[out] loss_ptr Combination of flags informing you what kind of losses will occur. * @return The best pixel format to convert to or -1 if none was found. */ -enum PixelFormat avcodec_find_best_pix_fmt2(enum PixelFormat *pix_fmt_list, - enum PixelFormat src_pix_fmt, - int has_alpha, int *loss_ptr); +enum AVPixelFormat avcodec_find_best_pix_fmt2(enum AVPixelFormat *pix_fmt_list, + enum AVPixelFormat src_pix_fmt, + int has_alpha, int *loss_ptr); -enum PixelFormat avcodec_default_get_format(struct AVCodecContext *s, const enum PixelFormat * fmt); +enum AVPixelFormat avcodec_default_get_format(struct AVCodecContext *s, const enum AVPixelFormat * fmt); /** * @} diff --git a/libavcodec/avs.c b/libavcodec/avs.c index b8ee1e1ff2..f6535a26a6 100644 --- a/libavcodec/avs.c +++ b/libavcodec/avs.c @@ -157,7 +157,7 @@ avs_decode_frame(AVCodecContext * avctx, static av_cold int avs_decode_init(AVCodecContext * avctx) { - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; avcodec_set_dimensions(avctx, 318, 198); return 0; } diff --git a/libavcodec/bethsoftvideo.c b/libavcodec/bethsoftvideo.c index 3900f7512c..c1f95ed889 100644 --- a/libavcodec/bethsoftvideo.c +++ b/libavcodec/bethsoftvideo.c @@ -43,7 +43,7 @@ static av_cold int bethsoftvid_decode_init(AVCodecContext *avctx) vid->frame.reference = 1; vid->frame.buffer_hints = FF_BUFFER_HINTS_VALID | FF_BUFFER_HINTS_PRESERVE | FF_BUFFER_HINTS_REUSABLE; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/bfi.c b/libavcodec/bfi.c index 8355850183..34c5bd3be5 100644 --- a/libavcodec/bfi.c +++ b/libavcodec/bfi.c @@ -39,7 +39,7 @@ typedef struct BFIContext { static av_cold int bfi_decode_init(AVCodecContext *avctx) { BFIContext *bfi = avctx->priv_data; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; bfi->dst = av_mallocz(avctx->width * avctx->height); return 0; } diff --git a/libavcodec/bink.c b/libavcodec/bink.c index b726bae45a..b6c8c4b240 100644 --- a/libavcodec/bink.c +++ b/libavcodec/bink.c @@ -1296,7 +1296,7 @@ static av_cold int decode_init(AVCodecContext *avctx) return 1; } - avctx->pix_fmt = c->has_alpha ? PIX_FMT_YUVA420P : PIX_FMT_YUV420P; + avctx->pix_fmt = c->has_alpha ? AV_PIX_FMT_YUVA420P : AV_PIX_FMT_YUV420P; avctx->idct_algo = FF_IDCT_BINK; ff_dsputil_init(&c->dsp, avctx); diff --git a/libavcodec/bmp.c b/libavcodec/bmp.c index 3af25cd333..b38e056315 100644 --- a/libavcodec/bmp.c +++ b/libavcodec/bmp.c @@ -136,7 +136,7 @@ static int bmp_decode_frame(AVCodecContext *avctx, avctx->width = width; avctx->height = height > 0? height: -height; - avctx->pix_fmt = PIX_FMT_NONE; + avctx->pix_fmt = AV_PIX_FMT_NONE; switch(depth){ case 32: @@ -155,21 +155,21 @@ static int bmp_decode_frame(AVCodecContext *avctx, rgb[2] = 0; } - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; break; case 24: - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; break; case 16: if(comp == BMP_RGB) - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; else if (comp == BMP_BITFIELDS) { if (rgb[0] == 0xF800 && rgb[1] == 0x07E0 && rgb[2] == 0x001F) - avctx->pix_fmt = PIX_FMT_RGB565; + avctx->pix_fmt = AV_PIX_FMT_RGB565; else if (rgb[0] == 0x7C00 && rgb[1] == 0x03E0 && rgb[2] == 0x001F) - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; else if (rgb[0] == 0x0F00 && rgb[1] == 0x00F0 && rgb[2] == 0x000F) - avctx->pix_fmt = PIX_FMT_RGB444; + avctx->pix_fmt = AV_PIX_FMT_RGB444; else { av_log(avctx, AV_LOG_ERROR, "Unknown bitfields %0X %0X %0X\n", rgb[0], rgb[1], rgb[2]); return AVERROR(EINVAL); @@ -178,14 +178,14 @@ static int bmp_decode_frame(AVCodecContext *avctx, break; case 8: if(hsize - ihsize - 14 > 0) - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; else - avctx->pix_fmt = PIX_FMT_GRAY8; + avctx->pix_fmt = AV_PIX_FMT_GRAY8; break; case 1: case 4: if(hsize - ihsize - 14 > 0){ - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; }else{ av_log(avctx, AV_LOG_ERROR, "Unknown palette for %d-colour BMP\n", 1<<depth); return -1; @@ -196,7 +196,7 @@ static int bmp_decode_frame(AVCodecContext *avctx, return -1; } - if(avctx->pix_fmt == PIX_FMT_NONE){ + if(avctx->pix_fmt == AV_PIX_FMT_NONE){ av_log(avctx, AV_LOG_ERROR, "unsupported pixel format\n"); return -1; } @@ -236,7 +236,7 @@ static int bmp_decode_frame(AVCodecContext *avctx, linesize = p->linesize[0]; } - if(avctx->pix_fmt == PIX_FMT_PAL8){ + if(avctx->pix_fmt == AV_PIX_FMT_PAL8){ int colors = 1 << depth; memset(p->data[1], 0, 1024); diff --git a/libavcodec/bmpenc.c b/libavcodec/bmpenc.c index 37cf6e13e4..a798f34409 100644 --- a/libavcodec/bmpenc.c +++ b/libavcodec/bmpenc.c @@ -37,23 +37,23 @@ static av_cold int bmp_encode_init(AVCodecContext *avctx){ avctx->coded_frame = &s->picture; switch (avctx->pix_fmt) { - case PIX_FMT_BGR24: + case AV_PIX_FMT_BGR24: avctx->bits_per_coded_sample = 24; break; - case PIX_FMT_RGB555: - case PIX_FMT_RGB565: - case PIX_FMT_RGB444: + case AV_PIX_FMT_RGB555: + case AV_PIX_FMT_RGB565: + case AV_PIX_FMT_RGB444: avctx->bits_per_coded_sample = 16; break; - case PIX_FMT_RGB8: - case PIX_FMT_BGR8: - case PIX_FMT_RGB4_BYTE: - case PIX_FMT_BGR4_BYTE: - case PIX_FMT_GRAY8: - case PIX_FMT_PAL8: + case AV_PIX_FMT_RGB8: + case AV_PIX_FMT_BGR8: + case AV_PIX_FMT_RGB4_BYTE: + case AV_PIX_FMT_BGR4_BYTE: + case AV_PIX_FMT_GRAY8: + case AV_PIX_FMT_PAL8: avctx->bits_per_coded_sample = 8; break; - case PIX_FMT_MONOBLACK: + case AV_PIX_FMT_MONOBLACK: avctx->bits_per_coded_sample = 1; break; default: @@ -78,26 +78,26 @@ static int bmp_encode_frame(AVCodecContext *avctx, AVPacket *pkt, p->pict_type= AV_PICTURE_TYPE_I; p->key_frame= 1; switch (avctx->pix_fmt) { - case PIX_FMT_RGB444: + case AV_PIX_FMT_RGB444: compression = BMP_BITFIELDS; pal = rgb444_masks; // abuse pal to hold color masks pal_entries = 3; break; - case PIX_FMT_RGB565: + case AV_PIX_FMT_RGB565: compression = BMP_BITFIELDS; pal = rgb565_masks; // abuse pal to hold color masks pal_entries = 3; break; - case PIX_FMT_RGB8: - case PIX_FMT_BGR8: - case PIX_FMT_RGB4_BYTE: - case PIX_FMT_BGR4_BYTE: - case PIX_FMT_GRAY8: + case AV_PIX_FMT_RGB8: + case AV_PIX_FMT_BGR8: + case AV_PIX_FMT_RGB4_BYTE: + case AV_PIX_FMT_BGR4_BYTE: + case AV_PIX_FMT_GRAY8: ff_set_systematic_pal2((uint32_t*)p->data[1], avctx->pix_fmt); - case PIX_FMT_PAL8: + case AV_PIX_FMT_PAL8: pal = (uint32_t *)p->data[1]; break; - case PIX_FMT_MONOBLACK: + case AV_PIX_FMT_MONOBLACK: pal = monoblack_pal; break; } @@ -166,12 +166,12 @@ AVCodec ff_bmp_encoder = { .priv_data_size = sizeof(BMPContext), .init = bmp_encode_init, .encode2 = bmp_encode_frame, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_BGR24, - PIX_FMT_RGB555, PIX_FMT_RGB444, PIX_FMT_RGB565, - PIX_FMT_RGB8, PIX_FMT_BGR8, PIX_FMT_RGB4_BYTE, PIX_FMT_BGR4_BYTE, PIX_FMT_GRAY8, PIX_FMT_PAL8, - PIX_FMT_MONOBLACK, - PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_BGR24, + AV_PIX_FMT_RGB555, AV_PIX_FMT_RGB444, AV_PIX_FMT_RGB565, + AV_PIX_FMT_RGB8, AV_PIX_FMT_BGR8, AV_PIX_FMT_RGB4_BYTE, AV_PIX_FMT_BGR4_BYTE, AV_PIX_FMT_GRAY8, AV_PIX_FMT_PAL8, + AV_PIX_FMT_MONOBLACK, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("BMP (Windows and OS/2 bitmap)"), }; diff --git a/libavcodec/bmv.c b/libavcodec/bmv.c index 11f3c4fc58..876c13f923 100644 --- a/libavcodec/bmv.c +++ b/libavcodec/bmv.c @@ -266,7 +266,7 @@ static av_cold int decode_init(AVCodecContext *avctx) BMVDecContext * const c = avctx->priv_data; c->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; c->pic.reference = 1; if (avctx->get_buffer(avctx, &c->pic) < 0) { diff --git a/libavcodec/c93.c b/libavcodec/c93.c index b0bffbce83..07acb26a72 100644 --- a/libavcodec/c93.c +++ b/libavcodec/c93.c @@ -47,7 +47,7 @@ typedef enum { static av_cold int decode_init(AVCodecContext *avctx) { - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/cavs.c b/libavcodec/cavs.c index 02d572ee11..a68a62801b 100644 --- a/libavcodec/cavs.c +++ b/libavcodec/cavs.c @@ -712,7 +712,7 @@ av_cold int ff_cavs_init(AVCodecContext *avctx) { ff_cavsdsp_init(&h->cdsp, avctx); s->avctx = avctx; - avctx->pix_fmt= PIX_FMT_YUV420P; + avctx->pix_fmt= AV_PIX_FMT_YUV420P; h->luma_scan[0] = 0; h->luma_scan[1] = 8; diff --git a/libavcodec/cdgraphics.c b/libavcodec/cdgraphics.c index bfbb736a2f..21de07effb 100644 --- a/libavcodec/cdgraphics.c +++ b/libavcodec/cdgraphics.c @@ -86,7 +86,7 @@ static av_cold int cdg_decode_init(AVCodecContext *avctx) avctx->width = CDG_FULL_WIDTH; avctx->height = CDG_FULL_HEIGHT; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/cdxl.c b/libavcodec/cdxl.c index 86b7515df9..bc4dfe74fc 100644 --- a/libavcodec/cdxl.c +++ b/libavcodec/cdxl.c @@ -247,11 +247,11 @@ static int cdxl_decode_frame(AVCodecContext *avctx, void *data, if (c->video_size < aligned_width * avctx->height * c->bpp / 8) return AVERROR_INVALIDDATA; if (!encoding && c->palette_size && c->bpp <= 8) { - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; } else if (encoding == 1 && (c->bpp == 6 || c->bpp == 8)) { if (c->palette_size != (1 << (c->bpp - 1))) return AVERROR_INVALIDDATA; - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; } else { av_log_ask_for_sample(avctx, "unsupported encoding %d and bpp %d\n", encoding, c->bpp); diff --git a/libavcodec/cinepak.c b/libavcodec/cinepak.c index 2b069925cb..e6e2df2a27 100644 --- a/libavcodec/cinepak.c +++ b/libavcodec/cinepak.c @@ -407,10 +407,10 @@ static av_cold int cinepak_decode_init(AVCodecContext *avctx) // check for paletted data if (avctx->bits_per_coded_sample != 8) { s->palette_video = 0; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; } else { s->palette_video = 1; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; } s->frame.data[0] = NULL; diff --git a/libavcodec/cljr.c b/libavcodec/cljr.c index 3c75d8329f..13b454727a 100644 --- a/libavcodec/cljr.c +++ b/libavcodec/cljr.c @@ -102,7 +102,7 @@ static int decode_frame(AVCodecContext *avctx, static av_cold int decode_init(AVCodecContext *avctx) { - avctx->pix_fmt = PIX_FMT_YUV411P; + avctx->pix_fmt = AV_PIX_FMT_YUV411P; return common_init(avctx); } @@ -175,8 +175,8 @@ AVCodec ff_cljr_encoder = { .priv_data_size = sizeof(CLJRContext), .init = common_init, .encode2 = encode_frame, - .pix_fmts = (const enum PixelFormat[]) { PIX_FMT_YUV411P, - PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_YUV411P, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Cirrus Logic AccuPak"), }; #endif diff --git a/libavcodec/cllc.c b/libavcodec/cllc.c index ebc466b4a5..467d81953e 100644 --- a/libavcodec/cllc.c +++ b/libavcodec/cllc.c @@ -330,7 +330,7 @@ static int cllc_decode_frame(AVCodecContext *avctx, void *data, switch (coding_type) { case 1: case 2: - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; avctx->bits_per_raw_sample = 8; ret = avctx->get_buffer(avctx, pic); @@ -345,7 +345,7 @@ static int cllc_decode_frame(AVCodecContext *avctx, void *data, break; case 3: - avctx->pix_fmt = PIX_FMT_ARGB; + avctx->pix_fmt = AV_PIX_FMT_ARGB; avctx->bits_per_raw_sample = 8; ret = avctx->get_buffer(avctx, pic); diff --git a/libavcodec/cscd.c b/libavcodec/cscd.c index 1e210efb78..8d0cbb66c8 100644 --- a/libavcodec/cscd.c +++ b/libavcodec/cscd.c @@ -222,9 +222,9 @@ static av_cold int decode_init(AVCodecContext *avctx) { CamStudioContext *c = avctx->priv_data; int stride; switch (avctx->bits_per_coded_sample) { - case 16: avctx->pix_fmt = PIX_FMT_RGB555; break; - case 24: avctx->pix_fmt = PIX_FMT_BGR24; break; - case 32: avctx->pix_fmt = PIX_FMT_RGB32; break; + case 16: avctx->pix_fmt = AV_PIX_FMT_RGB555; break; + case 24: avctx->pix_fmt = AV_PIX_FMT_BGR24; break; + case 32: avctx->pix_fmt = AV_PIX_FMT_RGB32; break; default: av_log(avctx, AV_LOG_ERROR, "CamStudio codec error: invalid depth %i bpp\n", diff --git a/libavcodec/cyuv.c b/libavcodec/cyuv.c index 1cf7013e8e..18767d78e8 100644 --- a/libavcodec/cyuv.c +++ b/libavcodec/cyuv.c @@ -53,7 +53,7 @@ static av_cold int cyuv_decode_init(AVCodecContext *avctx) if (s->width & 0x3) return -1; s->height = avctx->height; - avctx->pix_fmt = PIX_FMT_YUV411P; + avctx->pix_fmt = AV_PIX_FMT_YUV411P; return 0; } diff --git a/libavcodec/dfa.c b/libavcodec/dfa.c index d464acb187..39f0f6465d 100644 --- a/libavcodec/dfa.c +++ b/libavcodec/dfa.c @@ -38,7 +38,7 @@ static av_cold int dfa_decode_init(AVCodecContext *avctx) DfaContext *s = avctx->priv_data; int ret; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; if ((ret = av_image_check_size(avctx->width, avctx->height, 0, avctx)) < 0) return ret; diff --git a/libavcodec/dirac.c b/libavcodec/dirac.c index 489b1f5a01..070ea0f522 100644 --- a/libavcodec/dirac.c +++ b/libavcodec/dirac.c @@ -107,9 +107,9 @@ static const struct { }; /* [DIRAC_STD] Table 10.2 Supported chroma sampling formats + luma Offset */ -static const enum PixelFormat dirac_pix_fmt[2][3] = { - { PIX_FMT_YUV444P, PIX_FMT_YUV422P, PIX_FMT_YUV420P }, - { PIX_FMT_YUVJ444P, PIX_FMT_YUVJ422P, PIX_FMT_YUVJ420P }, +static const enum AVPixelFormat dirac_pix_fmt[2][3] = { + { AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P }, + { AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P }, }; /* [DIRAC_STD] 10.3 Parse Source Parameters. diff --git a/libavcodec/dnxhddec.c b/libavcodec/dnxhddec.c index 523030510c..e71409b33f 100644 --- a/libavcodec/dnxhddec.c +++ b/libavcodec/dnxhddec.c @@ -124,7 +124,7 @@ static int dnxhd_decode_header(DNXHDContext *ctx, const uint8_t *buf, int buf_si av_dlog(ctx->avctx, "width %d, height %d\n", ctx->width, ctx->height); if (buf[0x21] & 0x40) { - ctx->avctx->pix_fmt = PIX_FMT_YUV422P10; + ctx->avctx->pix_fmt = AV_PIX_FMT_YUV422P10; ctx->avctx->bits_per_raw_sample = 10; if (ctx->bit_depth != 10) { ff_dsputil_init(&ctx->dsp, ctx->avctx); @@ -132,7 +132,7 @@ static int dnxhd_decode_header(DNXHDContext *ctx, const uint8_t *buf, int buf_si ctx->decode_dct_block = dnxhd_decode_dct_block_10; } } else { - ctx->avctx->pix_fmt = PIX_FMT_YUV422P; + ctx->avctx->pix_fmt = AV_PIX_FMT_YUV422P; ctx->avctx->bits_per_raw_sample = 8; if (ctx->bit_depth != 8) { ff_dsputil_init(&ctx->dsp, ctx->avctx); diff --git a/libavcodec/dnxhdenc.c b/libavcodec/dnxhdenc.c index 203b8b5189..8531fe0ea5 100644 --- a/libavcodec/dnxhdenc.c +++ b/libavcodec/dnxhdenc.c @@ -233,10 +233,10 @@ static int dnxhd_encode_init(AVCodecContext *avctx) int i, index, bit_depth; switch (avctx->pix_fmt) { - case PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV422P: bit_depth = 8; break; - case PIX_FMT_YUV422P10: + case AV_PIX_FMT_YUV422P10: bit_depth = 10; break; default: @@ -1005,9 +1005,9 @@ AVCodec ff_dnxhd_encoder = { .encode2 = dnxhd_encode_picture, .close = dnxhd_encode_end, .capabilities = CODEC_CAP_SLICE_THREADS, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV422P, - PIX_FMT_YUV422P10, - PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV422P, + AV_PIX_FMT_YUV422P10, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("VC3/DNxHD"), .priv_class = &class, }; diff --git a/libavcodec/dpx.c b/libavcodec/dpx.c index 18d74fea70..f1a4e86c87 100644 --- a/libavcodec/dpx.c +++ b/libavcodec/dpx.c @@ -126,24 +126,24 @@ static int decode_frame(AVCodecContext *avctx, switch (bits_per_color) { case 8: if (elements == 4) { - avctx->pix_fmt = PIX_FMT_RGBA; + avctx->pix_fmt = AV_PIX_FMT_RGBA; } else { - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; } source_packet_size = elements; target_packet_size = elements; break; case 10: - avctx->pix_fmt = PIX_FMT_RGB48; + avctx->pix_fmt = AV_PIX_FMT_RGB48; target_packet_size = 6; source_packet_size = 4; break; case 12: case 16: if (endian) { - avctx->pix_fmt = PIX_FMT_RGB48BE; + avctx->pix_fmt = AV_PIX_FMT_RGB48BE; } else { - avctx->pix_fmt = PIX_FMT_RGB48LE; + avctx->pix_fmt = AV_PIX_FMT_RGB48LE; } target_packet_size = 6; source_packet_size = elements * 2; diff --git a/libavcodec/dpxenc.c b/libavcodec/dpxenc.c index ce84df0360..d2631615a1 100644 --- a/libavcodec/dpxenc.c +++ b/libavcodec/dpxenc.c @@ -45,14 +45,14 @@ static av_cold int encode_init(AVCodecContext *avctx) s->descriptor = 50; /* RGB */ switch (avctx->pix_fmt) { - case PIX_FMT_RGB24: + case AV_PIX_FMT_RGB24: break; - case PIX_FMT_RGBA: + case AV_PIX_FMT_RGBA: s->descriptor = 51; /* RGBA */ break; - case PIX_FMT_RGB48LE: + case AV_PIX_FMT_RGB48LE: s->big_endian = 0; - case PIX_FMT_RGB48BE: + case AV_PIX_FMT_RGB48BE: s->bits_per_component = avctx->bits_per_raw_sample ? avctx->bits_per_raw_sample : 16; break; default: @@ -180,11 +180,11 @@ AVCodec ff_dpx_encoder = { .priv_data_size = sizeof(DPXContext), .init = encode_init, .encode2 = encode_frame, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_RGB24, - PIX_FMT_RGBA, - PIX_FMT_RGB48LE, - PIX_FMT_RGB48BE, - PIX_FMT_NONE}, + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_RGB24, + AV_PIX_FMT_RGBA, + AV_PIX_FMT_RGB48LE, + AV_PIX_FMT_RGB48BE, + AV_PIX_FMT_NONE}, .long_name = NULL_IF_CONFIG_SMALL("DPX image"), }; diff --git a/libavcodec/dsicinav.c b/libavcodec/dsicinav.c index b8fad7e550..1492717f66 100644 --- a/libavcodec/dsicinav.c +++ b/libavcodec/dsicinav.c @@ -93,7 +93,7 @@ static av_cold int cinvideo_decode_init(AVCodecContext *avctx) unsigned int i; cin->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; cin->frame.data[0] = NULL; diff --git a/libavcodec/dv.c b/libavcodec/dv.c index 67e6fd7bc5..7067f0caf2 100644 --- a/libavcodec/dv.c +++ b/libavcodec/dv.c @@ -137,19 +137,19 @@ static inline void dv_calc_mb_coordinates(const DVprofile *d, int chan, int seq, break; case 720: switch (d->pix_fmt) { - case PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV422P: x = shuf3[m] + slot/3; y = serpent1[slot] + ((((seq + off[m]) % d->difseg_size)<<1) + chan)*3; tbl[m] = (x<<1)|(y<<8); break; - case PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV420P: x = shuf3[m] + slot/3; y = serpent1[slot] + ((seq + off[m]) % d->difseg_size)*3; tbl[m] = (x<<1)|(y<<9); break; - case PIX_FMT_YUV411P: + case AV_PIX_FMT_YUV411P: i = (seq + off[m]) % d->difseg_size; k = slot + ((m==1||m==2)?3:0); @@ -677,8 +677,8 @@ static int dv_encode_video_segment(AVCodecContext *avctx, void *arg) dv_calculate_mb_xy(s, work_chunk, mb_index, &mb_x, &mb_y); /* initializing luminance blocks */ - if ((s->sys->pix_fmt == PIX_FMT_YUV420P) || - (s->sys->pix_fmt == PIX_FMT_YUV411P && mb_x >= (704 / 8)) || + if ((s->sys->pix_fmt == AV_PIX_FMT_YUV420P) || + (s->sys->pix_fmt == AV_PIX_FMT_YUV411P && mb_x >= (704 / 8)) || (s->sys->height >= 720 && mb_y != 134)) { y_stride = s->picture.linesize[0] << 3; } else { @@ -703,13 +703,13 @@ static int dv_encode_video_segment(AVCodecContext *avctx, void *arg) enc_blk += 4; /* initializing chrominance blocks */ - c_offset = (((mb_y >> (s->sys->pix_fmt == PIX_FMT_YUV420P)) * s->picture.linesize[1] + - (mb_x >> ((s->sys->pix_fmt == PIX_FMT_YUV411P) ? 2 : 1))) << 3); + c_offset = (((mb_y >> (s->sys->pix_fmt == AV_PIX_FMT_YUV420P)) * s->picture.linesize[1] + + (mb_x >> ((s->sys->pix_fmt == AV_PIX_FMT_YUV411P) ? 2 : 1))) << 3); for (j = 2; j; j--) { uint8_t *c_ptr = s->picture.data[j] + c_offset; linesize = s->picture.linesize[j]; y_stride = (mb_y == 134) ? 8 : (s->picture.linesize[j] << 3); - if (s->sys->pix_fmt == PIX_FMT_YUV411P && mb_x >= (704 / 8)) { + if (s->sys->pix_fmt == AV_PIX_FMT_YUV411P && mb_x >= (704 / 8)) { uint8_t* d; uint8_t* b = scratch; for (i = 0; i < 8; i++) { @@ -806,7 +806,7 @@ static inline int dv_write_pack(enum dv_pack_type pack_id, DVVideoContext *c, * 2. It is not at all clear what STYPE is used for 4:2:0 PAL * compression scheme (if any). */ - int apt = (c->sys->pix_fmt == PIX_FMT_YUV420P ? 0 : 1); + int apt = (c->sys->pix_fmt == AV_PIX_FMT_YUV420P ? 0 : 1); uint8_t aspect = 0; if ((int)(av_q2d(c->avctx->sample_aspect_ratio) * c->avctx->width / c->avctx->height * 10) >= 17) /* 16:9 */ @@ -984,8 +984,8 @@ AVCodec ff_dvvideo_encoder = { .init = dvvideo_init_encoder, .encode2 = dvvideo_encode_frame, .capabilities = CODEC_CAP_SLICE_THREADS, - .pix_fmts = (const enum PixelFormat[]) { - PIX_FMT_YUV411P, PIX_FMT_YUV422P, PIX_FMT_YUV420P, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]) { + AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("DV (Digital Video)"), }; diff --git a/libavcodec/dv_profile.c b/libavcodec/dv_profile.c index b494f97fc1..be392bbcaa 100644 --- a/libavcodec/dv_profile.c +++ b/libavcodec/dv_profile.c @@ -90,7 +90,7 @@ static const DVprofile dv_profiles[] = { .sar = {{8, 9}, {32, 27}}, .work_chunks = &work_chunks_dv25ntsc[0], .idct_factor = &dv_idct_factor_sd[0], - .pix_fmt = PIX_FMT_YUV411P, + .pix_fmt = AV_PIX_FMT_YUV411P, .bpm = 6, .block_sizes = block_sizes_dv2550, .audio_stride = 90, @@ -110,7 +110,7 @@ static const DVprofile dv_profiles[] = { .sar = {{16, 15}, {64, 45}}, .work_chunks = &work_chunks_dv25pal[0], .idct_factor = &dv_idct_factor_sd[0], - .pix_fmt = PIX_FMT_YUV420P, + .pix_fmt = AV_PIX_FMT_YUV420P, .bpm = 6, .block_sizes = block_sizes_dv2550, .audio_stride = 108, @@ -130,7 +130,7 @@ static const DVprofile dv_profiles[] = { .sar = {{16, 15}, {64, 45}}, .work_chunks = &work_chunks_dv25pal411[0], .idct_factor = &dv_idct_factor_sd[0], - .pix_fmt = PIX_FMT_YUV411P, + .pix_fmt = AV_PIX_FMT_YUV411P, .bpm = 6, .block_sizes = block_sizes_dv2550, .audio_stride = 108, @@ -150,7 +150,7 @@ static const DVprofile dv_profiles[] = { .sar = {{8, 9}, {32, 27}}, .work_chunks = &work_chunks_dv50ntsc[0], .idct_factor = &dv_idct_factor_sd[0], - .pix_fmt = PIX_FMT_YUV422P, + .pix_fmt = AV_PIX_FMT_YUV422P, .bpm = 6, .block_sizes = block_sizes_dv2550, .audio_stride = 90, @@ -170,7 +170,7 @@ static const DVprofile dv_profiles[] = { .sar = {{16, 15}, {64, 45}}, .work_chunks = &work_chunks_dv50pal[0], .idct_factor = &dv_idct_factor_sd[0], - .pix_fmt = PIX_FMT_YUV422P, + .pix_fmt = AV_PIX_FMT_YUV422P, .bpm = 6, .block_sizes = block_sizes_dv2550, .audio_stride = 108, @@ -190,7 +190,7 @@ static const DVprofile dv_profiles[] = { .sar = {{1, 1}, {3, 2}}, .work_chunks = &work_chunks_dv100ntsci[0], .idct_factor = &dv_idct_factor_hd1080[0], - .pix_fmt = PIX_FMT_YUV422P, + .pix_fmt = AV_PIX_FMT_YUV422P, .bpm = 8, .block_sizes = block_sizes_dv100, .audio_stride = 90, @@ -210,7 +210,7 @@ static const DVprofile dv_profiles[] = { .sar = {{1, 1}, {4, 3}}, .work_chunks = &work_chunks_dv100pali[0], .idct_factor = &dv_idct_factor_hd1080[0], - .pix_fmt = PIX_FMT_YUV422P, + .pix_fmt = AV_PIX_FMT_YUV422P, .bpm = 8, .block_sizes = block_sizes_dv100, .audio_stride = 108, @@ -230,7 +230,7 @@ static const DVprofile dv_profiles[] = { .sar = {{1, 1}, {4, 3}}, .work_chunks = &work_chunks_dv100ntscp[0], .idct_factor = &dv_idct_factor_hd720[0], - .pix_fmt = PIX_FMT_YUV422P, + .pix_fmt = AV_PIX_FMT_YUV422P, .bpm = 8, .block_sizes = block_sizes_dv100, .audio_stride = 90, @@ -250,7 +250,7 @@ static const DVprofile dv_profiles[] = { .sar = {{1, 1}, {4, 3}}, .work_chunks = &work_chunks_dv100palp[0], .idct_factor = &dv_idct_factor_hd720[0], - .pix_fmt = PIX_FMT_YUV422P, + .pix_fmt = AV_PIX_FMT_YUV422P, .bpm = 8, .block_sizes = block_sizes_dv100, .audio_stride = 90, @@ -270,7 +270,7 @@ static const DVprofile dv_profiles[] = { .sar = {{16, 15}, {64, 45}}, .work_chunks = &work_chunks_dv25pal[0], .idct_factor = &dv_idct_factor_sd[0], - .pix_fmt = PIX_FMT_YUV420P, + .pix_fmt = AV_PIX_FMT_YUV420P, .bpm = 6, .block_sizes = block_sizes_dv2550, .audio_stride = 108, diff --git a/libavcodec/dv_profile.h b/libavcodec/dv_profile.h index 4fcf3e12ed..c6d2278b0d 100644 --- a/libavcodec/dv_profile.h +++ b/libavcodec/dv_profile.h @@ -49,7 +49,7 @@ typedef struct DVprofile { AVRational sar[2]; /* sample aspect ratios for 4:3 and 16:9 */ DVwork_chunk *work_chunks; /* each thread gets its own chunk of frame to work on */ uint32_t *idct_factor; /* set of iDCT factor tables */ - enum PixelFormat pix_fmt; /* picture pixel format */ + enum AVPixelFormat pix_fmt; /* picture pixel format */ int bpm; /* blocks per macroblock */ const uint8_t *block_sizes; /* AC block sizes, in bits */ int audio_stride; /* size of audio_shuffle table */ diff --git a/libavcodec/dvdec.c b/libavcodec/dvdec.c index 58889ec526..2917de93fa 100644 --- a/libavcodec/dvdec.c +++ b/libavcodec/dvdec.c @@ -253,8 +253,8 @@ static int dv_decode_video_segment(AVCodecContext *avctx, void *arg) dv_calculate_mb_xy(s, work_chunk, mb_index, &mb_x, &mb_y); /* idct_put'ting luminance */ - if ((s->sys->pix_fmt == PIX_FMT_YUV420P) || - (s->sys->pix_fmt == PIX_FMT_YUV411P && mb_x >= (704 / 8)) || + if ((s->sys->pix_fmt == AV_PIX_FMT_YUV420P) || + (s->sys->pix_fmt == AV_PIX_FMT_YUV411P && mb_x >= (704 / 8)) || (s->sys->height >= 720 && mb_y != 134)) { y_stride = (s->picture.linesize[0] << ((!is_field_mode[mb_index]) * log2_blocksize)); } else { @@ -274,11 +274,11 @@ static int dv_decode_video_segment(AVCodecContext *avctx, void *arg) block += 4*64; /* idct_put'ting chrominance */ - c_offset = (((mb_y >> (s->sys->pix_fmt == PIX_FMT_YUV420P)) * s->picture.linesize[1] + - (mb_x >> ((s->sys->pix_fmt == PIX_FMT_YUV411P) ? 2 : 1))) << log2_blocksize); + c_offset = (((mb_y >> (s->sys->pix_fmt == AV_PIX_FMT_YUV420P)) * s->picture.linesize[1] + + (mb_x >> ((s->sys->pix_fmt == AV_PIX_FMT_YUV411P) ? 2 : 1))) << log2_blocksize); for (j = 2; j; j--) { uint8_t *c_ptr = s->picture.data[j] + c_offset; - if (s->sys->pix_fmt == PIX_FMT_YUV411P && mb_x >= (704 / 8)) { + if (s->sys->pix_fmt == AV_PIX_FMT_YUV411P && mb_x >= (704 / 8)) { uint64_t aligned_pixels[64/8]; uint8_t *pixels = (uint8_t*)aligned_pixels; uint8_t *c_ptr1, *ptr1; diff --git a/libavcodec/dxa.c b/libavcodec/dxa.c index a6e2a11155..84159a4ae9 100644 --- a/libavcodec/dxa.c +++ b/libavcodec/dxa.c @@ -292,7 +292,7 @@ static av_cold int decode_init(AVCodecContext *avctx) { DxaDecContext * const c = avctx->priv_data; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; c->dsize = avctx->width * avctx->height * 2; if((c->decomp_buf = av_malloc(c->dsize)) == NULL) { diff --git a/libavcodec/dxtory.c b/libavcodec/dxtory.c index 83abc7253a..f741078092 100644 --- a/libavcodec/dxtory.c +++ b/libavcodec/dxtory.c @@ -26,7 +26,7 @@ static av_cold int decode_init(AVCodecContext *avctx) { - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; avctx->coded_frame = avcodec_alloc_frame(); if (!avctx->coded_frame) return AVERROR(ENOMEM); diff --git a/libavcodec/dxva2_h264.c b/libavcodec/dxva2_h264.c index 086c9a2670..2fd0767f89 100644 --- a/libavcodec/dxva2_h264.c +++ b/libavcodec/dxva2_h264.c @@ -443,7 +443,7 @@ AVHWAccel ff_h264_dxva2_hwaccel = { .name = "h264_dxva2", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_H264, - .pix_fmt = PIX_FMT_DXVA2_VLD, + .pix_fmt = AV_PIX_FMT_DXVA2_VLD, .start_frame = start_frame, .decode_slice = decode_slice, .end_frame = end_frame, diff --git a/libavcodec/dxva2_mpeg2.c b/libavcodec/dxva2_mpeg2.c index 5c9343cfc1..fa6ae7b7a3 100644 --- a/libavcodec/dxva2_mpeg2.c +++ b/libavcodec/dxva2_mpeg2.c @@ -264,7 +264,7 @@ AVHWAccel ff_mpeg2_dxva2_hwaccel = { .name = "mpeg2_dxva2", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_MPEG2VIDEO, - .pix_fmt = PIX_FMT_DXVA2_VLD, + .pix_fmt = AV_PIX_FMT_DXVA2_VLD, .start_frame = start_frame, .decode_slice = decode_slice, .end_frame = end_frame, diff --git a/libavcodec/dxva2_vc1.c b/libavcodec/dxva2_vc1.c index 487f5fca17..5aed2f3e75 100644 --- a/libavcodec/dxva2_vc1.c +++ b/libavcodec/dxva2_vc1.c @@ -269,7 +269,7 @@ AVHWAccel ff_wmv3_dxva2_hwaccel = { .name = "wmv3_dxva2", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_WMV3, - .pix_fmt = PIX_FMT_DXVA2_VLD, + .pix_fmt = AV_PIX_FMT_DXVA2_VLD, .start_frame = start_frame, .decode_slice = decode_slice, .end_frame = end_frame, @@ -281,7 +281,7 @@ AVHWAccel ff_vc1_dxva2_hwaccel = { .name = "vc1_dxva2", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_VC1, - .pix_fmt = PIX_FMT_DXVA2_VLD, + .pix_fmt = AV_PIX_FMT_DXVA2_VLD, .start_frame = start_frame, .decode_slice = decode_slice, .end_frame = end_frame, diff --git a/libavcodec/eacmv.c b/libavcodec/eacmv.c index 50cb686fbd..532bf94757 100644 --- a/libavcodec/eacmv.c +++ b/libavcodec/eacmv.c @@ -45,7 +45,7 @@ typedef struct CmvContext { static av_cold int cmv_decode_init(AVCodecContext *avctx){ CmvContext *s = avctx->priv_data; s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/eamad.c b/libavcodec/eamad.c index c3ec170f4e..c45b588ae7 100644 --- a/libavcodec/eamad.c +++ b/libavcodec/eamad.c @@ -61,7 +61,7 @@ static av_cold int decode_init(AVCodecContext *avctx) { MadContext *s = avctx->priv_data; s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; ff_dsputil_init(&s->dsp, avctx); ff_init_scantable_permutation(s->dsp.idct_permutation, FF_NO_IDCT_PERM); ff_init_scantable(s->dsp.idct_permutation, &s->scantable, ff_zigzag_direct); diff --git a/libavcodec/eatgq.c b/libavcodec/eatgq.c index faba0c702d..d6ee635252 100644 --- a/libavcodec/eatgq.c +++ b/libavcodec/eatgq.c @@ -53,7 +53,7 @@ static av_cold int tgq_decode_init(AVCodecContext *avctx){ ff_init_scantable_permutation(idct_permutation, FF_NO_IDCT_PERM); ff_init_scantable(idct_permutation, &s->scantable, ff_zigzag_direct); avctx->time_base = (AVRational){1, 15}; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; return 0; } diff --git a/libavcodec/eatgv.c b/libavcodec/eatgv.c index 118cc0bdcc..b29c99418c 100644 --- a/libavcodec/eatgv.c +++ b/libavcodec/eatgv.c @@ -54,7 +54,7 @@ static av_cold int tgv_decode_init(AVCodecContext *avctx){ TgvContext *s = avctx->priv_data; s->avctx = avctx; avctx->time_base = (AVRational){1, 15}; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/eatqi.c b/libavcodec/eatqi.c index 75edcd60c5..4c0732785e 100644 --- a/libavcodec/eatqi.c +++ b/libavcodec/eatqi.c @@ -52,7 +52,7 @@ static av_cold int tqi_decode_init(AVCodecContext *avctx) ff_init_scantable(s->dsp.idct_permutation, &s->intra_scantable, ff_zigzag_direct); s->qscale = 1; avctx->time_base = (AVRational){1, 15}; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; ff_mpeg12_init_vlcs(); return 0; } diff --git a/libavcodec/escape124.c b/libavcodec/escape124.c index 0c3339b158..40224fb99d 100644 --- a/libavcodec/escape124.c +++ b/libavcodec/escape124.c @@ -61,7 +61,7 @@ static av_cold int escape124_decode_init(AVCodecContext *avctx) { Escape124Context *s = avctx->priv_data; - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; s->num_superblocks = ((unsigned)avctx->width / 8) * ((unsigned)avctx->height / 8); diff --git a/libavcodec/ffv1.c b/libavcodec/ffv1.c index a257803349..e6ee96d722 100644 --- a/libavcodec/ffv1.c +++ b/libavcodec/ffv1.c @@ -901,9 +901,9 @@ static av_cold int encode_init(AVCodecContext *avctx) avctx->coded_frame= &s->picture; switch(avctx->pix_fmt){ - case PIX_FMT_YUV444P16: - case PIX_FMT_YUV422P16: - case PIX_FMT_YUV420P16: + case AV_PIX_FMT_YUV444P16: + case AV_PIX_FMT_YUV422P16: + case AV_PIX_FMT_YUV420P16: if(avctx->bits_per_raw_sample <=8){ av_log(avctx, AV_LOG_ERROR, "bits_per_raw_sample invalid\n"); return -1; @@ -913,14 +913,14 @@ static av_cold int encode_init(AVCodecContext *avctx) return -1; } s->version= FFMAX(s->version, 1); - case PIX_FMT_YUV444P: - case PIX_FMT_YUV422P: - case PIX_FMT_YUV420P: - case PIX_FMT_YUV411P: - case PIX_FMT_YUV410P: + case AV_PIX_FMT_YUV444P: + case AV_PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV411P: + case AV_PIX_FMT_YUV410P: s->colorspace= 0; break; - case PIX_FMT_RGB32: + case AV_PIX_FMT_RGB32: s->colorspace= 1; break; default: @@ -1547,20 +1547,20 @@ static int read_header(FFV1Context *f){ if(f->colorspace==0){ if(f->avctx->bits_per_raw_sample<=8){ switch(16*f->chroma_h_shift + f->chroma_v_shift){ - case 0x00: f->avctx->pix_fmt= PIX_FMT_YUV444P; break; - case 0x10: f->avctx->pix_fmt= PIX_FMT_YUV422P; break; - case 0x11: f->avctx->pix_fmt= PIX_FMT_YUV420P; break; - case 0x20: f->avctx->pix_fmt= PIX_FMT_YUV411P; break; - case 0x22: f->avctx->pix_fmt= PIX_FMT_YUV410P; break; + case 0x00: f->avctx->pix_fmt= AV_PIX_FMT_YUV444P; break; + case 0x10: f->avctx->pix_fmt= AV_PIX_FMT_YUV422P; break; + case 0x11: f->avctx->pix_fmt= AV_PIX_FMT_YUV420P; break; + case 0x20: f->avctx->pix_fmt= AV_PIX_FMT_YUV411P; break; + case 0x22: f->avctx->pix_fmt= AV_PIX_FMT_YUV410P; break; default: av_log(f->avctx, AV_LOG_ERROR, "format not supported\n"); return -1; } }else{ switch(16*f->chroma_h_shift + f->chroma_v_shift){ - case 0x00: f->avctx->pix_fmt= PIX_FMT_YUV444P16; break; - case 0x10: f->avctx->pix_fmt= PIX_FMT_YUV422P16; break; - case 0x11: f->avctx->pix_fmt= PIX_FMT_YUV420P16; break; + case 0x00: f->avctx->pix_fmt= AV_PIX_FMT_YUV444P16; break; + case 0x10: f->avctx->pix_fmt= AV_PIX_FMT_YUV422P16; break; + case 0x11: f->avctx->pix_fmt= AV_PIX_FMT_YUV420P16; break; default: av_log(f->avctx, AV_LOG_ERROR, "format not supported\n"); return -1; @@ -1571,7 +1571,7 @@ static int read_header(FFV1Context *f){ av_log(f->avctx, AV_LOG_ERROR, "chroma subsampling not supported in this colorspace\n"); return -1; } - f->avctx->pix_fmt= PIX_FMT_RGB32; + f->avctx->pix_fmt= AV_PIX_FMT_RGB32; }else{ av_log(f->avctx, AV_LOG_ERROR, "colorspace not supported\n"); return -1; @@ -1759,10 +1759,10 @@ AVCodec ff_ffv1_encoder = { .encode2 = encode_frame, .close = common_end, .capabilities = CODEC_CAP_SLICE_THREADS, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_YUV420P, PIX_FMT_YUV444P, PIX_FMT_YUV422P, PIX_FMT_YUV411P, - PIX_FMT_YUV410P, PIX_FMT_RGB32, PIX_FMT_YUV420P16, PIX_FMT_YUV422P16, - PIX_FMT_YUV444P16, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV411P, + AV_PIX_FMT_YUV410P, AV_PIX_FMT_RGB32, AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, + AV_PIX_FMT_YUV444P16, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("FFmpeg video codec #1"), }; diff --git a/libavcodec/flashsv.c b/libavcodec/flashsv.c index aaaf95622f..1f21c8bd29 100644 --- a/libavcodec/flashsv.c +++ b/libavcodec/flashsv.c @@ -114,7 +114,7 @@ static av_cold int flashsv_decode_init(AVCodecContext *avctx) av_log(avctx, AV_LOG_ERROR, "Inflate init error: %d\n", zret); return 1; } - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; s->frame.data[0] = NULL; return 0; @@ -461,7 +461,7 @@ AVCodec ff_flashsv_decoder = { .close = flashsv_decode_end, .decode = flashsv_decode_frame, .capabilities = CODEC_CAP_DR1, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_BGR24, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_BGR24, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Flash Screen Video v1"), }; #endif /* CONFIG_FLASHSV_DECODER */ @@ -524,7 +524,7 @@ AVCodec ff_flashsv2_decoder = { .close = flashsv2_decode_end, .decode = flashsv_decode_frame, .capabilities = CODEC_CAP_DR1, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_BGR24, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_BGR24, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Flash Screen Video v2"), }; #endif /* CONFIG_FLASHSV2_DECODER */ diff --git a/libavcodec/flashsvenc.c b/libavcodec/flashsvenc.c index 992a727053..0b788802d6 100644 --- a/libavcodec/flashsvenc.c +++ b/libavcodec/flashsvenc.c @@ -287,6 +287,6 @@ AVCodec ff_flashsv_encoder = { .init = flashsv_encode_init, .encode2 = flashsv_encode_frame, .close = flashsv_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_BGR24, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_BGR24, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Flash Screen Video"), }; diff --git a/libavcodec/flicvideo.c b/libavcodec/flicvideo.c index 831ada093b..2b9b290ba7 100644 --- a/libavcodec/flicvideo.c +++ b/libavcodec/flicvideo.c @@ -111,10 +111,10 @@ static av_cold int flic_decode_init(AVCodecContext *avctx) } switch (depth) { - case 8 : avctx->pix_fmt = PIX_FMT_PAL8; break; - case 15 : avctx->pix_fmt = PIX_FMT_RGB555; break; - case 16 : avctx->pix_fmt = PIX_FMT_RGB565; break; - case 24 : avctx->pix_fmt = PIX_FMT_BGR24; /* Supposedly BGR, but havent any files to test with */ + case 8 : avctx->pix_fmt = AV_PIX_FMT_PAL8; break; + case 15 : avctx->pix_fmt = AV_PIX_FMT_RGB555; break; + case 16 : avctx->pix_fmt = AV_PIX_FMT_RGB565; break; + case 24 : avctx->pix_fmt = AV_PIX_FMT_BGR24; /* Supposedly BGR, but havent any files to test with */ av_log(avctx, AV_LOG_ERROR, "24Bpp FLC/FLX is unsupported due to no test files.\n"); return -1; default : @@ -701,16 +701,16 @@ static int flic_decode_frame(AVCodecContext *avctx, { const uint8_t *buf = avpkt->data; int buf_size = avpkt->size; - if (avctx->pix_fmt == PIX_FMT_PAL8) { + if (avctx->pix_fmt == AV_PIX_FMT_PAL8) { return flic_decode_frame_8BPP(avctx, data, data_size, buf, buf_size); } - else if ((avctx->pix_fmt == PIX_FMT_RGB555) || - (avctx->pix_fmt == PIX_FMT_RGB565)) { + else if ((avctx->pix_fmt == AV_PIX_FMT_RGB555) || + (avctx->pix_fmt == AV_PIX_FMT_RGB565)) { return flic_decode_frame_15_16BPP(avctx, data, data_size, buf, buf_size); } - else if (avctx->pix_fmt == PIX_FMT_BGR24) { + else if (avctx->pix_fmt == AV_PIX_FMT_BGR24) { return flic_decode_frame_24BPP(avctx, data, data_size, buf, buf_size); } diff --git a/libavcodec/flvenc.c b/libavcodec/flvenc.c index 3010eb0e68..5427cbcfc6 100644 --- a/libavcodec/flvenc.c +++ b/libavcodec/flvenc.c @@ -94,7 +94,7 @@ AVCodec ff_flv_encoder = { .init = ff_MPV_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("FLV / Sorenson Spark / Sorenson H.263 (Flash Video)"), .priv_class = &flv_class, }; diff --git a/libavcodec/fraps.c b/libavcodec/fraps.c index 3107ba70bf..3df12809b0 100644 --- a/libavcodec/fraps.c +++ b/libavcodec/fraps.c @@ -61,7 +61,7 @@ static av_cold int decode_init(AVCodecContext *avctx) FrapsContext * const s = avctx->priv_data; avctx->coded_frame = &s->frame; - avctx->pix_fmt= PIX_FMT_NONE; /* set in decode_frame */ + avctx->pix_fmt= AV_PIX_FMT_NONE; /* set in decode_frame */ s->avctx = avctx; s->tmpbuf = NULL; @@ -139,7 +139,7 @@ static int decode_frame(AVCodecContext *avctx, uint32_t *luma1,*luma2,*cb,*cr; uint32_t offs[4]; int i, j, is_chroma, planes; - enum PixelFormat pix_fmt; + enum AVPixelFormat pix_fmt; header = AV_RL32(buf); version = header & 0xff; @@ -156,7 +156,7 @@ static int decode_frame(AVCodecContext *avctx, if (header_size == 8) buf+=4; - pix_fmt = version & 1 ? PIX_FMT_BGR24 : PIX_FMT_YUVJ420P; + pix_fmt = version & 1 ? AV_PIX_FMT_BGR24 : AV_PIX_FMT_YUVJ420P; if (avctx->pix_fmt != pix_fmt && f->data[0]) { avctx->release_buffer(avctx, f); } diff --git a/libavcodec/frwu.c b/libavcodec/frwu.c index 8544da90f5..63fc193c5f 100644 --- a/libavcodec/frwu.c +++ b/libavcodec/frwu.c @@ -29,7 +29,7 @@ static av_cold int decode_init(AVCodecContext *avctx) av_log(avctx, AV_LOG_ERROR, "frwu needs even width\n"); return AVERROR(EINVAL); } - avctx->pix_fmt = PIX_FMT_UYVY422; + avctx->pix_fmt = AV_PIX_FMT_UYVY422; avctx->coded_frame = avcodec_alloc_frame(); if (!avctx->coded_frame) diff --git a/libavcodec/gif.c b/libavcodec/gif.c index 24773faed1..af57fff112 100644 --- a/libavcodec/gif.c +++ b/libavcodec/gif.c @@ -187,9 +187,9 @@ AVCodec ff_gif_encoder = { .init = gif_encode_init, .encode2 = gif_encode_frame, .close = gif_encode_close, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_RGB8, PIX_FMT_BGR8, PIX_FMT_RGB4_BYTE, PIX_FMT_BGR4_BYTE, - PIX_FMT_GRAY8, PIX_FMT_PAL8, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_RGB8, AV_PIX_FMT_BGR8, AV_PIX_FMT_RGB4_BYTE, AV_PIX_FMT_BGR4_BYTE, + AV_PIX_FMT_GRAY8, AV_PIX_FMT_PAL8, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("GIF (Graphics Interchange Format)"), }; diff --git a/libavcodec/gifdec.c b/libavcodec/gifdec.c index 68715c2f66..4bb77895ad 100644 --- a/libavcodec/gifdec.c +++ b/libavcodec/gifdec.c @@ -294,7 +294,7 @@ static int gif_decode_frame(AVCodecContext *avctx, void *data, int *data_size, A if (gif_read_header1(s) < 0) return -1; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; if (av_image_check_size(s->screen_width, s->screen_height, 0, avctx)) return -1; avcodec_set_dimensions(avctx, s->screen_width, s->screen_height); diff --git a/libavcodec/h261dec.c b/libavcodec/h261dec.c index 57fe57a1ef..778088870a 100644 --- a/libavcodec/h261dec.c +++ b/libavcodec/h261dec.c @@ -85,7 +85,7 @@ static av_cold int h261_decode_init(AVCodecContext *avctx){ s->out_format = FMT_H261; s->low_delay= 1; - avctx->pix_fmt= PIX_FMT_YUV420P; + avctx->pix_fmt= AV_PIX_FMT_YUV420P; s->codec_id= avctx->codec->id; diff --git a/libavcodec/h261enc.c b/libavcodec/h261enc.c index 1a9a7748df..d6b4cfcd18 100644 --- a/libavcodec/h261enc.c +++ b/libavcodec/h261enc.c @@ -331,7 +331,7 @@ AVCodec ff_h261_encoder = { .init = ff_MPV_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("H.261"), .priv_class = &h261_class, }; diff --git a/libavcodec/h263dec.c b/libavcodec/h263dec.c index f93d79a675..72bf05243c 100644 --- a/libavcodec/h263dec.c +++ b/libavcodec/h263dec.c @@ -59,7 +59,7 @@ av_cold int ff_h263_decode_init(AVCodecContext *avctx) s->decode_mb= ff_h263_decode_mb; s->low_delay= 1; if (avctx->codec->id == AV_CODEC_ID_MSS2) - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; else avctx->pix_fmt = avctx->get_format(avctx, avctx->codec->pix_fmts); s->unrestricted_mv= 1; diff --git a/libavcodec/h264.c b/libavcodec/h264.c index 068c57a056..3c5870ce0d 100644 --- a/libavcodec/h264.c +++ b/libavcodec/h264.c @@ -59,12 +59,12 @@ static const uint8_t div6[QP_MAX_NUM + 1] = { 7, 7, 7, 7, 7, 7, 8, 8, 8, 8, 8, 8, 9, 9, 9, 9, 9, 9, 10, 10, 10, 10, }; -static const enum PixelFormat hwaccel_pixfmt_list_h264_jpeg_420[] = { - PIX_FMT_DXVA2_VLD, - PIX_FMT_VAAPI_VLD, - PIX_FMT_VDA_VLD, - PIX_FMT_YUVJ420P, - PIX_FMT_NONE +static const enum AVPixelFormat hwaccel_pixfmt_list_h264_jpeg_420[] = { + AV_PIX_FMT_DXVA2_VLD, + AV_PIX_FMT_VAAPI_VLD, + AV_PIX_FMT_VDA_VLD, + AV_PIX_FMT_YUVJ420P, + AV_PIX_FMT_NONE }; /** @@ -2502,35 +2502,35 @@ static int decode_slice_header(H264Context *h, H264Context *h0) case 9: if (CHROMA444) { if (s->avctx->colorspace == AVCOL_SPC_RGB) { - s->avctx->pix_fmt = PIX_FMT_GBRP9; + s->avctx->pix_fmt = AV_PIX_FMT_GBRP9; } else - s->avctx->pix_fmt = PIX_FMT_YUV444P9; + s->avctx->pix_fmt = AV_PIX_FMT_YUV444P9; } else if (CHROMA422) - s->avctx->pix_fmt = PIX_FMT_YUV422P9; + s->avctx->pix_fmt = AV_PIX_FMT_YUV422P9; else - s->avctx->pix_fmt = PIX_FMT_YUV420P9; + s->avctx->pix_fmt = AV_PIX_FMT_YUV420P9; break; case 10: if (CHROMA444) { if (s->avctx->colorspace == AVCOL_SPC_RGB) { - s->avctx->pix_fmt = PIX_FMT_GBRP10; + s->avctx->pix_fmt = AV_PIX_FMT_GBRP10; } else - s->avctx->pix_fmt = PIX_FMT_YUV444P10; + s->avctx->pix_fmt = AV_PIX_FMT_YUV444P10; } else if (CHROMA422) - s->avctx->pix_fmt = PIX_FMT_YUV422P10; + s->avctx->pix_fmt = AV_PIX_FMT_YUV422P10; else - s->avctx->pix_fmt = PIX_FMT_YUV420P10; + s->avctx->pix_fmt = AV_PIX_FMT_YUV420P10; break; case 8: if (CHROMA444) { if (s->avctx->colorspace == AVCOL_SPC_RGB) { - s->avctx->pix_fmt = PIX_FMT_GBRP; + s->avctx->pix_fmt = AV_PIX_FMT_GBRP; } else - s->avctx->pix_fmt = s->avctx->color_range == AVCOL_RANGE_JPEG ? PIX_FMT_YUVJ444P - : PIX_FMT_YUV444P; + s->avctx->pix_fmt = s->avctx->color_range == AVCOL_RANGE_JPEG ? AV_PIX_FMT_YUVJ444P + : AV_PIX_FMT_YUV444P; } else if (CHROMA422) { - s->avctx->pix_fmt = s->avctx->color_range == AVCOL_RANGE_JPEG ? PIX_FMT_YUVJ422P - : PIX_FMT_YUV422P; + s->avctx->pix_fmt = s->avctx->color_range == AVCOL_RANGE_JPEG ? AV_PIX_FMT_YUVJ422P + : AV_PIX_FMT_YUV422P; } else { s->avctx->pix_fmt = s->avctx->get_format(s->avctx, s->avctx->codec->pix_fmts ? @@ -4125,8 +4125,8 @@ AVCodec ff_h264_vdpau_decoder = { .capabilities = CODEC_CAP_DR1 | CODEC_CAP_DELAY | CODEC_CAP_HWACCEL_VDPAU, .flush = flush_dpb, .long_name = NULL_IF_CONFIG_SMALL("H.264 / AVC / MPEG-4 AVC / MPEG-4 part 10 (VDPAU acceleration)"), - .pix_fmts = (const enum PixelFormat[]) { PIX_FMT_VDPAU_H264, - PIX_FMT_NONE}, + .pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_VDPAU_H264, + AV_PIX_FMT_NONE}, .profiles = NULL_IF_CONFIG_SMALL(profiles), }; #endif diff --git a/libavcodec/huffyuv.c b/libavcodec/huffyuv.c index e3787e757b..3da1de8fc5 100644 --- a/libavcodec/huffyuv.c +++ b/libavcodec/huffyuv.c @@ -456,21 +456,21 @@ static av_cold int decode_init(AVCodecContext *avctx) switch (s->bitstream_bpp) { case 12: - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; break; case 16: if (s->yuy2) { - avctx->pix_fmt = PIX_FMT_YUYV422; + avctx->pix_fmt = AV_PIX_FMT_YUYV422; } else { - avctx->pix_fmt = PIX_FMT_YUV422P; + avctx->pix_fmt = AV_PIX_FMT_YUV422P; } break; case 24: case 32: if (s->bgr32) { - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; } else { - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; } break; default: @@ -545,13 +545,13 @@ static av_cold int encode_init(AVCodecContext *avctx) avctx->coded_frame = &s->picture; switch (avctx->pix_fmt) { - case PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV420P: s->bitstream_bpp = 12; break; - case PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV422P: s->bitstream_bpp = 16; break; - case PIX_FMT_RGB32: + case AV_PIX_FMT_RGB32: s->bitstream_bpp = 24; break; default: @@ -573,7 +573,7 @@ static av_cold int encode_init(AVCodecContext *avctx) }else s->context= 0; if (avctx->codec->id == AV_CODEC_ID_HUFFYUV) { - if (avctx->pix_fmt == PIX_FMT_YUV420P) { + if (avctx->pix_fmt == AV_PIX_FMT_YUV420P) { av_log(avctx, AV_LOG_ERROR, "Error: YV12 is not supported by huffyuv; use " "vcodec=ffvhuff or format=422p\n"); @@ -1256,8 +1256,8 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, init_put_bits(&s->pb, pkt->data + size, pkt->size - size); - if (avctx->pix_fmt == PIX_FMT_YUV422P || - avctx->pix_fmt == PIX_FMT_YUV420P) { + if (avctx->pix_fmt == AV_PIX_FMT_YUV422P || + avctx->pix_fmt == AV_PIX_FMT_YUV420P) { int lefty, leftu, leftv, y, cy; put_bits(&s->pb, 8, leftv = p->data[2][0]); @@ -1361,7 +1361,7 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, encode_422_bitstream(s, 0, width); } } - } else if(avctx->pix_fmt == PIX_FMT_RGB32) { + } else if(avctx->pix_fmt == AV_PIX_FMT_RGB32) { uint8_t *data = p->data[0] + (height - 1) * p->linesize[0]; const int stride = -p->linesize[0]; const int fake_stride = -fake_ystride; @@ -1479,8 +1479,8 @@ AVCodec ff_huffyuv_encoder = { .init = encode_init, .encode2 = encode_frame, .close = encode_end, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_YUV422P, PIX_FMT_RGB32, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_YUV422P, AV_PIX_FMT_RGB32, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Huffyuv / HuffYUV"), }; @@ -1495,8 +1495,8 @@ AVCodec ff_ffvhuff_encoder = { .init = encode_init, .encode2 = encode_frame, .close = encode_end, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_YUV420P, PIX_FMT_YUV422P, PIX_FMT_RGB32, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_RGB32, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Huffyuv FFmpeg variant"), }; diff --git a/libavcodec/idcinvideo.c b/libavcodec/idcinvideo.c index b6b9eb710f..476b628d82 100644 --- a/libavcodec/idcinvideo.c +++ b/libavcodec/idcinvideo.c @@ -151,7 +151,7 @@ static av_cold int idcin_decode_init(AVCodecContext *avctx) unsigned char *histograms; s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; /* make sure the Huffman tables make it */ if (s->avctx->extradata_size != HUFFMAN_TABLE_SIZE) { diff --git a/libavcodec/iff.c b/libavcodec/iff.c index 2556acd3ee..69efa2e87c 100644 --- a/libavcodec/iff.c +++ b/libavcodec/iff.c @@ -152,10 +152,10 @@ static av_cold int decode_init(AVCodecContext *avctx) if (avctx->bits_per_coded_sample <= 8) { avctx->pix_fmt = (avctx->bits_per_coded_sample < 8 || - avctx->extradata_size) ? PIX_FMT_PAL8 - : PIX_FMT_GRAY8; + avctx->extradata_size) ? AV_PIX_FMT_PAL8 + : AV_PIX_FMT_GRAY8; } else if (avctx->bits_per_coded_sample <= 32) { - avctx->pix_fmt = PIX_FMT_BGR32; + avctx->pix_fmt = AV_PIX_FMT_BGR32; } else { return AVERROR_INVALIDDATA; } @@ -263,14 +263,14 @@ static int decode_frame_ilbm(AVCodecContext *avctx, } else if ((res = avctx->get_buffer(avctx, &s->frame)) < 0) { av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); return res; - } else if (avctx->bits_per_coded_sample <= 8 && avctx->pix_fmt != PIX_FMT_GRAY8) { + } else if (avctx->bits_per_coded_sample <= 8 && avctx->pix_fmt != AV_PIX_FMT_GRAY8) { if ((res = ff_cmap_read_palette(avctx, (uint32_t*)s->frame.data[1])) < 0) return res; } s->init = 1; if (avctx->codec_tag == MKTAG('I','L','B','M')) { // interleaved - if (avctx->pix_fmt == PIX_FMT_PAL8 || avctx->pix_fmt == PIX_FMT_GRAY8) { + if (avctx->pix_fmt == AV_PIX_FMT_PAL8 || avctx->pix_fmt == AV_PIX_FMT_GRAY8) { for(y = 0; y < avctx->height; y++ ) { uint8_t *row = &s->frame.data[0][ y*s->frame.linesize[0] ]; memset(row, 0, avctx->width); @@ -279,7 +279,7 @@ static int decode_frame_ilbm(AVCodecContext *avctx, buf += s->planesize; } } - } else { // PIX_FMT_BGR32 + } else { // AV_PIX_FMT_BGR32 for(y = 0; y < avctx->height; y++ ) { uint8_t *row = &s->frame.data[0][y*s->frame.linesize[0]]; memset(row, 0, avctx->width << 2); @@ -289,7 +289,7 @@ static int decode_frame_ilbm(AVCodecContext *avctx, } } } - } else if (avctx->pix_fmt == PIX_FMT_PAL8 || avctx->pix_fmt == PIX_FMT_GRAY8) { // IFF-PBM + } else if (avctx->pix_fmt == AV_PIX_FMT_PAL8 || avctx->pix_fmt == AV_PIX_FMT_GRAY8) { // IFF-PBM for(y = 0; y < avctx->height; y++ ) { uint8_t *row = &s->frame.data[0][y * s->frame.linesize[0]]; memcpy(row, buf, FFMIN(avctx->width, buf_end - buf)); @@ -320,14 +320,14 @@ static int decode_frame_byterun1(AVCodecContext *avctx, } else if ((res = avctx->get_buffer(avctx, &s->frame)) < 0) { av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); return res; - } else if (avctx->bits_per_coded_sample <= 8 && avctx->pix_fmt != PIX_FMT_GRAY8) { + } else if (avctx->bits_per_coded_sample <= 8 && avctx->pix_fmt != AV_PIX_FMT_GRAY8) { if ((res = ff_cmap_read_palette(avctx, (uint32_t*)s->frame.data[1])) < 0) return res; } s->init = 1; if (avctx->codec_tag == MKTAG('I','L','B','M')) { //interleaved - if (avctx->pix_fmt == PIX_FMT_PAL8 || avctx->pix_fmt == PIX_FMT_GRAY8) { + if (avctx->pix_fmt == AV_PIX_FMT_PAL8 || avctx->pix_fmt == AV_PIX_FMT_GRAY8) { for(y = 0; y < avctx->height ; y++ ) { uint8_t *row = &s->frame.data[0][ y*s->frame.linesize[0] ]; memset(row, 0, avctx->width); @@ -336,7 +336,7 @@ static int decode_frame_byterun1(AVCodecContext *avctx, decodeplane8(row, s->planebuf, s->planesize, plane); } } - } else { //PIX_FMT_BGR32 + } else { //AV_PIX_FMT_BGR32 for(y = 0; y < avctx->height ; y++ ) { uint8_t *row = &s->frame.data[0][y*s->frame.linesize[0]]; memset(row, 0, avctx->width << 2); diff --git a/libavcodec/imgconvert.c b/libavcodec/imgconvert.c index 547bee2405..c88ce0c308 100644 --- a/libavcodec/imgconvert.c +++ b/libavcodec/imgconvert.c @@ -68,87 +68,87 @@ typedef struct PixFmtInfo { } PixFmtInfo; /* this table gives more information about formats */ -static const PixFmtInfo pix_fmt_info[PIX_FMT_NB] = { +static const PixFmtInfo pix_fmt_info[AV_PIX_FMT_NB] = { /* YUV formats */ - [PIX_FMT_YUV420P] = { + [AV_PIX_FMT_YUV420P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_YUV422P] = { + [AV_PIX_FMT_YUV422P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_YUV444P] = { + [AV_PIX_FMT_YUV444P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_YUYV422] = { + [AV_PIX_FMT_YUYV422] = { .nb_channels = 1, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_UYVY422] = { + [AV_PIX_FMT_UYVY422] = { .nb_channels = 1, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_YUV410P] = { + [AV_PIX_FMT_YUV410P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_YUV411P] = { + [AV_PIX_FMT_YUV411P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_YUV440P] = { + [AV_PIX_FMT_YUV440P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_YUV420P16LE] = { + [AV_PIX_FMT_YUV420P16LE] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 16, }, - [PIX_FMT_YUV422P16LE] = { + [AV_PIX_FMT_YUV422P16LE] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 16, }, - [PIX_FMT_YUV444P16LE] = { + [AV_PIX_FMT_YUV444P16LE] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 16, }, - [PIX_FMT_YUV420P16BE] = { + [AV_PIX_FMT_YUV420P16BE] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 16, }, - [PIX_FMT_YUV422P16BE] = { + [AV_PIX_FMT_YUV422P16BE] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 16, }, - [PIX_FMT_YUV444P16BE] = { + [AV_PIX_FMT_YUV444P16BE] = { .nb_channels = 3, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, @@ -157,7 +157,7 @@ static const PixFmtInfo pix_fmt_info[PIX_FMT_NB] = { /* YUV formats with alpha plane */ - [PIX_FMT_YUVA420P] = { + [AV_PIX_FMT_YUVA420P] = { .nb_channels = 4, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, @@ -165,25 +165,25 @@ static const PixFmtInfo pix_fmt_info[PIX_FMT_NB] = { }, /* JPEG YUV */ - [PIX_FMT_YUVJ420P] = { + [AV_PIX_FMT_YUVJ420P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV_JPEG, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_YUVJ422P] = { + [AV_PIX_FMT_YUVJ422P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV_JPEG, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_YUVJ444P] = { + [AV_PIX_FMT_YUVJ444P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV_JPEG, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_YUVJ440P] = { + [AV_PIX_FMT_YUVJ440P] = { .nb_channels = 3, .color_type = FF_COLOR_YUV_JPEG, .pixel_type = FF_PIXEL_PLANAR, @@ -191,67 +191,67 @@ static const PixFmtInfo pix_fmt_info[PIX_FMT_NB] = { }, /* RGB formats */ - [PIX_FMT_RGB24] = { + [AV_PIX_FMT_RGB24] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_BGR24] = { + [AV_PIX_FMT_BGR24] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_ARGB] = { + [AV_PIX_FMT_ARGB] = { .nb_channels = 4, .is_alpha = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_RGB48BE] = { + [AV_PIX_FMT_RGB48BE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 16, }, - [PIX_FMT_RGB48LE] = { + [AV_PIX_FMT_RGB48LE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 16, }, - [PIX_FMT_RGB565BE] = { + [AV_PIX_FMT_RGB565BE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 5, }, - [PIX_FMT_RGB565LE] = { + [AV_PIX_FMT_RGB565LE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 5, }, - [PIX_FMT_RGB555BE] = { + [AV_PIX_FMT_RGB555BE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 5, }, - [PIX_FMT_RGB555LE] = { + [AV_PIX_FMT_RGB555LE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 5, }, - [PIX_FMT_RGB444BE] = { + [AV_PIX_FMT_RGB444BE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 4, }, - [PIX_FMT_RGB444LE] = { + [AV_PIX_FMT_RGB444LE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, @@ -259,31 +259,31 @@ static const PixFmtInfo pix_fmt_info[PIX_FMT_NB] = { }, /* gray / mono formats */ - [PIX_FMT_GRAY16BE] = { + [AV_PIX_FMT_GRAY16BE] = { .nb_channels = 1, .color_type = FF_COLOR_GRAY, .pixel_type = FF_PIXEL_PLANAR, .depth = 16, }, - [PIX_FMT_GRAY16LE] = { + [AV_PIX_FMT_GRAY16LE] = { .nb_channels = 1, .color_type = FF_COLOR_GRAY, .pixel_type = FF_PIXEL_PLANAR, .depth = 16, }, - [PIX_FMT_GRAY8] = { + [AV_PIX_FMT_GRAY8] = { .nb_channels = 1, .color_type = FF_COLOR_GRAY, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_MONOWHITE] = { + [AV_PIX_FMT_MONOWHITE] = { .nb_channels = 1, .color_type = FF_COLOR_GRAY, .pixel_type = FF_PIXEL_PLANAR, .depth = 1, }, - [PIX_FMT_MONOBLACK] = { + [AV_PIX_FMT_MONOBLACK] = { .nb_channels = 1, .color_type = FF_COLOR_GRAY, .pixel_type = FF_PIXEL_PLANAR, @@ -291,116 +291,116 @@ static const PixFmtInfo pix_fmt_info[PIX_FMT_NB] = { }, /* paletted formats */ - [PIX_FMT_PAL8] = { + [AV_PIX_FMT_PAL8] = { .nb_channels = 4, .is_alpha = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PALETTE, .depth = 8, }, - [PIX_FMT_UYYVYY411] = { + [AV_PIX_FMT_UYYVYY411] = { .nb_channels = 1, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_ABGR] = { + [AV_PIX_FMT_ABGR] = { .nb_channels = 4, .is_alpha = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_BGR565BE] = { + [AV_PIX_FMT_BGR565BE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 5, }, - [PIX_FMT_BGR565LE] = { + [AV_PIX_FMT_BGR565LE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 5, }, - [PIX_FMT_BGR555BE] = { + [AV_PIX_FMT_BGR555BE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 5, }, - [PIX_FMT_BGR555LE] = { + [AV_PIX_FMT_BGR555LE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 5, }, - [PIX_FMT_BGR444BE] = { + [AV_PIX_FMT_BGR444BE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 4, }, - [PIX_FMT_BGR444LE] = { + [AV_PIX_FMT_BGR444LE] = { .nb_channels = 3, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 4, }, - [PIX_FMT_RGB8] = { + [AV_PIX_FMT_RGB8] = { .nb_channels = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_RGB4] = { + [AV_PIX_FMT_RGB4] = { .nb_channels = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 4, }, - [PIX_FMT_RGB4_BYTE] = { + [AV_PIX_FMT_RGB4_BYTE] = { .nb_channels = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_BGR8] = { + [AV_PIX_FMT_BGR8] = { .nb_channels = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_BGR4] = { + [AV_PIX_FMT_BGR4] = { .nb_channels = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 4, }, - [PIX_FMT_BGR4_BYTE] = { + [AV_PIX_FMT_BGR4_BYTE] = { .nb_channels = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_NV12] = { + [AV_PIX_FMT_NV12] = { .nb_channels = 2, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_NV21] = { + [AV_PIX_FMT_NV21] = { .nb_channels = 2, .color_type = FF_COLOR_YUV, .pixel_type = FF_PIXEL_PLANAR, .depth = 8, }, - [PIX_FMT_BGRA] = { + [AV_PIX_FMT_BGRA] = { .nb_channels = 4, .is_alpha = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, .depth = 8, }, - [PIX_FMT_RGBA] = { + [AV_PIX_FMT_RGBA] = { .nb_channels = 4, .is_alpha = 1, .color_type = FF_COLOR_RGB, .pixel_type = FF_PIXEL_PACKED, @@ -408,19 +408,19 @@ static const PixFmtInfo pix_fmt_info[PIX_FMT_NB] = { }, }; -void avcodec_get_chroma_sub_sample(enum PixelFormat pix_fmt, int *h_shift, int *v_shift) +void avcodec_get_chroma_sub_sample(enum AVPixelFormat pix_fmt, int *h_shift, int *v_shift) { *h_shift = av_pix_fmt_descriptors[pix_fmt].log2_chroma_w; *v_shift = av_pix_fmt_descriptors[pix_fmt].log2_chroma_h; } -int ff_is_hwaccel_pix_fmt(enum PixelFormat pix_fmt) +int ff_is_hwaccel_pix_fmt(enum AVPixelFormat pix_fmt) { return av_pix_fmt_descriptors[pix_fmt].flags & PIX_FMT_HWACCEL; } int avpicture_fill(AVPicture *picture, uint8_t *ptr, - enum PixelFormat pix_fmt, int width, int height) + enum AVPixelFormat pix_fmt, int width, int height) { int ret; @@ -433,7 +433,7 @@ int avpicture_fill(AVPicture *picture, uint8_t *ptr, return av_image_fill_pointers(picture->data, pix_fmt, height, ptr, picture->linesize); } -int avpicture_layout(const AVPicture* src, enum PixelFormat pix_fmt, int width, int height, +int avpicture_layout(const AVPicture* src, enum AVPixelFormat pix_fmt, int width, int height, unsigned char *dest, int dest_size) { int i, j, nb_planes = 0, linesizes[4]; @@ -466,7 +466,7 @@ int avpicture_layout(const AVPicture* src, enum PixelFormat pix_fmt, int width, return size; } -int avpicture_get_size(enum PixelFormat pix_fmt, int width, int height) +int avpicture_get_size(enum AVPixelFormat pix_fmt, int width, int height) { AVPicture dummy_pict; if(av_image_check_size(width, height, 0, NULL)) @@ -477,7 +477,7 @@ int avpicture_get_size(enum PixelFormat pix_fmt, int width, int height) return avpicture_fill(&dummy_pict, NULL, pix_fmt, width, height); } -int avcodec_get_pix_fmt_loss(enum PixelFormat dst_pix_fmt, enum PixelFormat src_pix_fmt, +int avcodec_get_pix_fmt_loss(enum AVPixelFormat dst_pix_fmt, enum AVPixelFormat src_pix_fmt, int has_alpha) { const PixFmtInfo *pf, *ps; @@ -491,10 +491,10 @@ int avcodec_get_pix_fmt_loss(enum PixelFormat dst_pix_fmt, enum PixelFormat src_ loss = 0; pf = &pix_fmt_info[dst_pix_fmt]; if (pf->depth < ps->depth || - ((dst_pix_fmt == PIX_FMT_RGB555BE || dst_pix_fmt == PIX_FMT_RGB555LE || - dst_pix_fmt == PIX_FMT_BGR555BE || dst_pix_fmt == PIX_FMT_BGR555LE) && - (src_pix_fmt == PIX_FMT_RGB565BE || src_pix_fmt == PIX_FMT_RGB565LE || - src_pix_fmt == PIX_FMT_BGR565BE || src_pix_fmt == PIX_FMT_BGR565LE))) + ((dst_pix_fmt == AV_PIX_FMT_RGB555BE || dst_pix_fmt == AV_PIX_FMT_RGB555LE || + dst_pix_fmt == AV_PIX_FMT_BGR555BE || dst_pix_fmt == AV_PIX_FMT_BGR555LE) && + (src_pix_fmt == AV_PIX_FMT_RGB565BE || src_pix_fmt == AV_PIX_FMT_RGB565LE || + src_pix_fmt == AV_PIX_FMT_BGR565BE || src_pix_fmt == AV_PIX_FMT_BGR565LE))) loss |= FF_LOSS_DEPTH; if (dst_desc->log2_chroma_w > src_desc->log2_chroma_w || dst_desc->log2_chroma_h > src_desc->log2_chroma_h) @@ -536,7 +536,7 @@ int avcodec_get_pix_fmt_loss(enum PixelFormat dst_pix_fmt, enum PixelFormat src_ return loss; } -static int avg_bits_per_pixel(enum PixelFormat pix_fmt) +static int avg_bits_per_pixel(enum AVPixelFormat pix_fmt) { int bits; const PixFmtInfo *pf; @@ -546,23 +546,23 @@ static int avg_bits_per_pixel(enum PixelFormat pix_fmt) switch(pf->pixel_type) { case FF_PIXEL_PACKED: switch(pix_fmt) { - case PIX_FMT_YUYV422: - case PIX_FMT_UYVY422: - case PIX_FMT_RGB565BE: - case PIX_FMT_RGB565LE: - case PIX_FMT_RGB555BE: - case PIX_FMT_RGB555LE: - case PIX_FMT_RGB444BE: - case PIX_FMT_RGB444LE: - case PIX_FMT_BGR565BE: - case PIX_FMT_BGR565LE: - case PIX_FMT_BGR555BE: - case PIX_FMT_BGR555LE: - case PIX_FMT_BGR444BE: - case PIX_FMT_BGR444LE: + case AV_PIX_FMT_YUYV422: + case AV_PIX_FMT_UYVY422: + case AV_PIX_FMT_RGB565BE: + case AV_PIX_FMT_RGB565LE: + case AV_PIX_FMT_RGB555BE: + case AV_PIX_FMT_RGB555LE: + case AV_PIX_FMT_RGB444BE: + case AV_PIX_FMT_RGB444LE: + case AV_PIX_FMT_BGR565BE: + case AV_PIX_FMT_BGR565LE: + case AV_PIX_FMT_BGR555BE: + case AV_PIX_FMT_BGR555LE: + case AV_PIX_FMT_BGR444BE: + case AV_PIX_FMT_BGR444LE: bits = 16; break; - case PIX_FMT_UYYVYY411: + case AV_PIX_FMT_UYYVYY411: bits = 12; break; default: @@ -588,25 +588,25 @@ static int avg_bits_per_pixel(enum PixelFormat pix_fmt) return bits; } -static enum PixelFormat avcodec_find_best_pix_fmt1(enum PixelFormat *pix_fmt_list, - enum PixelFormat src_pix_fmt, +static enum AVPixelFormat avcodec_find_best_pix_fmt1(enum AVPixelFormat *pix_fmt_list, + enum AVPixelFormat src_pix_fmt, int has_alpha, int loss_mask) { int dist, i, loss, min_dist; - enum PixelFormat dst_pix_fmt; + enum AVPixelFormat dst_pix_fmt; /* find exact color match with smallest size */ - dst_pix_fmt = PIX_FMT_NONE; + dst_pix_fmt = AV_PIX_FMT_NONE; min_dist = 0x7fffffff; i = 0; - while (pix_fmt_list[i] != PIX_FMT_NONE) { - enum PixelFormat pix_fmt = pix_fmt_list[i]; + while (pix_fmt_list[i] != AV_PIX_FMT_NONE) { + enum AVPixelFormat pix_fmt = pix_fmt_list[i]; - if (i > PIX_FMT_NB) { + if (i > AV_PIX_FMT_NB) { av_log(NULL, AV_LOG_ERROR, "Pixel format list longer than expected, " "it is either not properly terminated or contains duplicates\n"); - return PIX_FMT_NONE; + return AV_PIX_FMT_NONE; } loss = avcodec_get_pix_fmt_loss(pix_fmt, src_pix_fmt, has_alpha) & loss_mask; @@ -623,10 +623,10 @@ static enum PixelFormat avcodec_find_best_pix_fmt1(enum PixelFormat *pix_fmt_lis } #if FF_API_FIND_BEST_PIX_FMT -enum PixelFormat avcodec_find_best_pix_fmt(int64_t pix_fmt_mask, enum PixelFormat src_pix_fmt, +enum AVPixelFormat avcodec_find_best_pix_fmt(int64_t pix_fmt_mask, enum AVPixelFormat src_pix_fmt, int has_alpha, int *loss_ptr) { - enum PixelFormat list[64]; + enum AVPixelFormat list[64]; int i, j = 0; // test only the first 64 pixel formats to avoid undefined behaviour @@ -634,17 +634,17 @@ enum PixelFormat avcodec_find_best_pix_fmt(int64_t pix_fmt_mask, enum PixelForma if (pix_fmt_mask & (1ULL << i)) list[j++] = i; } - list[j] = PIX_FMT_NONE; + list[j] = AV_PIX_FMT_NONE; return avcodec_find_best_pix_fmt2(list, src_pix_fmt, has_alpha, loss_ptr); } #endif /* FF_API_FIND_BEST_PIX_FMT */ -enum PixelFormat avcodec_find_best_pix_fmt2(enum PixelFormat *pix_fmt_list, - enum PixelFormat src_pix_fmt, +enum AVPixelFormat avcodec_find_best_pix_fmt2(enum AVPixelFormat *pix_fmt_list, + enum AVPixelFormat src_pix_fmt, int has_alpha, int *loss_ptr) { - enum PixelFormat dst_pix_fmt; + enum AVPixelFormat dst_pix_fmt; int loss_mask, i; static const int loss_mask_order[] = { ~0, /* no loss first */ @@ -667,7 +667,7 @@ enum PixelFormat avcodec_find_best_pix_fmt2(enum PixelFormat *pix_fmt_list, if (loss_mask == 0) break; } - return PIX_FMT_NONE; + return AV_PIX_FMT_NONE; found: if (loss_ptr) *loss_ptr = avcodec_get_pix_fmt_loss(dst_pix_fmt, src_pix_fmt, has_alpha); @@ -675,7 +675,7 @@ enum PixelFormat avcodec_find_best_pix_fmt2(enum PixelFormat *pix_fmt_list, } void av_picture_copy(AVPicture *dst, const AVPicture *src, - enum PixelFormat pix_fmt, int width, int height) + enum AVPixelFormat pix_fmt, int width, int height) { av_image_copy(dst->data, dst->linesize, src->data, src->linesize, pix_fmt, width, height); @@ -769,7 +769,7 @@ void ff_shrink88(uint8_t *dst, int dst_wrap, int avpicture_alloc(AVPicture *picture, - enum PixelFormat pix_fmt, int width, int height) + enum AVPixelFormat pix_fmt, int width, int height) { int ret; @@ -795,12 +795,12 @@ static inline int is_yuv_planar(const PixFmtInfo *ps) } int av_picture_crop(AVPicture *dst, const AVPicture *src, - enum PixelFormat pix_fmt, int top_band, int left_band) + enum AVPixelFormat pix_fmt, int top_band, int left_band) { int y_shift; int x_shift; - if (pix_fmt < 0 || pix_fmt >= PIX_FMT_NB || !is_yuv_planar(&pix_fmt_info[pix_fmt])) + if (pix_fmt < 0 || pix_fmt >= AV_PIX_FMT_NB || !is_yuv_planar(&pix_fmt_info[pix_fmt])) return -1; y_shift = av_pix_fmt_descriptors[pix_fmt].log2_chroma_h; @@ -817,7 +817,7 @@ int av_picture_crop(AVPicture *dst, const AVPicture *src, } int av_picture_pad(AVPicture *dst, const AVPicture *src, int height, int width, - enum PixelFormat pix_fmt, int padtop, int padbottom, int padleft, int padright, + enum AVPixelFormat pix_fmt, int padtop, int padbottom, int padleft, int padright, int *color) { uint8_t *optr; @@ -826,7 +826,7 @@ int av_picture_pad(AVPicture *dst, const AVPicture *src, int height, int width, int yheight; int i, y; - if (pix_fmt < 0 || pix_fmt >= PIX_FMT_NB || + if (pix_fmt < 0 || pix_fmt >= AV_PIX_FMT_NB || !is_yuv_planar(&pix_fmt_info[pix_fmt])) return -1; for (i = 0; i < 3; i++) { @@ -985,17 +985,17 @@ static void deinterlace_bottom_field_inplace(uint8_t *src1, int src_wrap, } int avpicture_deinterlace(AVPicture *dst, const AVPicture *src, - enum PixelFormat pix_fmt, int width, int height) + enum AVPixelFormat pix_fmt, int width, int height) { int i; - if (pix_fmt != PIX_FMT_YUV420P && - pix_fmt != PIX_FMT_YUVJ420P && - pix_fmt != PIX_FMT_YUV422P && - pix_fmt != PIX_FMT_YUVJ422P && - pix_fmt != PIX_FMT_YUV444P && - pix_fmt != PIX_FMT_YUV411P && - pix_fmt != PIX_FMT_GRAY8) + if (pix_fmt != AV_PIX_FMT_YUV420P && + pix_fmt != AV_PIX_FMT_YUVJ420P && + pix_fmt != AV_PIX_FMT_YUV422P && + pix_fmt != AV_PIX_FMT_YUVJ422P && + pix_fmt != AV_PIX_FMT_YUV444P && + pix_fmt != AV_PIX_FMT_YUV411P && + pix_fmt != AV_PIX_FMT_GRAY8) return -1; if ((width & 3) != 0 || (height & 3) != 0) return -1; @@ -1003,22 +1003,22 @@ int avpicture_deinterlace(AVPicture *dst, const AVPicture *src, for(i=0;i<3;i++) { if (i == 1) { switch(pix_fmt) { - case PIX_FMT_YUVJ420P: - case PIX_FMT_YUV420P: + case AV_PIX_FMT_YUVJ420P: + case AV_PIX_FMT_YUV420P: width >>= 1; height >>= 1; break; - case PIX_FMT_YUV422P: - case PIX_FMT_YUVJ422P: + case AV_PIX_FMT_YUV422P: + case AV_PIX_FMT_YUVJ422P: width >>= 1; break; - case PIX_FMT_YUV411P: + case AV_PIX_FMT_YUV411P: width >>= 2; break; default: break; } - if (pix_fmt == PIX_FMT_GRAY8) { + if (pix_fmt == AV_PIX_FMT_GRAY8) { break; } } diff --git a/libavcodec/indeo2.c b/libavcodec/indeo2.c index ac0da05065..0fa79223e4 100644 --- a/libavcodec/indeo2.c +++ b/libavcodec/indeo2.c @@ -203,7 +203,7 @@ static av_cold int ir2_decode_init(AVCodecContext *avctx){ ic->avctx = avctx; - avctx->pix_fmt= PIX_FMT_YUV410P; + avctx->pix_fmt= AV_PIX_FMT_YUV410P; ir2_vlc.table = vlc_tables; ir2_vlc.table_allocated = 1 << CODE_VLC_BITS; diff --git a/libavcodec/indeo3.c b/libavcodec/indeo3.c index 18d9e10270..c99cee801c 100644 --- a/libavcodec/indeo3.c +++ b/libavcodec/indeo3.c @@ -1009,7 +1009,7 @@ static av_cold int decode_init(AVCodecContext *avctx) ctx->avctx = avctx; ctx->width = avctx->width; ctx->height = avctx->height; - avctx->pix_fmt = PIX_FMT_YUV410P; + avctx->pix_fmt = AV_PIX_FMT_YUV410P; build_requant_tab(); diff --git a/libavcodec/indeo4.c b/libavcodec/indeo4.c index c2b8425b64..b837ffb627 100644 --- a/libavcodec/indeo4.c +++ b/libavcodec/indeo4.c @@ -586,7 +586,7 @@ static av_cold int decode_init(AVCodecContext *avctx) ctx->pic_conf.pic_width = 0; ctx->pic_conf.pic_height = 0; - avctx->pix_fmt = PIX_FMT_YUV410P; + avctx->pix_fmt = AV_PIX_FMT_YUV410P; ctx->decode_pic_hdr = decode_pic_hdr; ctx->decode_band_hdr = decode_band_hdr; diff --git a/libavcodec/indeo5.c b/libavcodec/indeo5.c index 4044a04ba2..e9f7733f31 100644 --- a/libavcodec/indeo5.c +++ b/libavcodec/indeo5.c @@ -627,7 +627,7 @@ static av_cold int decode_init(AVCodecContext *avctx) ctx->switch_buffers = switch_buffers; ctx->is_nonnull_frame = is_nonnull_frame; - avctx->pix_fmt = PIX_FMT_YUV410P; + avctx->pix_fmt = AV_PIX_FMT_YUV410P; return 0; } diff --git a/libavcodec/internal.h b/libavcodec/internal.h index 0bfc9247be..e5b1958596 100644 --- a/libavcodec/internal.h +++ b/libavcodec/internal.h @@ -36,7 +36,7 @@ typedef struct InternalBuffer { int linesize[AV_NUM_DATA_POINTERS]; int width; int height; - enum PixelFormat pix_fmt; + enum AVPixelFormat pix_fmt; uint8_t **extended_data; int audio_data_size; int nb_channels; @@ -86,7 +86,7 @@ struct AVCodecDefault { /** * Determine whether pix_fmt is a hardware accelerated format. */ -int ff_is_hwaccel_pix_fmt(enum PixelFormat pix_fmt); +int ff_is_hwaccel_pix_fmt(enum AVPixelFormat pix_fmt); /** * Return the hardware accelerated codec for codec codec_id and @@ -96,7 +96,7 @@ int ff_is_hwaccel_pix_fmt(enum PixelFormat pix_fmt); * @param pix_fmt the pixel format to match * @return the hardware accelerated codec, or NULL if none was found. */ -AVHWAccel *ff_find_hwaccel(enum AVCodecID codec_id, enum PixelFormat pix_fmt); +AVHWAccel *ff_find_hwaccel(enum AVCodecID codec_id, enum AVPixelFormat pix_fmt); /** * Return the index into tab at which {a,b} match elements {[0],[1]} of tab. diff --git a/libavcodec/interplayvideo.c b/libavcodec/interplayvideo.c index 7d15a24ba0..60ccf37a17 100644 --- a/libavcodec/interplayvideo.c +++ b/libavcodec/interplayvideo.c @@ -941,7 +941,7 @@ static av_cold int ipvideo_decode_init(AVCodecContext *avctx) s->avctx = avctx; s->is_16bpp = avctx->bits_per_coded_sample == 16; - avctx->pix_fmt = s->is_16bpp ? PIX_FMT_RGB555 : PIX_FMT_PAL8; + avctx->pix_fmt = s->is_16bpp ? AV_PIX_FMT_RGB555 : AV_PIX_FMT_PAL8; ff_dsputil_init(&s->dsp, avctx); diff --git a/libavcodec/jpeglsenc.c b/libavcodec/jpeglsenc.c index 15e0d111ec..fea2a5baef 100644 --- a/libavcodec/jpeglsenc.c +++ b/libavcodec/jpeglsenc.c @@ -244,7 +244,7 @@ static int encode_picture_ls(AVCodecContext *avctx, AVPacket *pkt, p->pict_type= AV_PICTURE_TYPE_I; p->key_frame= 1; - if(avctx->pix_fmt == PIX_FMT_GRAY8 || avctx->pix_fmt == PIX_FMT_GRAY16) + if(avctx->pix_fmt == AV_PIX_FMT_GRAY8 || avctx->pix_fmt == AV_PIX_FMT_GRAY16) comps = 1; else comps = 3; @@ -264,7 +264,7 @@ static int encode_picture_ls(AVCodecContext *avctx, AVPacket *pkt, put_marker(&pb, SOI); put_marker(&pb, SOF48); put_bits(&pb, 16, 8 + comps * 3); // header size depends on components - put_bits(&pb, 8, (avctx->pix_fmt == PIX_FMT_GRAY16) ? 16 : 8); // bpp + put_bits(&pb, 8, (avctx->pix_fmt == AV_PIX_FMT_GRAY16) ? 16 : 8); // bpp put_bits(&pb, 16, avctx->height); put_bits(&pb, 16, avctx->width); put_bits(&pb, 8, comps); // components @@ -288,7 +288,7 @@ static int encode_picture_ls(AVCodecContext *avctx, AVPacket *pkt, state = av_mallocz(sizeof(JLSState)); /* initialize JPEG-LS state from JPEG parameters */ state->near = near; - state->bpp = (avctx->pix_fmt == PIX_FMT_GRAY16) ? 16 : 8; + state->bpp = (avctx->pix_fmt == AV_PIX_FMT_GRAY16) ? 16 : 8; ff_jpegls_reset_coding_parameters(state, 0); ff_jpegls_init_state(state); @@ -297,7 +297,7 @@ static int encode_picture_ls(AVCodecContext *avctx, AVPacket *pkt, zero = av_mallocz(p->linesize[0]); last = zero; cur = p->data[0]; - if(avctx->pix_fmt == PIX_FMT_GRAY8){ + if(avctx->pix_fmt == AV_PIX_FMT_GRAY8){ int t = 0; for(i = 0; i < avctx->height; i++) { @@ -306,7 +306,7 @@ static int encode_picture_ls(AVCodecContext *avctx, AVPacket *pkt, last = cur; cur += p->linesize[0]; } - }else if(avctx->pix_fmt == PIX_FMT_GRAY16){ + }else if(avctx->pix_fmt == AV_PIX_FMT_GRAY16){ int t = 0; for(i = 0; i < avctx->height; i++) { @@ -315,7 +315,7 @@ static int encode_picture_ls(AVCodecContext *avctx, AVPacket *pkt, last = cur; cur += p->linesize[0]; } - }else if(avctx->pix_fmt == PIX_FMT_RGB24){ + }else if(avctx->pix_fmt == AV_PIX_FMT_RGB24){ int j, width; int Rc[3] = {0, 0, 0}; @@ -328,7 +328,7 @@ static int encode_picture_ls(AVCodecContext *avctx, AVPacket *pkt, last = cur; cur += s->picture.linesize[0]; } - }else if(avctx->pix_fmt == PIX_FMT_BGR24){ + }else if(avctx->pix_fmt == AV_PIX_FMT_BGR24){ int j, width; int Rc[3] = {0, 0, 0}; @@ -385,7 +385,7 @@ static av_cold int encode_init_ls(AVCodecContext *ctx) { c->avctx = ctx; ctx->coded_frame = &c->picture; - if(ctx->pix_fmt != PIX_FMT_GRAY8 && ctx->pix_fmt != PIX_FMT_GRAY16 && ctx->pix_fmt != PIX_FMT_RGB24 && ctx->pix_fmt != PIX_FMT_BGR24){ + if(ctx->pix_fmt != AV_PIX_FMT_GRAY8 && ctx->pix_fmt != AV_PIX_FMT_GRAY16 && ctx->pix_fmt != AV_PIX_FMT_RGB24 && ctx->pix_fmt != AV_PIX_FMT_BGR24){ av_log(ctx, AV_LOG_ERROR, "Only grayscale and RGB24/BGR24 images are supported\n"); return -1; } @@ -399,9 +399,9 @@ AVCodec ff_jpegls_encoder = { //FIXME avoid MPV_* lossless JPEG should not need .priv_data_size = sizeof(JpeglsContext), .init = encode_init_ls, .encode2 = encode_picture_ls, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_BGR24, PIX_FMT_RGB24, PIX_FMT_GRAY8, PIX_FMT_GRAY16, - PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_BGR24, AV_PIX_FMT_RGB24, AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY16, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("JPEG-LS"), }; diff --git a/libavcodec/jvdec.c b/libavcodec/jvdec.c index 2a7b52cd62..ed44e152cc 100644 --- a/libavcodec/jvdec.c +++ b/libavcodec/jvdec.c @@ -40,7 +40,7 @@ typedef struct JvContext { static av_cold int decode_init(AVCodecContext *avctx) { JvContext *s = avctx->priv_data; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; ff_dsputil_init(&s->dsp, avctx); return 0; } diff --git a/libavcodec/kgv1dec.c b/libavcodec/kgv1dec.c index d19563f7e4..8f3294383d 100644 --- a/libavcodec/kgv1dec.c +++ b/libavcodec/kgv1dec.c @@ -169,7 +169,7 @@ static av_cold int decode_init(AVCodecContext *avctx) KgvContext * const c = avctx->priv_data; c->avctx = avctx; - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; avctx->flags |= CODEC_FLAG_EMU_EDGE; return 0; diff --git a/libavcodec/kmvc.c b/libavcodec/kmvc.c index a8c6dbc1cf..2b6774001b 100644 --- a/libavcodec/kmvc.c +++ b/libavcodec/kmvc.c @@ -397,7 +397,7 @@ static av_cold int decode_init(AVCodecContext * avctx) c->setpal = 1; } - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/lagarith.c b/libavcodec/lagarith.c index a34c28941b..f8db42e0ea 100644 --- a/libavcodec/lagarith.c +++ b/libavcodec/lagarith.c @@ -259,7 +259,7 @@ static void lag_pred_line(LagarithContext *l, uint8_t *buf, if (line == 1) { /* Second line, left predict first pixel, the rest of the line is median predicted * NOTE: In the case of RGB this pixel is top predicted */ - TL = l->avctx->pix_fmt == PIX_FMT_YUV420P ? buf[-stride] : L; + TL = l->avctx->pix_fmt == AV_PIX_FMT_YUV420P ? buf[-stride] : L; } else { /* Top left is 2 rows back, last pixel */ TL = buf[width - (2 * stride) - 1]; @@ -472,7 +472,7 @@ static int lag_decode_arith_plane(LagarithContext *l, uint8_t *dst, return -1; } - if (l->avctx->pix_fmt != PIX_FMT_YUV422P) { + if (l->avctx->pix_fmt != AV_PIX_FMT_YUV422P) { for (i = 0; i < height; i++) { lag_pred_line(l, dst, width, stride, i); dst += stride; @@ -524,7 +524,7 @@ static int lag_decode_frame(AVCodecContext *avctx, switch (frametype) { case FRAME_SOLID_RGBA: - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; if (ff_thread_get_buffer(avctx, p) < 0) { av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); @@ -539,14 +539,14 @@ static int lag_decode_frame(AVCodecContext *avctx, } break; case FRAME_ARITH_RGBA: - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; planes = 4; offset_ry += 4; offs[3] = AV_RL32(buf + 9); case FRAME_ARITH_RGB24: case FRAME_U_RGB24: if (frametype == FRAME_ARITH_RGB24 || frametype == FRAME_U_RGB24) - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; if (ff_thread_get_buffer(avctx, p) < 0) { av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); @@ -606,7 +606,7 @@ static int lag_decode_frame(AVCodecContext *avctx, } break; case FRAME_ARITH_YUY2: - avctx->pix_fmt = PIX_FMT_YUV422P; + avctx->pix_fmt = AV_PIX_FMT_YUV422P; if (ff_thread_get_buffer(avctx, p) < 0) { av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); @@ -632,7 +632,7 @@ static int lag_decode_frame(AVCodecContext *avctx, buf + offset_bv, buf_size - offset_bv); break; case FRAME_ARITH_YV12: - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; if (ff_thread_get_buffer(avctx, p) < 0) { av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); diff --git a/libavcodec/lcldec.c b/libavcodec/lcldec.c index d2482138ab..0de7410355 100644 --- a/libavcodec/lcldec.c +++ b/libavcodec/lcldec.c @@ -495,37 +495,37 @@ static av_cold int decode_init(AVCodecContext *avctx) case IMGTYPE_YUV111: c->decomp_size = basesize * 3; max_decomp_size = max_basesize * 3; - avctx->pix_fmt = PIX_FMT_YUV444P; + avctx->pix_fmt = AV_PIX_FMT_YUV444P; av_log(avctx, AV_LOG_DEBUG, "Image type is YUV 1:1:1.\n"); break; case IMGTYPE_YUV422: c->decomp_size = basesize * 2; max_decomp_size = max_basesize * 2; - avctx->pix_fmt = PIX_FMT_YUV422P; + avctx->pix_fmt = AV_PIX_FMT_YUV422P; av_log(avctx, AV_LOG_DEBUG, "Image type is YUV 4:2:2.\n"); break; case IMGTYPE_RGB24: c->decomp_size = basesize * 3; max_decomp_size = max_basesize * 3; - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; av_log(avctx, AV_LOG_DEBUG, "Image type is RGB 24.\n"); break; case IMGTYPE_YUV411: c->decomp_size = basesize / 2 * 3; max_decomp_size = max_basesize / 2 * 3; - avctx->pix_fmt = PIX_FMT_YUV411P; + avctx->pix_fmt = AV_PIX_FMT_YUV411P; av_log(avctx, AV_LOG_DEBUG, "Image type is YUV 4:1:1.\n"); break; case IMGTYPE_YUV211: c->decomp_size = basesize * 2; max_decomp_size = max_basesize * 2; - avctx->pix_fmt = PIX_FMT_YUV422P; + avctx->pix_fmt = AV_PIX_FMT_YUV422P; av_log(avctx, AV_LOG_DEBUG, "Image type is YUV 2:1:1.\n"); break; case IMGTYPE_YUV420: c->decomp_size = basesize / 2 * 3; max_decomp_size = max_basesize / 2 * 3; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; av_log(avctx, AV_LOG_DEBUG, "Image type is YUV 4:2:0.\n"); break; default: diff --git a/libavcodec/lclenc.c b/libavcodec/lclenc.c index 6540e9e82f..0fb303cca6 100644 --- a/libavcodec/lclenc.c +++ b/libavcodec/lclenc.c @@ -89,7 +89,7 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, p->pict_type= AV_PICTURE_TYPE_I; p->key_frame= 1; - if(avctx->pix_fmt != PIX_FMT_BGR24){ + if(avctx->pix_fmt != AV_PIX_FMT_BGR24){ av_log(avctx, AV_LOG_ERROR, "Format not supported!\n"); return -1; } @@ -192,6 +192,6 @@ AVCodec ff_zlib_encoder = { .init = encode_init, .encode2 = encode_frame, .close = encode_end, - .pix_fmts = (const enum PixelFormat[]) { PIX_FMT_BGR24, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_BGR24, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("LCL (LossLess Codec Library) ZLIB"), }; diff --git a/libavcodec/libopenjpegdec.c b/libavcodec/libopenjpegdec.c index 454cdf90d1..bdadcd3385 100644 --- a/libavcodec/libopenjpegdec.c +++ b/libavcodec/libopenjpegdec.c @@ -40,27 +40,27 @@ // pix_fmts with lower bpp have to be listed before // similar pix_fmts with higher bpp. -#define RGB_PIXEL_FORMATS PIX_FMT_RGB24, PIX_FMT_RGBA, \ - PIX_FMT_RGB48 - -#define GRAY_PIXEL_FORMATS PIX_FMT_GRAY8, PIX_FMT_Y400A, \ - PIX_FMT_GRAY16 - -#define YUV_PIXEL_FORMATS PIX_FMT_YUV410P, PIX_FMT_YUV411P, \ - PIX_FMT_YUVA420P, \ - PIX_FMT_YUV420P, PIX_FMT_YUV422P, \ - PIX_FMT_YUV440P, PIX_FMT_YUV444P, \ - PIX_FMT_YUV420P9, PIX_FMT_YUV422P9, \ - PIX_FMT_YUV444P9, \ - PIX_FMT_YUV420P10, PIX_FMT_YUV422P10, \ - PIX_FMT_YUV444P10, \ - PIX_FMT_YUV420P16, PIX_FMT_YUV422P16, \ - PIX_FMT_YUV444P16 - -static const enum PixelFormat rgb_pix_fmts[] = {RGB_PIXEL_FORMATS}; -static const enum PixelFormat gray_pix_fmts[] = {GRAY_PIXEL_FORMATS}; -static const enum PixelFormat yuv_pix_fmts[] = {YUV_PIXEL_FORMATS}; -static const enum PixelFormat any_pix_fmts[] = {RGB_PIXEL_FORMATS, +#define RGB_PIXEL_FORMATS AV_PIX_FMT_RGB24, AV_PIX_FMT_RGBA, \ + AV_PIX_FMT_RGB48 + +#define GRAY_PIXEL_FORMATS AV_PIX_FMT_GRAY8, AV_PIX_FMT_Y400A, \ + AV_PIX_FMT_GRAY16 + +#define YUV_PIXEL_FORMATS AV_PIX_FMT_YUV410P, AV_PIX_FMT_YUV411P, \ + AV_PIX_FMT_YUVA420P, \ + AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, \ + AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUV444P, \ + AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, \ + AV_PIX_FMT_YUV444P9, \ + AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, \ + AV_PIX_FMT_YUV444P10, \ + AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, \ + AV_PIX_FMT_YUV444P16 + +static const enum AVPixelFormat rgb_pix_fmts[] = {RGB_PIXEL_FORMATS}; +static const enum AVPixelFormat gray_pix_fmts[] = {GRAY_PIXEL_FORMATS}; +static const enum AVPixelFormat yuv_pix_fmts[] = {YUV_PIXEL_FORMATS}; +static const enum AVPixelFormat any_pix_fmts[] = {RGB_PIXEL_FORMATS, GRAY_PIXEL_FORMATS, YUV_PIXEL_FORMATS}; @@ -73,7 +73,7 @@ typedef struct { } LibOpenJPEGContext; static int libopenjpeg_matches_pix_fmt(const opj_image_t *img, - enum PixelFormat pix_fmt) + enum AVPixelFormat pix_fmt) { AVPixFmtDescriptor des = av_pix_fmt_descriptors[pix_fmt]; int match = 1; @@ -110,10 +110,10 @@ static int libopenjpeg_matches_pix_fmt(const opj_image_t *img, return match; } -static enum PixelFormat libopenjpeg_guess_pix_fmt(const opj_image_t *image) +static enum AVPixelFormat libopenjpeg_guess_pix_fmt(const opj_image_t *image) { int index; - const enum PixelFormat *possible_fmts = NULL; + const enum AVPixelFormat *possible_fmts = NULL; int possible_fmts_nb = 0; switch (image->color_space) { @@ -141,14 +141,14 @@ static enum PixelFormat libopenjpeg_guess_pix_fmt(const opj_image_t *image) } } - return PIX_FMT_NONE; + return AV_PIX_FMT_NONE; } -static inline int libopenjpeg_ispacked(enum PixelFormat pix_fmt) +static inline int libopenjpeg_ispacked(enum AVPixelFormat pix_fmt) { int i, component_plane; - if (pix_fmt == PIX_FMT_GRAY16) + if (pix_fmt == AV_PIX_FMT_GRAY16) return 0; component_plane = av_pix_fmt_descriptors[pix_fmt].comp[0].plane; @@ -328,14 +328,14 @@ static int libopenjpeg_decode_frame(AVCodecContext *avctx, avcodec_set_dimensions(avctx, width, height); - if (avctx->pix_fmt != PIX_FMT_NONE) + if (avctx->pix_fmt != AV_PIX_FMT_NONE) if (!libopenjpeg_matches_pix_fmt(image, avctx->pix_fmt)) - avctx->pix_fmt = PIX_FMT_NONE; + avctx->pix_fmt = AV_PIX_FMT_NONE; - if (avctx->pix_fmt == PIX_FMT_NONE) + if (avctx->pix_fmt == AV_PIX_FMT_NONE) avctx->pix_fmt = libopenjpeg_guess_pix_fmt(image); - if (avctx->pix_fmt == PIX_FMT_NONE) { + if (avctx->pix_fmt == AV_PIX_FMT_NONE) { av_log(avctx, AV_LOG_ERROR, "Unable to determine pixel format\n"); ret = AVERROR_INVALIDDATA; goto done; diff --git a/libavcodec/libopenjpegenc.c b/libavcodec/libopenjpegenc.c index aad7ead0a2..972259abf6 100644 --- a/libavcodec/libopenjpegenc.c +++ b/libavcodec/libopenjpegenc.c @@ -87,32 +87,32 @@ static opj_image_t *libopenjpeg_create_image(AVCodecContext *avctx, sub_dy[2] = 1 << av_pix_fmt_descriptors[avctx->pix_fmt].log2_chroma_h; switch (avctx->pix_fmt) { - case PIX_FMT_GRAY8: - case PIX_FMT_GRAY16: - case PIX_FMT_Y400A: + case AV_PIX_FMT_GRAY8: + case AV_PIX_FMT_GRAY16: + case AV_PIX_FMT_Y400A: color_space = CLRSPC_GRAY; break; - case PIX_FMT_RGB24: - case PIX_FMT_RGBA: - case PIX_FMT_RGB48: + case AV_PIX_FMT_RGB24: + case AV_PIX_FMT_RGBA: + case AV_PIX_FMT_RGB48: color_space = CLRSPC_SRGB; break; - case PIX_FMT_YUV410P: - case PIX_FMT_YUV411P: - case PIX_FMT_YUV420P: - case PIX_FMT_YUV422P: - case PIX_FMT_YUV440P: - case PIX_FMT_YUV444P: - case PIX_FMT_YUVA420P: - case PIX_FMT_YUV420P9: - case PIX_FMT_YUV422P9: - case PIX_FMT_YUV444P9: - case PIX_FMT_YUV420P10: - case PIX_FMT_YUV422P10: - case PIX_FMT_YUV444P10: - case PIX_FMT_YUV420P16: - case PIX_FMT_YUV422P16: - case PIX_FMT_YUV444P16: + case AV_PIX_FMT_YUV410P: + case AV_PIX_FMT_YUV411P: + case AV_PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV440P: + case AV_PIX_FMT_YUV444P: + case AV_PIX_FMT_YUVA420P: + case AV_PIX_FMT_YUV420P9: + case AV_PIX_FMT_YUV422P9: + case AV_PIX_FMT_YUV444P9: + case AV_PIX_FMT_YUV420P10: + case AV_PIX_FMT_YUV422P10: + case AV_PIX_FMT_YUV444P10: + case AV_PIX_FMT_YUV420P16: + case AV_PIX_FMT_YUV422P16: + case AV_PIX_FMT_YUV444P16: color_space = CLRSPC_SYCC; break; default: @@ -302,34 +302,34 @@ static int libopenjpeg_encode_frame(AVCodecContext *avctx, AVPacket *pkt, image->y1 = (avctx->height - 1) * ctx->enc_params.subsampling_dy + 1; switch (avctx->pix_fmt) { - case PIX_FMT_RGB24: - case PIX_FMT_RGBA: - case PIX_FMT_Y400A: + case AV_PIX_FMT_RGB24: + case AV_PIX_FMT_RGBA: + case AV_PIX_FMT_Y400A: libopenjpeg_copy_packed8(avctx, frame, image); break; - case PIX_FMT_RGB48: + case AV_PIX_FMT_RGB48: libopenjpeg_copy_packed16(avctx, frame, image); break; - case PIX_FMT_GRAY8: - case PIX_FMT_YUV410P: - case PIX_FMT_YUV411P: - case PIX_FMT_YUV420P: - case PIX_FMT_YUV422P: - case PIX_FMT_YUV440P: - case PIX_FMT_YUV444P: - case PIX_FMT_YUVA420P: + case AV_PIX_FMT_GRAY8: + case AV_PIX_FMT_YUV410P: + case AV_PIX_FMT_YUV411P: + case AV_PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV440P: + case AV_PIX_FMT_YUV444P: + case AV_PIX_FMT_YUVA420P: libopenjpeg_copy_unpacked8(avctx, frame, image); break; - case PIX_FMT_GRAY16: - case PIX_FMT_YUV420P9: - case PIX_FMT_YUV422P9: - case PIX_FMT_YUV444P9: - case PIX_FMT_YUV444P10: - case PIX_FMT_YUV422P10: - case PIX_FMT_YUV420P10: - case PIX_FMT_YUV444P16: - case PIX_FMT_YUV422P16: - case PIX_FMT_YUV420P16: + case AV_PIX_FMT_GRAY16: + case AV_PIX_FMT_YUV420P9: + case AV_PIX_FMT_YUV422P9: + case AV_PIX_FMT_YUV444P9: + case AV_PIX_FMT_YUV444P10: + case AV_PIX_FMT_YUV422P10: + case AV_PIX_FMT_YUV420P10: + case AV_PIX_FMT_YUV444P16: + case AV_PIX_FMT_YUV422P16: + case AV_PIX_FMT_YUV420P16: libopenjpeg_copy_unpacked16(avctx, frame, image); break; default: @@ -421,16 +421,16 @@ AVCodec ff_libopenjpeg_encoder = { .encode2 = libopenjpeg_encode_frame, .close = libopenjpeg_encode_close, .capabilities = 0, - .pix_fmts = (const enum PixelFormat[]) { - PIX_FMT_RGB24, PIX_FMT_RGBA, PIX_FMT_RGB48, - PIX_FMT_GRAY8, PIX_FMT_GRAY16, PIX_FMT_Y400A, - PIX_FMT_YUV420P, PIX_FMT_YUV422P, PIX_FMT_YUVA420P, - PIX_FMT_YUV440P, PIX_FMT_YUV444P, - PIX_FMT_YUV411P, PIX_FMT_YUV410P, - PIX_FMT_YUV420P9, PIX_FMT_YUV422P9, PIX_FMT_YUV444P9, - PIX_FMT_YUV420P10, PIX_FMT_YUV422P10, PIX_FMT_YUV444P10, - PIX_FMT_YUV420P16, PIX_FMT_YUV422P16, PIX_FMT_YUV444P16, - PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]) { + AV_PIX_FMT_RGB24, AV_PIX_FMT_RGBA, AV_PIX_FMT_RGB48, + AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY16, AV_PIX_FMT_Y400A, + AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUVA420P, + AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUV444P, + AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P, + AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV444P9, + AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, + AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV444P16, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("OpenJPEG JPEG 2000"), .priv_class = &class, diff --git a/libavcodec/libschroedinger.h b/libavcodec/libschroedinger.h index f33014d766..bf179d2778 100644 --- a/libavcodec/libschroedinger.h +++ b/libavcodec/libschroedinger.h @@ -101,13 +101,13 @@ void *ff_schro_queue_pop(FFSchroQueue *queue); void ff_schro_queue_free(FFSchroQueue *queue, void (*free_func)(void *)); static const struct { - enum PixelFormat ff_pix_fmt; + enum AVPixelFormat ff_pix_fmt; SchroChromaFormat schro_pix_fmt; SchroFrameFormat schro_frame_fmt; } schro_pixel_format_map[] = { - { PIX_FMT_YUV420P, SCHRO_CHROMA_420, SCHRO_FRAME_FORMAT_U8_420 }, - { PIX_FMT_YUV422P, SCHRO_CHROMA_422, SCHRO_FRAME_FORMAT_U8_422 }, - { PIX_FMT_YUV444P, SCHRO_CHROMA_444, SCHRO_FRAME_FORMAT_U8_444 }, + { AV_PIX_FMT_YUV420P, SCHRO_CHROMA_420, SCHRO_FRAME_FORMAT_U8_420 }, + { AV_PIX_FMT_YUV422P, SCHRO_CHROMA_422, SCHRO_FRAME_FORMAT_U8_422 }, + { AV_PIX_FMT_YUV444P, SCHRO_CHROMA_444, SCHRO_FRAME_FORMAT_U8_444 }, }; /** diff --git a/libavcodec/libschroedingerdec.c b/libavcodec/libschroedingerdec.c index 110e8d0ae1..3cd9f9ace2 100644 --- a/libavcodec/libschroedingerdec.c +++ b/libavcodec/libschroedingerdec.c @@ -138,7 +138,7 @@ static SchroBuffer *find_next_parse_unit(SchroParseUnitContext *parse_ctx) /** * Returns Libav chroma format. */ -static enum PixelFormat get_chroma_format(SchroChromaFormat schro_pix_fmt) +static enum AVPixelFormat get_chroma_format(SchroChromaFormat schro_pix_fmt) { int num_formats = sizeof(schro_pixel_format_map) / sizeof(schro_pixel_format_map[0]); @@ -147,7 +147,7 @@ static enum PixelFormat get_chroma_format(SchroChromaFormat schro_pix_fmt) for (idx = 0; idx < num_formats; ++idx) if (schro_pixel_format_map[idx].schro_pix_fmt == schro_pix_fmt) return schro_pixel_format_map[idx].ff_pix_fmt; - return PIX_FMT_NONE; + return AV_PIX_FMT_NONE; } static av_cold int libschroedinger_decode_init(AVCodecContext *avccontext) diff --git a/libavcodec/libschroedingerenc.c b/libavcodec/libschroedingerenc.c index 865b093fa7..40022861be 100644 --- a/libavcodec/libschroedingerenc.c +++ b/libavcodec/libschroedingerenc.c @@ -448,8 +448,8 @@ AVCodec ff_libschroedinger_encoder = { .encode2 = libschroedinger_encode_frame, .close = libschroedinger_encode_close, .capabilities = CODEC_CAP_DELAY, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_YUV420P, PIX_FMT_YUV422P, PIX_FMT_YUV444P, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("libschroedinger Dirac 2.2"), }; diff --git a/libavcodec/libtheoraenc.c b/libavcodec/libtheoraenc.c index 3f861e1639..6c4eddf301 100644 --- a/libavcodec/libtheoraenc.c +++ b/libavcodec/libtheoraenc.c @@ -187,11 +187,11 @@ static av_cold int encode_init(AVCodecContext* avc_context) else t_info.colorspace = TH_CS_UNSPECIFIED; - if (avc_context->pix_fmt == PIX_FMT_YUV420P) + if (avc_context->pix_fmt == AV_PIX_FMT_YUV420P) t_info.pixel_fmt = TH_PF_420; - else if (avc_context->pix_fmt == PIX_FMT_YUV422P) + else if (avc_context->pix_fmt == AV_PIX_FMT_YUV422P) t_info.pixel_fmt = TH_PF_422; - else if (avc_context->pix_fmt == PIX_FMT_YUV444P) + else if (avc_context->pix_fmt == AV_PIX_FMT_YUV444P) t_info.pixel_fmt = TH_PF_444; else { av_log(avc_context, AV_LOG_ERROR, "Unsupported pix_fmt\n"); @@ -370,8 +370,8 @@ AVCodec ff_libtheora_encoder = { .close = encode_close, .encode2 = encode_frame, .capabilities = CODEC_CAP_DELAY, // needed to get the statsfile summary - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_YUV420P, PIX_FMT_YUV422P, PIX_FMT_YUV444P, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("libtheora Theora"), }; diff --git a/libavcodec/libvpxdec.c b/libavcodec/libvpxdec.c index 648187c319..56f706ed50 100644 --- a/libavcodec/libvpxdec.c +++ b/libavcodec/libvpxdec.c @@ -54,7 +54,7 @@ static av_cold int vp8_init(AVCodecContext *avctx) return AVERROR(EINVAL); } - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; return 0; } diff --git a/libavcodec/libvpxenc.c b/libavcodec/libvpxenc.c index 65f9d6db1f..f505d0ef43 100644 --- a/libavcodec/libvpxenc.c +++ b/libavcodec/libvpxenc.c @@ -580,7 +580,7 @@ AVCodec ff_libvpx_encoder = { .encode2 = vp8_encode, .close = vp8_free, .capabilities = CODEC_CAP_DELAY | CODEC_CAP_AUTO_THREADS, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("libvpx VP8"), .priv_class = &class, .defaults = defaults, diff --git a/libavcodec/libx264.c b/libavcodec/libx264.c index 6f68e8f8fe..f536f7f965 100644 --- a/libavcodec/libx264.c +++ b/libavcodec/libx264.c @@ -206,18 +206,18 @@ static av_cold int X264_close(AVCodecContext *avctx) return 0; } -static int convert_pix_fmt(enum PixelFormat pix_fmt) +static int convert_pix_fmt(enum AVPixelFormat pix_fmt) { switch (pix_fmt) { - case PIX_FMT_YUV420P: - case PIX_FMT_YUVJ420P: - case PIX_FMT_YUV420P9: - case PIX_FMT_YUV420P10: return X264_CSP_I420; - case PIX_FMT_YUV422P: - case PIX_FMT_YUV422P10: return X264_CSP_I422; - case PIX_FMT_YUV444P: - case PIX_FMT_YUV444P9: - case PIX_FMT_YUV444P10: return X264_CSP_I444; + case AV_PIX_FMT_YUV420P: + case AV_PIX_FMT_YUVJ420P: + case AV_PIX_FMT_YUV420P9: + case AV_PIX_FMT_YUV420P10: return X264_CSP_I420; + case AV_PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV422P10: return X264_CSP_I422; + case AV_PIX_FMT_YUV444P: + case AV_PIX_FMT_YUV444P9: + case AV_PIX_FMT_YUV444P10: return X264_CSP_I444; }; return 0; } @@ -402,7 +402,7 @@ static av_cold int X264_init(AVCodecContext *avctx) x4->params.i_slice_count = avctx->slices; - x4->params.vui.b_fullrange = avctx->pix_fmt == PIX_FMT_YUVJ420P; + x4->params.vui.b_fullrange = avctx->pix_fmt == AV_PIX_FMT_YUVJ420P; if (avctx->flags & CODEC_FLAG_GLOBAL_HEADER) x4->params.b_repeat_headers = 0; @@ -447,23 +447,23 @@ static av_cold int X264_init(AVCodecContext *avctx) return 0; } -static const enum PixelFormat pix_fmts_8bit[] = { - PIX_FMT_YUV420P, - PIX_FMT_YUVJ420P, - PIX_FMT_YUV422P, - PIX_FMT_YUV444P, - PIX_FMT_NONE +static const enum AVPixelFormat pix_fmts_8bit[] = { + AV_PIX_FMT_YUV420P, + AV_PIX_FMT_YUVJ420P, + AV_PIX_FMT_YUV422P, + AV_PIX_FMT_YUV444P, + AV_PIX_FMT_NONE }; -static const enum PixelFormat pix_fmts_9bit[] = { - PIX_FMT_YUV420P9, - PIX_FMT_YUV444P9, - PIX_FMT_NONE +static const enum AVPixelFormat pix_fmts_9bit[] = { + AV_PIX_FMT_YUV420P9, + AV_PIX_FMT_YUV444P9, + AV_PIX_FMT_NONE }; -static const enum PixelFormat pix_fmts_10bit[] = { - PIX_FMT_YUV420P10, - PIX_FMT_YUV422P10, - PIX_FMT_YUV444P10, - PIX_FMT_NONE +static const enum AVPixelFormat pix_fmts_10bit[] = { + AV_PIX_FMT_YUV420P10, + AV_PIX_FMT_YUV422P10, + AV_PIX_FMT_YUV444P10, + AV_PIX_FMT_NONE }; static av_cold void X264_init_static(AVCodec *codec) diff --git a/libavcodec/libxavs.c b/libavcodec/libxavs.c index ddfdeaff82..16026ac86b 100644 --- a/libavcodec/libxavs.c +++ b/libavcodec/libxavs.c @@ -427,7 +427,7 @@ AVCodec ff_libxavs_encoder = { .encode2 = XAVS_frame, .close = XAVS_close, .capabilities = CODEC_CAP_DELAY | CODEC_CAP_AUTO_THREADS, - .pix_fmts = (const enum PixelFormat[]) { PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("libxavs Chinese AVS (Audio Video Standard)"), .priv_class = &class, .defaults = xavs_defaults, diff --git a/libavcodec/libxvid.c b/libavcodec/libxvid.c index e43b52cdf7..89021e774f 100644 --- a/libavcodec/libxvid.c +++ b/libavcodec/libxvid.c @@ -657,7 +657,7 @@ static int xvid_encode_frame(AVCodecContext *avctx, AVPacket *pkt, xvid_enc_frame.length = pkt->size; /* Initialize input image fields */ - if( avctx->pix_fmt != PIX_FMT_YUV420P ) { + if( avctx->pix_fmt != AV_PIX_FMT_YUV420P ) { av_log(avctx, AV_LOG_ERROR, "Xvid: Color spaces other than 420p not supported\n"); return -1; } @@ -783,6 +783,6 @@ AVCodec ff_libxvid_encoder = { .init = xvid_encode_init, .encode2 = xvid_encode_frame, .close = xvid_encode_close, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("libxvidcore MPEG-4 part 2"), }; diff --git a/libavcodec/ljpegenc.c b/libavcodec/ljpegenc.c index 331ec5ccb6..78ba2c962e 100644 --- a/libavcodec/ljpegenc.c +++ b/libavcodec/ljpegenc.c @@ -51,7 +51,7 @@ static int encode_picture_lossless(AVCodecContext *avctx, AVPacket *pkt, const int mb_height = (height + s->mjpeg_vsample[0] - 1) / s->mjpeg_vsample[0]; int ret, max_pkt_size = FF_MIN_BUFFER_SIZE; - if (avctx->pix_fmt == PIX_FMT_BGRA) + if (avctx->pix_fmt == AV_PIX_FMT_BGRA) max_pkt_size += width * height * 3 * 4; else { max_pkt_size += mb_width * mb_height * 3 * 4 @@ -72,7 +72,7 @@ static int encode_picture_lossless(AVCodecContext *avctx, AVPacket *pkt, s->header_bits= put_bits_count(&s->pb); - if(avctx->pix_fmt == PIX_FMT_BGRA){ + if(avctx->pix_fmt == AV_PIX_FMT_BGRA){ int x, y, i; const int linesize= p->linesize[0]; uint16_t (*buffer)[4]= (void *) s->rd_scratchpad; diff --git a/libavcodec/loco.c b/libavcodec/loco.c index f8e41684d0..5d0ce05113 100644 --- a/libavcodec/loco.c +++ b/libavcodec/loco.c @@ -254,16 +254,16 @@ static av_cold int decode_init(AVCodecContext *avctx){ l->mode = AV_RL32(avctx->extradata + 4); switch(l->mode) { case LOCO_CYUY2: case LOCO_YUY2: case LOCO_UYVY: - avctx->pix_fmt = PIX_FMT_YUV422P; + avctx->pix_fmt = AV_PIX_FMT_YUV422P; break; case LOCO_CRGB: case LOCO_RGB: - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; break; case LOCO_CYV12: case LOCO_YV12: - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; break; case LOCO_CRGBA: case LOCO_RGBA: - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; break; default: av_log(avctx, AV_LOG_INFO, "Unknown colorspace, index = %i\n", l->mode); diff --git a/libavcodec/mdec.c b/libavcodec/mdec.c index 1b90ec984b..a07d6e564b 100644 --- a/libavcodec/mdec.c +++ b/libavcodec/mdec.c @@ -235,7 +235,7 @@ static av_cold int decode_init(AVCodecContext *avctx){ avctx->idct_algo = FF_IDCT_SIMPLE; p->qstride= 0; p->qscale_table= av_mallocz(a->mb_width); - avctx->pix_fmt= PIX_FMT_YUVJ420P; + avctx->pix_fmt= AV_PIX_FMT_YUVJ420P; return 0; } diff --git a/libavcodec/mimic.c b/libavcodec/mimic.c index 5aa6a7b3bb..fd7efef343 100644 --- a/libavcodec/mimic.c +++ b/libavcodec/mimic.c @@ -340,7 +340,7 @@ static int mimic_decode_frame(AVCodecContext *avctx, void *data, ctx->avctx = avctx; avctx->width = width; avctx->height = height; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; for(i = 0; i < 3; i++) { ctx->num_vblocks[i] = -((-height) >> (3 + !!i)); ctx->num_hblocks[i] = width >> (3 + !!i) ; diff --git a/libavcodec/mjpegdec.c b/libavcodec/mjpegdec.c index 1bcbcfd1d4..afe1f3db65 100644 --- a/libavcodec/mjpegdec.c +++ b/libavcodec/mjpegdec.c @@ -321,22 +321,22 @@ int ff_mjpeg_decode_sof(MJpegDecodeContext *s) switch (pix_fmt_id) { case 0x11111100: if (s->rgb) - s->avctx->pix_fmt = PIX_FMT_BGRA; + s->avctx->pix_fmt = AV_PIX_FMT_BGRA; else - s->avctx->pix_fmt = s->cs_itu601 ? PIX_FMT_YUV444P : PIX_FMT_YUVJ444P; + s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV444P : AV_PIX_FMT_YUVJ444P; assert(s->nb_components == 3); break; case 0x11000000: - s->avctx->pix_fmt = PIX_FMT_GRAY8; + s->avctx->pix_fmt = AV_PIX_FMT_GRAY8; break; case 0x12111100: - s->avctx->pix_fmt = s->cs_itu601 ? PIX_FMT_YUV440P : PIX_FMT_YUVJ440P; + s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV440P : AV_PIX_FMT_YUVJ440P; break; case 0x21111100: - s->avctx->pix_fmt = s->cs_itu601 ? PIX_FMT_YUV422P : PIX_FMT_YUVJ422P; + s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV422P : AV_PIX_FMT_YUVJ422P; break; case 0x22111100: - s->avctx->pix_fmt = s->cs_itu601 ? PIX_FMT_YUV420P : PIX_FMT_YUVJ420P; + s->avctx->pix_fmt = s->cs_itu601 ? AV_PIX_FMT_YUV420P : AV_PIX_FMT_YUVJ420P; break; default: av_log(s->avctx, AV_LOG_ERROR, "Unhandled pixel format 0x%x\n", pix_fmt_id); @@ -344,11 +344,11 @@ int ff_mjpeg_decode_sof(MJpegDecodeContext *s) } if (s->ls) { if (s->nb_components > 1) - s->avctx->pix_fmt = PIX_FMT_RGB24; + s->avctx->pix_fmt = AV_PIX_FMT_RGB24; else if (s->bits <= 8) - s->avctx->pix_fmt = PIX_FMT_GRAY8; + s->avctx->pix_fmt = AV_PIX_FMT_GRAY8; else - s->avctx->pix_fmt = PIX_FMT_GRAY16; + s->avctx->pix_fmt = AV_PIX_FMT_GRAY16; } if (s->picture_ptr->data[0]) diff --git a/libavcodec/mjpegenc.c b/libavcodec/mjpegenc.c index ebe899326e..64e6f25fc6 100644 --- a/libavcodec/mjpegenc.c +++ b/libavcodec/mjpegenc.c @@ -181,9 +181,9 @@ static void jpeg_put_comments(MpegEncContext *s) AV_WB16(ptr, size); } - if( s->avctx->pix_fmt == PIX_FMT_YUV420P - ||s->avctx->pix_fmt == PIX_FMT_YUV422P - ||s->avctx->pix_fmt == PIX_FMT_YUV444P){ + if( s->avctx->pix_fmt == AV_PIX_FMT_YUV420P + ||s->avctx->pix_fmt == AV_PIX_FMT_YUV422P + ||s->avctx->pix_fmt == AV_PIX_FMT_YUV444P){ put_marker(p, COM); flush_put_bits(p); ptr = put_bits_ptr(p); @@ -211,7 +211,7 @@ void ff_mjpeg_encode_picture_header(MpegEncContext *s) } put_bits(&s->pb, 16, 17); - if(lossless && s->avctx->pix_fmt == PIX_FMT_BGRA) + if(lossless && s->avctx->pix_fmt == AV_PIX_FMT_BGRA) put_bits(&s->pb, 8, 9); /* 9 bits/component RCT */ else put_bits(&s->pb, 8, 8); /* 8 bits/component */ @@ -452,8 +452,8 @@ AVCodec ff_mjpeg_encoder = { .init = ff_MPV_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_YUVJ420P, PIX_FMT_YUVJ422P, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("MJPEG (Motion JPEG)"), }; diff --git a/libavcodec/mmvideo.c b/libavcodec/mmvideo.c index 2187d7f585..9959d5fe03 100644 --- a/libavcodec/mmvideo.c +++ b/libavcodec/mmvideo.c @@ -58,7 +58,7 @@ static av_cold int mm_decode_init(AVCodecContext *avctx) s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; s->frame.reference = 1; diff --git a/libavcodec/motionpixels.c b/libavcodec/motionpixels.c index 4a1c104a59..a33d8d3eb0 100644 --- a/libavcodec/motionpixels.c +++ b/libavcodec/motionpixels.c @@ -62,7 +62,7 @@ static av_cold int mp_decode_init(AVCodecContext *avctx) mp->offset_bits_len = av_log2(avctx->width * avctx->height) + 1; mp->vpt = av_mallocz(avctx->height * sizeof(YuvPixel)); mp->hpt = av_mallocz(h4 * w4 / 16 * sizeof(YuvPixel)); - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; return 0; } diff --git a/libavcodec/mpeg12.c b/libavcodec/mpeg12.c index fd8d71f1b4..a5dcccf3dc 100644 --- a/libavcodec/mpeg12.c +++ b/libavcodec/mpeg12.c @@ -1171,12 +1171,12 @@ static void quant_matrix_rebuild(uint16_t *matrix, const uint8_t *old_perm, } } -static const enum PixelFormat pixfmt_xvmc_mpg2_420[] = { - PIX_FMT_XVMC_MPEG2_IDCT, - PIX_FMT_XVMC_MPEG2_MC, - PIX_FMT_NONE }; +static const enum AVPixelFormat pixfmt_xvmc_mpg2_420[] = { + AV_PIX_FMT_XVMC_MPEG2_IDCT, + AV_PIX_FMT_XVMC_MPEG2_MC, + AV_PIX_FMT_NONE }; -static enum PixelFormat mpeg_get_pixelformat(AVCodecContext *avctx) +static enum AVPixelFormat mpeg_get_pixelformat(AVCodecContext *avctx) { Mpeg1Context *s1 = avctx->priv_data; MpegEncContext *s = &s1->mpeg_enc_ctx; @@ -1185,16 +1185,16 @@ static enum PixelFormat mpeg_get_pixelformat(AVCodecContext *avctx) return avctx->get_format(avctx, pixfmt_xvmc_mpg2_420); else if (avctx->codec->capabilities & CODEC_CAP_HWACCEL_VDPAU) { if (avctx->codec_id == AV_CODEC_ID_MPEG1VIDEO) - return PIX_FMT_VDPAU_MPEG1; + return AV_PIX_FMT_VDPAU_MPEG1; else - return PIX_FMT_VDPAU_MPEG2; + return AV_PIX_FMT_VDPAU_MPEG2; } else { if (s->chroma_format < 2) return avctx->get_format(avctx, ff_hwaccel_pixfmt_list_420); else if (s->chroma_format == 2) - return PIX_FMT_YUV422P; + return AV_PIX_FMT_YUV422P; else - return PIX_FMT_YUV444P; + return AV_PIX_FMT_YUV444P; } } @@ -1289,7 +1289,7 @@ static int mpeg_decode_postinit(AVCodecContext *avctx) avctx->pix_fmt = mpeg_get_pixelformat(avctx); avctx->hwaccel = ff_find_hwaccel(avctx->codec->id, avctx->pix_fmt); // until then pix_fmt may be changed right after codec init - if (avctx->pix_fmt == PIX_FMT_XVMC_MPEG2_IDCT || + if (avctx->pix_fmt == AV_PIX_FMT_XVMC_MPEG2_IDCT || avctx->hwaccel || s->avctx->codec->capabilities & CODEC_CAP_HWACCEL_VDPAU) if (avctx->idct_algo == FF_IDCT_AUTO) @@ -2036,7 +2036,7 @@ static int vcr2_init_sequence(AVCodecContext *avctx) avctx->pix_fmt = mpeg_get_pixelformat(avctx); avctx->hwaccel = ff_find_hwaccel(avctx->codec->id, avctx->pix_fmt); - if (avctx->pix_fmt == PIX_FMT_XVMC_MPEG2_IDCT || avctx->hwaccel || + if (avctx->pix_fmt == AV_PIX_FMT_XVMC_MPEG2_IDCT || avctx->hwaccel || s->avctx->codec->capabilities & CODEC_CAP_HWACCEL_VDPAU) if (avctx->idct_algo == FF_IDCT_AUTO) avctx->idct_algo = FF_IDCT_SIMPLE; @@ -2539,7 +2539,7 @@ static av_cold int mpeg_mc_decode_init(AVCodecContext *avctx) } mpeg_decode_init(avctx); - avctx->pix_fmt = PIX_FMT_XVMC_MPEG2_IDCT; + avctx->pix_fmt = AV_PIX_FMT_XVMC_MPEG2_IDCT; avctx->xvmc_acceleration = 2; // 2 - the blocks are packed! return 0; diff --git a/libavcodec/mpeg12enc.c b/libavcodec/mpeg12enc.c index 019ab5b1b7..7048b5dc97 100644 --- a/libavcodec/mpeg12enc.c +++ b/libavcodec/mpeg12enc.c @@ -959,8 +959,8 @@ AVCodec ff_mpeg1video_encoder = { .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, .supported_framerates = avpriv_frame_rate_tab+1, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, - PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, + AV_PIX_FMT_NONE }, .capabilities = CODEC_CAP_DELAY | CODEC_CAP_SLICE_THREADS, .long_name = NULL_IF_CONFIG_SMALL("MPEG-1 video"), .priv_class = &mpeg1_class, @@ -975,8 +975,8 @@ AVCodec ff_mpeg2video_encoder = { .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, .supported_framerates = avpriv_frame_rate_tab + 1, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_YUV420P, PIX_FMT_YUV422P, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_NONE }, .capabilities = CODEC_CAP_DELAY | CODEC_CAP_SLICE_THREADS, .long_name = NULL_IF_CONFIG_SMALL("MPEG-2 video"), diff --git a/libavcodec/mpeg4videodec.c b/libavcodec/mpeg4videodec.c index 4a4998cdaa..faa98668d9 100644 --- a/libavcodec/mpeg4videodec.c +++ b/libavcodec/mpeg4videodec.c @@ -2288,7 +2288,7 @@ AVCodec ff_mpeg4_vdpau_decoder = { .capabilities = CODEC_CAP_DR1 | CODEC_CAP_TRUNCATED | CODEC_CAP_DELAY | CODEC_CAP_HWACCEL_VDPAU, .long_name = NULL_IF_CONFIG_SMALL("MPEG-4 part 2 (VDPAU)"), - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_VDPAU_MPEG4, - PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_VDPAU_MPEG4, + AV_PIX_FMT_NONE }, }; #endif diff --git a/libavcodec/mpeg4videoenc.c b/libavcodec/mpeg4videoenc.c index 42d1ba8e2f..b145eb229b 100644 --- a/libavcodec/mpeg4videoenc.c +++ b/libavcodec/mpeg4videoenc.c @@ -1339,7 +1339,7 @@ AVCodec ff_mpeg4_encoder = { .init = encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .capabilities = CODEC_CAP_DELAY | CODEC_CAP_SLICE_THREADS, .long_name = NULL_IF_CONFIG_SMALL("MPEG-4 part 2"), .priv_class = &mpeg4enc_class, diff --git a/libavcodec/mpegvideo.c b/libavcodec/mpegvideo.c index ed40c10583..f739c2dc5a 100644 --- a/libavcodec/mpegvideo.c +++ b/libavcodec/mpegvideo.c @@ -125,17 +125,17 @@ const uint8_t *const ff_mpeg2_dc_scale_table[4] = { mpeg2_dc_scale_table3, }; -const enum PixelFormat ff_pixfmt_list_420[] = { - PIX_FMT_YUV420P, - PIX_FMT_NONE +const enum AVPixelFormat ff_pixfmt_list_420[] = { + AV_PIX_FMT_YUV420P, + AV_PIX_FMT_NONE }; -const enum PixelFormat ff_hwaccel_pixfmt_list_420[] = { - PIX_FMT_DXVA2_VLD, - PIX_FMT_VAAPI_VLD, - PIX_FMT_VDA_VLD, - PIX_FMT_YUV420P, - PIX_FMT_NONE +const enum AVPixelFormat ff_hwaccel_pixfmt_list_420[] = { + AV_PIX_FMT_DXVA2_VLD, + AV_PIX_FMT_VAAPI_VLD, + AV_PIX_FMT_VDA_VLD, + AV_PIX_FMT_YUV420P, + AV_PIX_FMT_NONE }; const uint8_t *avpriv_mpv_find_start_code(const uint8_t *restrict p, @@ -841,9 +841,9 @@ av_cold int ff_MPV_common_init(MpegEncContext *s) else if (s->codec_id != AV_CODEC_ID_H264) s->mb_height = (s->height + 15) / 16; - if (s->avctx->pix_fmt == PIX_FMT_NONE) { + if (s->avctx->pix_fmt == AV_PIX_FMT_NONE) { av_log(s->avctx, AV_LOG_ERROR, - "decoding to PIX_FMT_NONE is not supported.\n"); + "decoding to AV_PIX_FMT_NONE is not supported.\n"); return -1; } diff --git a/libavcodec/mpegvideo.h b/libavcodec/mpegvideo.h index 4c220ecb0a..23637d0541 100644 --- a/libavcodec/mpegvideo.h +++ b/libavcodec/mpegvideo.h @@ -809,8 +809,8 @@ void ff_MPV_motion(MpegEncContext *s, */ int ff_alloc_picture(MpegEncContext *s, Picture *pic, int shared); -extern const enum PixelFormat ff_pixfmt_list_420[]; -extern const enum PixelFormat ff_hwaccel_pixfmt_list_420[]; +extern const enum AVPixelFormat ff_pixfmt_list_420[]; +extern const enum AVPixelFormat ff_hwaccel_pixfmt_list_420[]; static inline void ff_update_block_index(MpegEncContext *s){ const int block_size = 8; diff --git a/libavcodec/mpegvideo_enc.c b/libavcodec/mpegvideo_enc.c index 8325ee0f74..59476a023f 100644 --- a/libavcodec/mpegvideo_enc.c +++ b/libavcodec/mpegvideo_enc.c @@ -278,50 +278,50 @@ av_cold int ff_MPV_encode_init(AVCodecContext *avctx) switch (avctx->codec_id) { case AV_CODEC_ID_MPEG2VIDEO: - if (avctx->pix_fmt != PIX_FMT_YUV420P && - avctx->pix_fmt != PIX_FMT_YUV422P) { + if (avctx->pix_fmt != AV_PIX_FMT_YUV420P && + avctx->pix_fmt != AV_PIX_FMT_YUV422P) { av_log(avctx, AV_LOG_ERROR, "only YUV420 and YUV422 are supported\n"); return -1; } break; case AV_CODEC_ID_LJPEG: - if (avctx->pix_fmt != PIX_FMT_YUVJ420P && - avctx->pix_fmt != PIX_FMT_YUVJ422P && - avctx->pix_fmt != PIX_FMT_YUVJ444P && - avctx->pix_fmt != PIX_FMT_BGRA && - ((avctx->pix_fmt != PIX_FMT_YUV420P && - avctx->pix_fmt != PIX_FMT_YUV422P && - avctx->pix_fmt != PIX_FMT_YUV444P) || + if (avctx->pix_fmt != AV_PIX_FMT_YUVJ420P && + avctx->pix_fmt != AV_PIX_FMT_YUVJ422P && + avctx->pix_fmt != AV_PIX_FMT_YUVJ444P && + avctx->pix_fmt != AV_PIX_FMT_BGRA && + ((avctx->pix_fmt != AV_PIX_FMT_YUV420P && + avctx->pix_fmt != AV_PIX_FMT_YUV422P && + avctx->pix_fmt != AV_PIX_FMT_YUV444P) || avctx->strict_std_compliance > FF_COMPLIANCE_UNOFFICIAL)) { av_log(avctx, AV_LOG_ERROR, "colorspace not supported in LJPEG\n"); return -1; } break; case AV_CODEC_ID_MJPEG: - if (avctx->pix_fmt != PIX_FMT_YUVJ420P && - avctx->pix_fmt != PIX_FMT_YUVJ422P && - ((avctx->pix_fmt != PIX_FMT_YUV420P && - avctx->pix_fmt != PIX_FMT_YUV422P) || + if (avctx->pix_fmt != AV_PIX_FMT_YUVJ420P && + avctx->pix_fmt != AV_PIX_FMT_YUVJ422P && + ((avctx->pix_fmt != AV_PIX_FMT_YUV420P && + avctx->pix_fmt != AV_PIX_FMT_YUV422P) || avctx->strict_std_compliance > FF_COMPLIANCE_UNOFFICIAL)) { av_log(avctx, AV_LOG_ERROR, "colorspace not supported in jpeg\n"); return -1; } break; default: - if (avctx->pix_fmt != PIX_FMT_YUV420P) { + if (avctx->pix_fmt != AV_PIX_FMT_YUV420P) { av_log(avctx, AV_LOG_ERROR, "only YUV420 is supported\n"); return -1; } } switch (avctx->pix_fmt) { - case PIX_FMT_YUVJ422P: - case PIX_FMT_YUV422P: + case AV_PIX_FMT_YUVJ422P: + case AV_PIX_FMT_YUV422P: s->chroma_format = CHROMA_422; break; - case PIX_FMT_YUVJ420P: - case PIX_FMT_YUV420P: + case AV_PIX_FMT_YUVJ420P: + case AV_PIX_FMT_YUV420P: default: s->chroma_format = CHROMA_420; break; @@ -640,7 +640,7 @@ av_cold int ff_MPV_encode_init(AVCodecContext *avctx) s->out_format = FMT_MJPEG; s->intra_only = 1; /* force intra only for jpeg */ if (avctx->codec->id == AV_CODEC_ID_LJPEG && - avctx->pix_fmt == PIX_FMT_BGRA) { + avctx->pix_fmt == AV_PIX_FMT_BGRA) { s->mjpeg_vsample[0] = s->mjpeg_hsample[0] = s->mjpeg_vsample[1] = s->mjpeg_hsample[1] = s->mjpeg_vsample[2] = s->mjpeg_hsample[2] = 1; @@ -1117,7 +1117,7 @@ static int estimate_best_b_count(MpegEncContext *s) c->me_cmp = s->avctx->me_cmp; c->mb_cmp = s->avctx->mb_cmp; c->me_sub_cmp = s->avctx->me_sub_cmp; - c->pix_fmt = PIX_FMT_YUV420P; + c->pix_fmt = AV_PIX_FMT_YUV420P; c->time_base = s->avctx->time_base; c->max_b_frames = s->max_b_frames; @@ -4138,7 +4138,7 @@ AVCodec ff_h263_encoder = { .init = ff_MPV_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts= (const enum PixelFormat[]){PIX_FMT_YUV420P, PIX_FMT_NONE}, + .pix_fmts= (const enum AVPixelFormat[]){AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE}, .long_name= NULL_IF_CONFIG_SMALL("H.263 / H.263-1996"), .priv_class = &h263_class, }; @@ -4167,7 +4167,7 @@ AVCodec ff_h263p_encoder = { .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, .capabilities = CODEC_CAP_SLICE_THREADS, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("H.263+ / H.263-1998 / H.263 version 2"), .priv_class = &h263p_class, }; @@ -4182,7 +4182,7 @@ AVCodec ff_msmpeg4v2_encoder = { .init = ff_MPV_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("MPEG-4 part 2 Microsoft variant version 2"), .priv_class = &msmpeg4v2_class, }; @@ -4197,7 +4197,7 @@ AVCodec ff_msmpeg4v3_encoder = { .init = ff_MPV_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("MPEG-4 part 2 Microsoft variant version 3"), .priv_class = &msmpeg4v3_class, }; @@ -4212,7 +4212,7 @@ AVCodec ff_wmv1_encoder = { .init = ff_MPV_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Windows Media Video 7"), .priv_class = &wmv1_class, }; diff --git a/libavcodec/msrle.c b/libavcodec/msrle.c index c938ed9358..06305bf276 100644 --- a/libavcodec/msrle.c +++ b/libavcodec/msrle.c @@ -56,10 +56,10 @@ static av_cold int msrle_decode_init(AVCodecContext *avctx) switch (avctx->bits_per_coded_sample) { case 4: case 8: - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; break; case 24: - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; break; default: av_log(avctx, AV_LOG_ERROR, "unsupported bits per sample\n"); diff --git a/libavcodec/mss1.c b/libavcodec/mss1.c index 1591bba535..f72385f9fe 100644 --- a/libavcodec/mss1.c +++ b/libavcodec/mss1.c @@ -197,7 +197,7 @@ static av_cold int mss1_decode_init(AVCodecContext *avctx) ret = ff_mss12_decode_init(&c->ctx, 0, &c->sc, NULL); - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return ret; } diff --git a/libavcodec/mss2.c b/libavcodec/mss2.c index fbdc72c5c7..597ebb646e 100644 --- a/libavcodec/mss2.c +++ b/libavcodec/mss2.c @@ -396,11 +396,11 @@ static int decode_wmv9(AVCodecContext *avctx, const uint8_t *buf, int buf_size, return AVERROR_INVALIDDATA; } - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; if (ff_MPV_frame_start(s, avctx) < 0) { av_log(v->s.avctx, AV_LOG_ERROR, "ff_MPV_frame_start error\n"); - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; return -1; } @@ -448,7 +448,7 @@ static int decode_wmv9(AVCodecContext *avctx, const uint8_t *buf, int buf_size, f->data[1], f->data[2], f->linesize[1], w, h); - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; return 0; } @@ -519,7 +519,7 @@ static int mss2_decode_frame(AVCodecContext *avctx, void *data, int *data_size, if (is_555 && (has_wmv9 || has_mv || c->slice_split && ctx->split_position)) return AVERROR_INVALIDDATA; - avctx->pix_fmt = is_555 ? PIX_FMT_RGB555 : PIX_FMT_RGB24; + avctx->pix_fmt = is_555 ? AV_PIX_FMT_RGB555 : AV_PIX_FMT_RGB24; if (ctx->pic.data[0] && ctx->pic.format != avctx->pix_fmt) avctx->release_buffer(avctx, &ctx->pic); @@ -848,8 +848,8 @@ static av_cold int mss2_decode_init(AVCodecContext *avctx) } ff_mss2dsp_init(&ctx->dsp); - avctx->pix_fmt = c->free_colours == 127 ? PIX_FMT_RGB555 - : PIX_FMT_RGB24; + avctx->pix_fmt = c->free_colours == 127 ? AV_PIX_FMT_RGB555 + : AV_PIX_FMT_RGB24; return 0; } diff --git a/libavcodec/mss3.c b/libavcodec/mss3.c index 327849f92b..a41db22249 100644 --- a/libavcodec/mss3.c +++ b/libavcodec/mss3.c @@ -835,7 +835,7 @@ static av_cold int mss3_decode_init(AVCodecContext *avctx) } } - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; avctx->coded_frame = &c->pic; init_coders(c); diff --git a/libavcodec/mss4.c b/libavcodec/mss4.c index 3e974e71f5..1c850bc1c3 100644 --- a/libavcodec/mss4.c +++ b/libavcodec/mss4.c @@ -649,7 +649,7 @@ static av_cold int mss4_decode_init(AVCodecContext *avctx) } } - avctx->pix_fmt = PIX_FMT_YUV444P; + avctx->pix_fmt = AV_PIX_FMT_YUV444P; avctx->coded_frame = &c->pic; return 0; diff --git a/libavcodec/msvideo1.c b/libavcodec/msvideo1.c index 0d772b9f44..644fecf66b 100644 --- a/libavcodec/msvideo1.c +++ b/libavcodec/msvideo1.c @@ -65,10 +65,10 @@ static av_cold int msvideo1_decode_init(AVCodecContext *avctx) /* figure out the colorspace based on the presence of a palette */ if (s->avctx->bits_per_coded_sample == 8) { s->mode_8bit = 1; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; } else { s->mode_8bit = 0; - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; } s->frame.data[0] = NULL; @@ -172,7 +172,7 @@ static void msvideo1_decode_8bit(Msvideo1Context *s) } /* make the palette available on the way out */ - if (s->avctx->pix_fmt == PIX_FMT_PAL8) + if (s->avctx->pix_fmt == AV_PIX_FMT_PAL8) memcpy(s->frame.data[1], s->pal, AVPALETTE_SIZE); } diff --git a/libavcodec/nuv.c b/libavcodec/nuv.c index b74ee31ca1..4beed8f918 100644 --- a/libavcodec/nuv.c +++ b/libavcodec/nuv.c @@ -74,8 +74,8 @@ static const uint8_t fallback_cquant[] = { static void copy_frame(AVFrame *f, const uint8_t *src, int width, int height) { AVPicture pic; - avpicture_fill(&pic, src, PIX_FMT_YUV420P, width, height); - av_picture_copy((AVPicture *)f, &pic, PIX_FMT_YUV420P, width, height); + avpicture_fill(&pic, src, AV_PIX_FMT_YUV420P, width, height); + av_picture_copy((AVPicture *)f, &pic, AV_PIX_FMT_YUV420P, width, height); } /** @@ -269,7 +269,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, static av_cold int decode_init(AVCodecContext *avctx) { NuvContext *c = avctx->priv_data; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; c->pic.data[0] = NULL; c->decomp_buf = NULL; c->quality = -1; diff --git a/libavcodec/options.c b/libavcodec/options.c index 25018b9b97..fc2a184221 100644 --- a/libavcodec/options.c +++ b/libavcodec/options.c @@ -94,7 +94,7 @@ int avcodec_get_context_defaults3(AVCodecContext *s, const AVCodec *codec) s->execute = avcodec_default_execute; s->execute2 = avcodec_default_execute2; s->sample_aspect_ratio = (AVRational){0,1}; - s->pix_fmt = PIX_FMT_NONE; + s->pix_fmt = AV_PIX_FMT_NONE; s->sample_fmt = AV_SAMPLE_FMT_NONE; s->reget_buffer = avcodec_default_reget_buffer; diff --git a/libavcodec/pamenc.c b/libavcodec/pamenc.c index 646954978d..dba47717e2 100644 --- a/libavcodec/pamenc.c +++ b/libavcodec/pamenc.c @@ -52,25 +52,25 @@ static int pam_encode_frame(AVCodecContext *avctx, AVPacket *pkt, h = avctx->height; w = avctx->width; switch (avctx->pix_fmt) { - case PIX_FMT_MONOWHITE: + case AV_PIX_FMT_MONOWHITE: n = (w + 7) >> 3; depth = 1; maxval = 1; tuple_type = "BLACKANDWHITE"; break; - case PIX_FMT_GRAY8: + case AV_PIX_FMT_GRAY8: n = w; depth = 1; maxval = 255; tuple_type = "GRAYSCALE"; break; - case PIX_FMT_RGB24: + case AV_PIX_FMT_RGB24: n = w * 3; depth = 3; maxval = 255; tuple_type = "RGB"; break; - case PIX_FMT_RGB32: + case AV_PIX_FMT_RGB32: n = w * 4; depth = 4; maxval = 255; @@ -87,7 +87,7 @@ static int pam_encode_frame(AVCodecContext *avctx, AVPacket *pkt, ptr = p->data[0]; linesize = p->linesize[0]; - if (avctx->pix_fmt == PIX_FMT_RGB32) { + if (avctx->pix_fmt == AV_PIX_FMT_RGB32) { int j; unsigned int v; @@ -121,9 +121,9 @@ AVCodec ff_pam_encoder = { .priv_data_size = sizeof(PNMContext), .init = ff_pnm_init, .encode2 = pam_encode_frame, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_RGB24, PIX_FMT_RGB32, PIX_FMT_GRAY8, PIX_FMT_MONOWHITE, - PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_RGB24, AV_PIX_FMT_RGB32, AV_PIX_FMT_GRAY8, AV_PIX_FMT_MONOWHITE, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("PAM (Portable AnyMap) image"), }; diff --git a/libavcodec/pcx.c b/libavcodec/pcx.c index 0a18acd486..45e751b389 100644 --- a/libavcodec/pcx.c +++ b/libavcodec/pcx.c @@ -122,7 +122,7 @@ static int pcx_decode_frame(AVCodecContext *avctx, void *data, int *data_size, switch ((nplanes<<8) + bits_per_pixel) { case 0x0308: - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; break; case 0x0108: case 0x0104: @@ -131,7 +131,7 @@ static int pcx_decode_frame(AVCodecContext *avctx, void *data, int *data_size, case 0x0401: case 0x0301: case 0x0201: - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; break; default: av_log(avctx, AV_LOG_ERROR, "invalid PCX file\n"); diff --git a/libavcodec/pcxenc.c b/libavcodec/pcxenc.c index 7fc0408b5c..e0ee20ab74 100644 --- a/libavcodec/pcxenc.c +++ b/libavcodec/pcxenc.c @@ -118,21 +118,21 @@ static int pcx_encode_frame(AVCodecContext *avctx, AVPacket *pkt, } switch (avctx->pix_fmt) { - case PIX_FMT_RGB24: + case AV_PIX_FMT_RGB24: bpp = 8; nplanes = 3; break; - case PIX_FMT_RGB8: - case PIX_FMT_BGR8: - case PIX_FMT_RGB4_BYTE: - case PIX_FMT_BGR4_BYTE: - case PIX_FMT_GRAY8: - case PIX_FMT_PAL8: + case AV_PIX_FMT_RGB8: + case AV_PIX_FMT_BGR8: + case AV_PIX_FMT_RGB4_BYTE: + case AV_PIX_FMT_BGR4_BYTE: + case AV_PIX_FMT_GRAY8: + case AV_PIX_FMT_PAL8: bpp = 8; nplanes = 1; pal = (uint32_t *)pict->data[1]; break; - case PIX_FMT_MONOBLACK: + case AV_PIX_FMT_MONOBLACK: bpp = 1; nplanes = 1; pal = monoblack_pal; @@ -209,12 +209,12 @@ AVCodec ff_pcx_encoder = { .priv_data_size = sizeof(PCXContext), .init = pcx_encode_init, .encode2 = pcx_encode_frame, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_RGB24, - PIX_FMT_RGB8, PIX_FMT_BGR8, PIX_FMT_RGB4_BYTE, PIX_FMT_BGR4_BYTE, - PIX_FMT_GRAY8, PIX_FMT_PAL8, - PIX_FMT_MONOBLACK, - PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_RGB24, + AV_PIX_FMT_RGB8, AV_PIX_FMT_BGR8, AV_PIX_FMT_RGB4_BYTE, AV_PIX_FMT_BGR4_BYTE, + AV_PIX_FMT_GRAY8, AV_PIX_FMT_PAL8, + AV_PIX_FMT_MONOBLACK, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("PC Paintbrush PCX image"), }; diff --git a/libavcodec/pgssubdec.c b/libavcodec/pgssubdec.c index 49ef325eca..9fd26d80aa 100644 --- a/libavcodec/pgssubdec.c +++ b/libavcodec/pgssubdec.c @@ -63,7 +63,7 @@ typedef struct PGSSubContext { static av_cold int init_decoder(AVCodecContext *avctx) { - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/pictordec.c b/libavcodec/pictordec.c index 109c8f1035..a0affc98e8 100644 --- a/libavcodec/pictordec.c +++ b/libavcodec/pictordec.c @@ -136,7 +136,7 @@ static int decode_frame(AVCodecContext *avctx, esize = 0; } - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; if (s->width != avctx->width && s->height != avctx->height) { if (av_image_check_size(s->width, s->height, 0, avctx) < 0) diff --git a/libavcodec/pngdec.c b/libavcodec/pngdec.c index da63c5f28c..aaedeff187 100644 --- a/libavcodec/pngdec.c +++ b/libavcodec/pngdec.c @@ -463,28 +463,28 @@ static int decode_frame(AVCodecContext *avctx, if (s->bit_depth == 8 && s->color_type == PNG_COLOR_TYPE_RGB) { - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; } else if (s->bit_depth == 8 && s->color_type == PNG_COLOR_TYPE_RGB_ALPHA) { - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; } else if (s->bit_depth == 8 && s->color_type == PNG_COLOR_TYPE_GRAY) { - avctx->pix_fmt = PIX_FMT_GRAY8; + avctx->pix_fmt = AV_PIX_FMT_GRAY8; } else if (s->bit_depth == 16 && s->color_type == PNG_COLOR_TYPE_GRAY) { - avctx->pix_fmt = PIX_FMT_GRAY16BE; + avctx->pix_fmt = AV_PIX_FMT_GRAY16BE; } else if (s->bit_depth == 16 && s->color_type == PNG_COLOR_TYPE_RGB) { - avctx->pix_fmt = PIX_FMT_RGB48BE; + avctx->pix_fmt = AV_PIX_FMT_RGB48BE; } else if (s->bit_depth == 1 && s->color_type == PNG_COLOR_TYPE_GRAY) { - avctx->pix_fmt = PIX_FMT_MONOBLACK; + avctx->pix_fmt = AV_PIX_FMT_MONOBLACK; } else if (s->bit_depth == 8 && s->color_type == PNG_COLOR_TYPE_PALETTE) { - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; } else if (s->bit_depth == 8 && s->color_type == PNG_COLOR_TYPE_GRAY_ALPHA) { - avctx->pix_fmt = PIX_FMT_Y400A; + avctx->pix_fmt = AV_PIX_FMT_Y400A; } else { goto fail; } diff --git a/libavcodec/pngenc.c b/libavcodec/pngenc.c index a0b667972c..00a800c795 100644 --- a/libavcodec/pngenc.c +++ b/libavcodec/pngenc.c @@ -249,23 +249,23 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, is_progressive = !!(avctx->flags & CODEC_FLAG_INTERLACED_DCT); switch(avctx->pix_fmt) { - case PIX_FMT_RGB32: + case AV_PIX_FMT_RGB32: bit_depth = 8; color_type = PNG_COLOR_TYPE_RGB_ALPHA; break; - case PIX_FMT_RGB24: + case AV_PIX_FMT_RGB24: bit_depth = 8; color_type = PNG_COLOR_TYPE_RGB; break; - case PIX_FMT_GRAY8: + case AV_PIX_FMT_GRAY8: bit_depth = 8; color_type = PNG_COLOR_TYPE_GRAY; break; - case PIX_FMT_MONOBLACK: + case AV_PIX_FMT_MONOBLACK: bit_depth = 1; color_type = PNG_COLOR_TYPE_GRAY; break; - case PIX_FMT_PAL8: + case AV_PIX_FMT_PAL8: bit_depth = 8; color_type = PNG_COLOR_TYPE_PALETTE; break; @@ -447,7 +447,7 @@ static av_cold int png_enc_init(AVCodecContext *avctx){ ff_dsputil_init(&s->dsp, avctx); s->filter_type = av_clip(avctx->prediction_method, PNG_FILTER_VALUE_NONE, PNG_FILTER_VALUE_MIXED); - if(avctx->pix_fmt == PIX_FMT_MONOBLACK) + if(avctx->pix_fmt == AV_PIX_FMT_MONOBLACK) s->filter_type = PNG_FILTER_VALUE_NONE; return 0; @@ -460,9 +460,9 @@ AVCodec ff_png_encoder = { .priv_data_size = sizeof(PNGEncContext), .init = png_enc_init, .encode2 = encode_frame, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_RGB24, PIX_FMT_RGB32, PIX_FMT_PAL8, PIX_FMT_GRAY8, - PIX_FMT_MONOBLACK, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_RGB24, AV_PIX_FMT_RGB32, AV_PIX_FMT_PAL8, AV_PIX_FMT_GRAY8, + AV_PIX_FMT_MONOBLACK, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("PNG (Portable Network Graphics) image"), }; diff --git a/libavcodec/pnm.c b/libavcodec/pnm.c index f60d0e03d3..2a89a723f3 100644 --- a/libavcodec/pnm.c +++ b/libavcodec/pnm.c @@ -68,14 +68,14 @@ int ff_pnm_decode_header(AVCodecContext *avctx, PNMContext * const s) return -1; if (s->type==1 || s->type==4) { - avctx->pix_fmt = PIX_FMT_MONOWHITE; + avctx->pix_fmt = AV_PIX_FMT_MONOWHITE; } else if (s->type==2 || s->type==5) { if (avctx->codec_id == AV_CODEC_ID_PGMYUV) - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; else - avctx->pix_fmt = PIX_FMT_GRAY8; + avctx->pix_fmt = AV_PIX_FMT_GRAY8; } else if (s->type==3 || s->type==6) { - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; } else if (s->type==7) { w = -1; h = -1; @@ -114,19 +114,19 @@ int ff_pnm_decode_header(AVCodecContext *avctx, PNMContext * const s) avctx->height = h; if (depth == 1) { if (maxval == 1) - avctx->pix_fmt = PIX_FMT_MONOWHITE; + avctx->pix_fmt = AV_PIX_FMT_MONOWHITE; else - avctx->pix_fmt = PIX_FMT_GRAY8; + avctx->pix_fmt = AV_PIX_FMT_GRAY8; } else if (depth == 3) { if (maxval < 256) { - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; } else { av_log(avctx, AV_LOG_ERROR, "16-bit components are only supported for grayscale\n"); - avctx->pix_fmt = PIX_FMT_NONE; + avctx->pix_fmt = AV_PIX_FMT_NONE; return -1; } } else if (depth == 4) { - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; } else { return -1; } @@ -142,7 +142,7 @@ int ff_pnm_decode_header(AVCodecContext *avctx, PNMContext * const s) avctx->height = atoi(buf1); if(av_image_check_size(avctx->width, avctx->height, 0, avctx)) return -1; - if (avctx->pix_fmt != PIX_FMT_MONOWHITE) { + if (avctx->pix_fmt != AV_PIX_FMT_MONOWHITE) { pnm_get(s, buf1, sizeof(buf1)); s->maxval = atoi(buf1); if (s->maxval <= 0) { @@ -150,23 +150,23 @@ int ff_pnm_decode_header(AVCodecContext *avctx, PNMContext * const s) s->maxval = 255; } if (s->maxval >= 256) { - if (avctx->pix_fmt == PIX_FMT_GRAY8) { - avctx->pix_fmt = PIX_FMT_GRAY16BE; + if (avctx->pix_fmt == AV_PIX_FMT_GRAY8) { + avctx->pix_fmt = AV_PIX_FMT_GRAY16BE; if (s->maxval != 65535) - avctx->pix_fmt = PIX_FMT_GRAY16; - } else if (avctx->pix_fmt == PIX_FMT_RGB24) { + avctx->pix_fmt = AV_PIX_FMT_GRAY16; + } else if (avctx->pix_fmt == AV_PIX_FMT_RGB24) { if (s->maxval > 255) - avctx->pix_fmt = PIX_FMT_RGB48BE; + avctx->pix_fmt = AV_PIX_FMT_RGB48BE; } else { av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format\n"); - avctx->pix_fmt = PIX_FMT_NONE; + avctx->pix_fmt = AV_PIX_FMT_NONE; return -1; } } }else s->maxval=1; /* more check if YUV420 */ - if (avctx->pix_fmt == PIX_FMT_YUV420P) { + if (avctx->pix_fmt == AV_PIX_FMT_YUV420P) { if ((avctx->width & 1) != 0) return -1; h = (avctx->height * 2); diff --git a/libavcodec/pnmdec.c b/libavcodec/pnmdec.c index f7d8eff8e4..6fdd3f932d 100644 --- a/libavcodec/pnmdec.c +++ b/libavcodec/pnmdec.c @@ -58,33 +58,33 @@ static int pnm_decode_frame(AVCodecContext *avctx, void *data, switch (avctx->pix_fmt) { default: return -1; - case PIX_FMT_RGB48BE: + case AV_PIX_FMT_RGB48BE: n = avctx->width * 6; components=3; sample_len=16; goto do_read; - case PIX_FMT_RGB24: + case AV_PIX_FMT_RGB24: n = avctx->width * 3; components=3; sample_len=8; goto do_read; - case PIX_FMT_GRAY8: + case AV_PIX_FMT_GRAY8: n = avctx->width; components=1; sample_len=8; if (s->maxval < 255) upgrade = 1; goto do_read; - case PIX_FMT_GRAY16BE: - case PIX_FMT_GRAY16LE: + case AV_PIX_FMT_GRAY16BE: + case AV_PIX_FMT_GRAY16LE: n = avctx->width * 2; components=1; sample_len=16; if (s->maxval < 65535) upgrade = 2; goto do_read; - case PIX_FMT_MONOWHITE: - case PIX_FMT_MONOBLACK: + case AV_PIX_FMT_MONOWHITE: + case AV_PIX_FMT_MONOBLACK: n = (avctx->width + 7) >> 3; components=1; sample_len=1; @@ -133,7 +133,7 @@ static int pnm_decode_frame(AVCodecContext *avctx, void *data, } } break; - case PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV420P: { unsigned char *ptr1, *ptr2; @@ -161,7 +161,7 @@ static int pnm_decode_frame(AVCodecContext *avctx, void *data, } } break; - case PIX_FMT_RGB32: + case AV_PIX_FMT_RGB32: ptr = p->data[0]; linesize = p->linesize[0]; if (s->bytestream + avctx->width * avctx->height * 4 > s->bytestream_end) diff --git a/libavcodec/pnmenc.c b/libavcodec/pnmenc.c index b4194b68c8..2863db7566 100644 --- a/libavcodec/pnmenc.c +++ b/libavcodec/pnmenc.c @@ -51,27 +51,27 @@ static int pnm_encode_frame(AVCodecContext *avctx, AVPacket *pkt, h = avctx->height; h1 = h; switch (avctx->pix_fmt) { - case PIX_FMT_MONOWHITE: + case AV_PIX_FMT_MONOWHITE: c = '4'; n = (avctx->width + 7) >> 3; break; - case PIX_FMT_GRAY8: + case AV_PIX_FMT_GRAY8: c = '5'; n = avctx->width; break; - case PIX_FMT_GRAY16BE: + case AV_PIX_FMT_GRAY16BE: c = '5'; n = avctx->width * 2; break; - case PIX_FMT_RGB24: + case AV_PIX_FMT_RGB24: c = '6'; n = avctx->width * 3; break; - case PIX_FMT_RGB48BE: + case AV_PIX_FMT_RGB48BE: c = '6'; n = avctx->width * 6; break; - case PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV420P: c = '5'; n = avctx->width; h1 = (h * 3) / 2; @@ -82,9 +82,9 @@ static int pnm_encode_frame(AVCodecContext *avctx, AVPacket *pkt, snprintf(s->bytestream, s->bytestream_end - s->bytestream, "P%c\n%d %d\n", c, avctx->width, h1); s->bytestream += strlen(s->bytestream); - if (avctx->pix_fmt != PIX_FMT_MONOWHITE) { + if (avctx->pix_fmt != AV_PIX_FMT_MONOWHITE) { snprintf(s->bytestream, s->bytestream_end - s->bytestream, - "%d\n", (avctx->pix_fmt != PIX_FMT_GRAY16BE && avctx->pix_fmt != PIX_FMT_RGB48BE) ? 255 : 65535); + "%d\n", (avctx->pix_fmt != AV_PIX_FMT_GRAY16BE && avctx->pix_fmt != AV_PIX_FMT_RGB48BE) ? 255 : 65535); s->bytestream += strlen(s->bytestream); } @@ -96,7 +96,7 @@ static int pnm_encode_frame(AVCodecContext *avctx, AVPacket *pkt, ptr += linesize; } - if (avctx->pix_fmt == PIX_FMT_YUV420P) { + if (avctx->pix_fmt == AV_PIX_FMT_YUV420P) { h >>= 1; n >>= 1; ptr1 = p->data[1]; @@ -126,8 +126,8 @@ AVCodec ff_pgm_encoder = { .priv_data_size = sizeof(PNMContext), .init = ff_pnm_init, .encode2 = pnm_encode_frame, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_GRAY8, PIX_FMT_GRAY16BE, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY16BE, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("PGM (Portable GrayMap) image"), }; @@ -141,7 +141,7 @@ AVCodec ff_pgmyuv_encoder = { .priv_data_size = sizeof(PNMContext), .init = ff_pnm_init, .encode2 = pnm_encode_frame, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("PGMYUV (Portable GrayMap YUV) image"), }; #endif @@ -154,8 +154,8 @@ AVCodec ff_ppm_encoder = { .priv_data_size = sizeof(PNMContext), .init = ff_pnm_init, .encode2 = pnm_encode_frame, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_RGB24, PIX_FMT_RGB48BE, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_RGB24, AV_PIX_FMT_RGB48BE, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("PPM (Portable PixelMap) image"), }; @@ -169,8 +169,8 @@ AVCodec ff_pbm_encoder = { .priv_data_size = sizeof(PNMContext), .init = ff_pnm_init, .encode2 = pnm_encode_frame, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_MONOWHITE, - PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_MONOWHITE, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("PBM (Portable BitMap) image"), }; #endif diff --git a/libavcodec/proresdec.c b/libavcodec/proresdec.c index 86ce1576f3..47a5820cf2 100644 --- a/libavcodec/proresdec.c +++ b/libavcodec/proresdec.c @@ -139,10 +139,10 @@ static int decode_frame_header(ProresContext *ctx, const uint8_t *buf, ctx->num_chroma_blocks = (1 << ctx->chroma_factor) >> 1; switch (ctx->chroma_factor) { case 2: - avctx->pix_fmt = PIX_FMT_YUV422P10; + avctx->pix_fmt = AV_PIX_FMT_YUV422P10; break; case 3: - avctx->pix_fmt = PIX_FMT_YUV444P10; + avctx->pix_fmt = AV_PIX_FMT_YUV444P10; break; default: av_log(avctx, AV_LOG_ERROR, diff --git a/libavcodec/proresenc.c b/libavcodec/proresenc.c index 8c0b4be68c..c4716d6c91 100644 --- a/libavcodec/proresenc.c +++ b/libavcodec/proresenc.c @@ -33,7 +33,7 @@ #define MAX_MBS_PER_SLICE 8 -#define MAX_PLANES 3 // should be increased to 4 when there's PIX_FMT_YUV444AP10 +#define MAX_PLANES 3 // should be increased to 4 when there's AV_PIX_FMT_YUV444AP10 enum { PRORES_PROFILE_PROXY = 0, @@ -904,7 +904,7 @@ static av_cold int encode_init(AVCodecContext *avctx) return AVERROR(EINVAL); } - ctx->chroma_factor = avctx->pix_fmt == PIX_FMT_YUV422P10 + ctx->chroma_factor = avctx->pix_fmt == AV_PIX_FMT_YUV422P10 ? CFACTOR_Y422 : CFACTOR_Y444; ctx->profile_info = prores_profile_info + ctx->profile; @@ -1071,8 +1071,8 @@ AVCodec ff_prores_encoder = { .encode2 = encode_frame, .capabilities = CODEC_CAP_SLICE_THREADS, .long_name = NULL_IF_CONFIG_SMALL("Apple ProRes (iCodec Pro)"), - .pix_fmts = (const enum PixelFormat[]) { - PIX_FMT_YUV422P10, PIX_FMT_YUV444P10, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]) { + AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, AV_PIX_FMT_NONE }, .priv_class = &proresenc_class, }; diff --git a/libavcodec/ptx.c b/libavcodec/ptx.c index e0f5a1b301..e6835e6f4a 100644 --- a/libavcodec/ptx.c +++ b/libavcodec/ptx.c @@ -59,7 +59,7 @@ static int ptx_decode_frame(AVCodecContext *avctx, void *data, int *data_size, return -1; } - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; if (buf_end - buf < offset) return AVERROR_INVALIDDATA; diff --git a/libavcodec/qdrw.c b/libavcodec/qdrw.c index 9e8c204f73..934a1fca1b 100644 --- a/libavcodec/qdrw.c +++ b/libavcodec/qdrw.c @@ -148,7 +148,7 @@ static int decode_frame(AVCodecContext *avctx, static av_cold int decode_init(AVCodecContext *avctx){ // QdrawContext * const a = avctx->priv_data; - avctx->pix_fmt= PIX_FMT_PAL8; + avctx->pix_fmt= AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/qpeg.c b/libavcodec/qpeg.c index 6092622beb..1ee764bb99 100644 --- a/libavcodec/qpeg.c +++ b/libavcodec/qpeg.c @@ -290,7 +290,7 @@ static av_cold int decode_init(AVCodecContext *avctx){ QpegContext * const a = avctx->priv_data; a->avctx = avctx; - avctx->pix_fmt= PIX_FMT_PAL8; + avctx->pix_fmt= AV_PIX_FMT_PAL8; a->refdata = av_malloc(avctx->width * avctx->height); return 0; diff --git a/libavcodec/qtrle.c b/libavcodec/qtrle.c index c289114d1f..3f173a596b 100644 --- a/libavcodec/qtrle.c +++ b/libavcodec/qtrle.c @@ -346,7 +346,7 @@ static av_cold int qtrle_decode_init(AVCodecContext *avctx) switch (avctx->bits_per_coded_sample) { case 1: case 33: - avctx->pix_fmt = PIX_FMT_MONOWHITE; + avctx->pix_fmt = AV_PIX_FMT_MONOWHITE; break; case 2: @@ -355,19 +355,19 @@ static av_cold int qtrle_decode_init(AVCodecContext *avctx) case 34: case 36: case 40: - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; break; case 16: - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; break; case 24: - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; break; case 32: - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; break; default: diff --git a/libavcodec/qtrleenc.c b/libavcodec/qtrleenc.c index ae70e733bd..bb686f571a 100644 --- a/libavcodec/qtrleenc.c +++ b/libavcodec/qtrleenc.c @@ -70,13 +70,13 @@ static av_cold int qtrle_encode_init(AVCodecContext *avctx) s->avctx=avctx; switch (avctx->pix_fmt) { - case PIX_FMT_RGB555BE: + case AV_PIX_FMT_RGB555BE: s->pixel_size = 2; break; - case PIX_FMT_RGB24: + case AV_PIX_FMT_RGB24: s->pixel_size = 3; break; - case PIX_FMT_ARGB: + case AV_PIX_FMT_ARGB: s->pixel_size = 4; break; default: @@ -335,8 +335,8 @@ AVCodec ff_qtrle_encoder = { .init = qtrle_encode_init, .encode2 = qtrle_encode_frame, .close = qtrle_encode_end, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_RGB24, PIX_FMT_RGB555BE, PIX_FMT_ARGB, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_RGB24, AV_PIX_FMT_RGB555BE, AV_PIX_FMT_ARGB, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("QuickTime Animation (RLE) video"), }; diff --git a/libavcodec/r210dec.c b/libavcodec/r210dec.c index 5b87ebdb2d..05f94b7e1a 100644 --- a/libavcodec/r210dec.c +++ b/libavcodec/r210dec.c @@ -26,7 +26,7 @@ static av_cold int decode_init(AVCodecContext *avctx) { - avctx->pix_fmt = PIX_FMT_RGB48; + avctx->pix_fmt = AV_PIX_FMT_RGB48; avctx->bits_per_raw_sample = 10; avctx->coded_frame = avcodec_alloc_frame(); diff --git a/libavcodec/raw.c b/libavcodec/raw.c index ecf5365210..6f0c9230b9 100644 --- a/libavcodec/raw.c +++ b/libavcodec/raw.c @@ -29,126 +29,126 @@ #include "libavutil/common.h" const PixelFormatTag ff_raw_pix_fmt_tags[] = { - { PIX_FMT_YUV420P, MKTAG('I', '4', '2', '0') }, /* Planar formats */ - { PIX_FMT_YUV420P, MKTAG('I', 'Y', 'U', 'V') }, - { PIX_FMT_YUV420P, MKTAG('Y', 'V', '1', '2') }, - { PIX_FMT_YUV410P, MKTAG('Y', 'U', 'V', '9') }, - { PIX_FMT_YUV410P, MKTAG('Y', 'V', 'U', '9') }, - { PIX_FMT_YUV411P, MKTAG('Y', '4', '1', 'B') }, - { PIX_FMT_YUV422P, MKTAG('Y', '4', '2', 'B') }, - { PIX_FMT_YUV422P, MKTAG('P', '4', '2', '2') }, - { PIX_FMT_YUV422P, MKTAG('Y', 'V', '1', '6') }, + { AV_PIX_FMT_YUV420P, MKTAG('I', '4', '2', '0') }, /* Planar formats */ + { AV_PIX_FMT_YUV420P, MKTAG('I', 'Y', 'U', 'V') }, + { AV_PIX_FMT_YUV420P, MKTAG('Y', 'V', '1', '2') }, + { AV_PIX_FMT_YUV410P, MKTAG('Y', 'U', 'V', '9') }, + { AV_PIX_FMT_YUV410P, MKTAG('Y', 'V', 'U', '9') }, + { AV_PIX_FMT_YUV411P, MKTAG('Y', '4', '1', 'B') }, + { AV_PIX_FMT_YUV422P, MKTAG('Y', '4', '2', 'B') }, + { AV_PIX_FMT_YUV422P, MKTAG('P', '4', '2', '2') }, + { AV_PIX_FMT_YUV422P, MKTAG('Y', 'V', '1', '6') }, /* yuvjXXX formats are deprecated hacks specific to libav*, they are identical to yuvXXX */ - { PIX_FMT_YUVJ420P, MKTAG('I', '4', '2', '0') }, /* Planar formats */ - { PIX_FMT_YUVJ420P, MKTAG('I', 'Y', 'U', 'V') }, - { PIX_FMT_YUVJ420P, MKTAG('Y', 'V', '1', '2') }, - { PIX_FMT_YUVJ422P, MKTAG('Y', '4', '2', 'B') }, - { PIX_FMT_YUVJ422P, MKTAG('P', '4', '2', '2') }, - { PIX_FMT_GRAY8, MKTAG('Y', '8', '0', '0') }, - { PIX_FMT_GRAY8, MKTAG('Y', '8', ' ', ' ') }, + { AV_PIX_FMT_YUVJ420P, MKTAG('I', '4', '2', '0') }, /* Planar formats */ + { AV_PIX_FMT_YUVJ420P, MKTAG('I', 'Y', 'U', 'V') }, + { AV_PIX_FMT_YUVJ420P, MKTAG('Y', 'V', '1', '2') }, + { AV_PIX_FMT_YUVJ422P, MKTAG('Y', '4', '2', 'B') }, + { AV_PIX_FMT_YUVJ422P, MKTAG('P', '4', '2', '2') }, + { AV_PIX_FMT_GRAY8, MKTAG('Y', '8', '0', '0') }, + { AV_PIX_FMT_GRAY8, MKTAG('Y', '8', ' ', ' ') }, - { PIX_FMT_YUYV422, MKTAG('Y', 'U', 'Y', '2') }, /* Packed formats */ - { PIX_FMT_YUYV422, MKTAG('Y', '4', '2', '2') }, - { PIX_FMT_YUYV422, MKTAG('V', '4', '2', '2') }, - { PIX_FMT_YUYV422, MKTAG('V', 'Y', 'U', 'Y') }, - { PIX_FMT_YUYV422, MKTAG('Y', 'U', 'N', 'V') }, - { PIX_FMT_UYVY422, MKTAG('U', 'Y', 'V', 'Y') }, - { PIX_FMT_UYVY422, MKTAG('H', 'D', 'Y', 'C') }, - { PIX_FMT_UYVY422, MKTAG('U', 'Y', 'N', 'V') }, - { PIX_FMT_UYVY422, MKTAG('U', 'Y', 'N', 'Y') }, - { PIX_FMT_UYVY422, MKTAG('u', 'y', 'v', '1') }, - { PIX_FMT_UYVY422, MKTAG('2', 'V', 'u', '1') }, - { PIX_FMT_UYVY422, MKTAG('A', 'V', 'R', 'n') }, /* Avid AVI Codec 1:1 */ - { PIX_FMT_UYVY422, MKTAG('A', 'V', '1', 'x') }, /* Avid 1:1x */ - { PIX_FMT_UYVY422, MKTAG('A', 'V', 'u', 'p') }, - { PIX_FMT_UYVY422, MKTAG('V', 'D', 'T', 'Z') }, /* SoftLab-NSK VideoTizer */ - { PIX_FMT_UYVY422, MKTAG('a', 'u', 'v', '2') }, - { PIX_FMT_UYYVYY411, MKTAG('Y', '4', '1', '1') }, - { PIX_FMT_GRAY8, MKTAG('G', 'R', 'E', 'Y') }, - { PIX_FMT_NV12, MKTAG('N', 'V', '1', '2') }, - { PIX_FMT_NV21, MKTAG('N', 'V', '2', '1') }, + { AV_PIX_FMT_YUYV422, MKTAG('Y', 'U', 'Y', '2') }, /* Packed formats */ + { AV_PIX_FMT_YUYV422, MKTAG('Y', '4', '2', '2') }, + { AV_PIX_FMT_YUYV422, MKTAG('V', '4', '2', '2') }, + { AV_PIX_FMT_YUYV422, MKTAG('V', 'Y', 'U', 'Y') }, + { AV_PIX_FMT_YUYV422, MKTAG('Y', 'U', 'N', 'V') }, + { AV_PIX_FMT_UYVY422, MKTAG('U', 'Y', 'V', 'Y') }, + { AV_PIX_FMT_UYVY422, MKTAG('H', 'D', 'Y', 'C') }, + { AV_PIX_FMT_UYVY422, MKTAG('U', 'Y', 'N', 'V') }, + { AV_PIX_FMT_UYVY422, MKTAG('U', 'Y', 'N', 'Y') }, + { AV_PIX_FMT_UYVY422, MKTAG('u', 'y', 'v', '1') }, + { AV_PIX_FMT_UYVY422, MKTAG('2', 'V', 'u', '1') }, + { AV_PIX_FMT_UYVY422, MKTAG('A', 'V', 'R', 'n') }, /* Avid AVI Codec 1:1 */ + { AV_PIX_FMT_UYVY422, MKTAG('A', 'V', '1', 'x') }, /* Avid 1:1x */ + { AV_PIX_FMT_UYVY422, MKTAG('A', 'V', 'u', 'p') }, + { AV_PIX_FMT_UYVY422, MKTAG('V', 'D', 'T', 'Z') }, /* SoftLab-NSK VideoTizer */ + { AV_PIX_FMT_UYVY422, MKTAG('a', 'u', 'v', '2') }, + { AV_PIX_FMT_UYYVYY411, MKTAG('Y', '4', '1', '1') }, + { AV_PIX_FMT_GRAY8, MKTAG('G', 'R', 'E', 'Y') }, + { AV_PIX_FMT_NV12, MKTAG('N', 'V', '1', '2') }, + { AV_PIX_FMT_NV21, MKTAG('N', 'V', '2', '1') }, /* nut */ - { PIX_FMT_RGB555LE, MKTAG('R', 'G', 'B', 15) }, - { PIX_FMT_BGR555LE, MKTAG('B', 'G', 'R', 15) }, - { PIX_FMT_RGB565LE, MKTAG('R', 'G', 'B', 16) }, - { PIX_FMT_BGR565LE, MKTAG('B', 'G', 'R', 16) }, - { PIX_FMT_RGB555BE, MKTAG(15 , 'B', 'G', 'R') }, - { PIX_FMT_BGR555BE, MKTAG(15 , 'R', 'G', 'B') }, - { PIX_FMT_RGB565BE, MKTAG(16 , 'B', 'G', 'R') }, - { PIX_FMT_BGR565BE, MKTAG(16 , 'R', 'G', 'B') }, - { PIX_FMT_RGB444LE, MKTAG('R', 'G', 'B', 12) }, - { PIX_FMT_BGR444LE, MKTAG('B', 'G', 'R', 12) }, - { PIX_FMT_RGB444BE, MKTAG(12 , 'B', 'G', 'R') }, - { PIX_FMT_BGR444BE, MKTAG(12 , 'R', 'G', 'B') }, - { PIX_FMT_RGBA, MKTAG('R', 'G', 'B', 'A') }, - { PIX_FMT_BGRA, MKTAG('B', 'G', 'R', 'A') }, - { PIX_FMT_ABGR, MKTAG('A', 'B', 'G', 'R') }, - { PIX_FMT_ARGB, MKTAG('A', 'R', 'G', 'B') }, - { PIX_FMT_RGB24, MKTAG('R', 'G', 'B', 24 ) }, - { PIX_FMT_BGR24, MKTAG('B', 'G', 'R', 24 ) }, - { PIX_FMT_YUV411P, MKTAG('4', '1', '1', 'P') }, - { PIX_FMT_YUV422P, MKTAG('4', '2', '2', 'P') }, - { PIX_FMT_YUVJ422P, MKTAG('4', '2', '2', 'P') }, - { PIX_FMT_YUV440P, MKTAG('4', '4', '0', 'P') }, - { PIX_FMT_YUVJ440P, MKTAG('4', '4', '0', 'P') }, - { PIX_FMT_YUV444P, MKTAG('4', '4', '4', 'P') }, - { PIX_FMT_YUVJ444P, MKTAG('4', '4', '4', 'P') }, - { PIX_FMT_MONOWHITE,MKTAG('B', '1', 'W', '0') }, - { PIX_FMT_MONOBLACK,MKTAG('B', '0', 'W', '1') }, - { PIX_FMT_BGR8, MKTAG('B', 'G', 'R', 8 ) }, - { PIX_FMT_RGB8, MKTAG('R', 'G', 'B', 8 ) }, - { PIX_FMT_BGR4, MKTAG('B', 'G', 'R', 4 ) }, - { PIX_FMT_RGB4, MKTAG('R', 'G', 'B', 4 ) }, - { PIX_FMT_RGB4_BYTE,MKTAG('B', '4', 'B', 'Y') }, - { PIX_FMT_BGR4_BYTE,MKTAG('R', '4', 'B', 'Y') }, - { PIX_FMT_RGB48LE, MKTAG('R', 'G', 'B', 48 ) }, - { PIX_FMT_RGB48BE, MKTAG( 48, 'R', 'G', 'B') }, - { PIX_FMT_BGR48LE, MKTAG('B', 'G', 'R', 48 ) }, - { PIX_FMT_BGR48BE, MKTAG( 48, 'B', 'G', 'R') }, - { PIX_FMT_GRAY16LE, MKTAG('Y', '1', 0 , 16 ) }, - { PIX_FMT_GRAY16BE, MKTAG(16 , 0 , '1', 'Y') }, - { PIX_FMT_YUV420P10LE, MKTAG('Y', '3', 11 , 10 ) }, - { PIX_FMT_YUV420P10BE, MKTAG(10 , 11 , '3', 'Y') }, - { PIX_FMT_YUV422P10LE, MKTAG('Y', '3', 10 , 10 ) }, - { PIX_FMT_YUV422P10BE, MKTAG(10 , 10 , '3', 'Y') }, - { PIX_FMT_YUV444P10LE, MKTAG('Y', '3', 0 , 10 ) }, - { PIX_FMT_YUV444P10BE, MKTAG(10 , 0 , '3', 'Y') }, - { PIX_FMT_YUV420P16LE, MKTAG('Y', '3', 11 , 16 ) }, - { PIX_FMT_YUV420P16BE, MKTAG(16 , 11 , '3', 'Y') }, - { PIX_FMT_YUV422P16LE, MKTAG('Y', '3', 10 , 16 ) }, - { PIX_FMT_YUV422P16BE, MKTAG(16 , 10 , '3', 'Y') }, - { PIX_FMT_YUV444P16LE, MKTAG('Y', '3', 0 , 16 ) }, - { PIX_FMT_YUV444P16BE, MKTAG(16 , 0 , '3', 'Y') }, - { PIX_FMT_YUVA420P, MKTAG('Y', '4', 11 , 8 ) }, - { PIX_FMT_Y400A, MKTAG('Y', '2', 0 , 8 ) }, + { AV_PIX_FMT_RGB555LE, MKTAG('R', 'G', 'B', 15) }, + { AV_PIX_FMT_BGR555LE, MKTAG('B', 'G', 'R', 15) }, + { AV_PIX_FMT_RGB565LE, MKTAG('R', 'G', 'B', 16) }, + { AV_PIX_FMT_BGR565LE, MKTAG('B', 'G', 'R', 16) }, + { AV_PIX_FMT_RGB555BE, MKTAG(15 , 'B', 'G', 'R') }, + { AV_PIX_FMT_BGR555BE, MKTAG(15 , 'R', 'G', 'B') }, + { AV_PIX_FMT_RGB565BE, MKTAG(16 , 'B', 'G', 'R') }, + { AV_PIX_FMT_BGR565BE, MKTAG(16 , 'R', 'G', 'B') }, + { AV_PIX_FMT_RGB444LE, MKTAG('R', 'G', 'B', 12) }, + { AV_PIX_FMT_BGR444LE, MKTAG('B', 'G', 'R', 12) }, + { AV_PIX_FMT_RGB444BE, MKTAG(12 , 'B', 'G', 'R') }, + { AV_PIX_FMT_BGR444BE, MKTAG(12 , 'R', 'G', 'B') }, + { AV_PIX_FMT_RGBA, MKTAG('R', 'G', 'B', 'A') }, + { AV_PIX_FMT_BGRA, MKTAG('B', 'G', 'R', 'A') }, + { AV_PIX_FMT_ABGR, MKTAG('A', 'B', 'G', 'R') }, + { AV_PIX_FMT_ARGB, MKTAG('A', 'R', 'G', 'B') }, + { AV_PIX_FMT_RGB24, MKTAG('R', 'G', 'B', 24 ) }, + { AV_PIX_FMT_BGR24, MKTAG('B', 'G', 'R', 24 ) }, + { AV_PIX_FMT_YUV411P, MKTAG('4', '1', '1', 'P') }, + { AV_PIX_FMT_YUV422P, MKTAG('4', '2', '2', 'P') }, + { AV_PIX_FMT_YUVJ422P, MKTAG('4', '2', '2', 'P') }, + { AV_PIX_FMT_YUV440P, MKTAG('4', '4', '0', 'P') }, + { AV_PIX_FMT_YUVJ440P, MKTAG('4', '4', '0', 'P') }, + { AV_PIX_FMT_YUV444P, MKTAG('4', '4', '4', 'P') }, + { AV_PIX_FMT_YUVJ444P, MKTAG('4', '4', '4', 'P') }, + { AV_PIX_FMT_MONOWHITE,MKTAG('B', '1', 'W', '0') }, + { AV_PIX_FMT_MONOBLACK,MKTAG('B', '0', 'W', '1') }, + { AV_PIX_FMT_BGR8, MKTAG('B', 'G', 'R', 8 ) }, + { AV_PIX_FMT_RGB8, MKTAG('R', 'G', 'B', 8 ) }, + { AV_PIX_FMT_BGR4, MKTAG('B', 'G', 'R', 4 ) }, + { AV_PIX_FMT_RGB4, MKTAG('R', 'G', 'B', 4 ) }, + { AV_PIX_FMT_RGB4_BYTE,MKTAG('B', '4', 'B', 'Y') }, + { AV_PIX_FMT_BGR4_BYTE,MKTAG('R', '4', 'B', 'Y') }, + { AV_PIX_FMT_RGB48LE, MKTAG('R', 'G', 'B', 48 ) }, + { AV_PIX_FMT_RGB48BE, MKTAG( 48, 'R', 'G', 'B') }, + { AV_PIX_FMT_BGR48LE, MKTAG('B', 'G', 'R', 48 ) }, + { AV_PIX_FMT_BGR48BE, MKTAG( 48, 'B', 'G', 'R') }, + { AV_PIX_FMT_GRAY16LE, MKTAG('Y', '1', 0 , 16 ) }, + { AV_PIX_FMT_GRAY16BE, MKTAG(16 , 0 , '1', 'Y') }, + { AV_PIX_FMT_YUV420P10LE, MKTAG('Y', '3', 11 , 10 ) }, + { AV_PIX_FMT_YUV420P10BE, MKTAG(10 , 11 , '3', 'Y') }, + { AV_PIX_FMT_YUV422P10LE, MKTAG('Y', '3', 10 , 10 ) }, + { AV_PIX_FMT_YUV422P10BE, MKTAG(10 , 10 , '3', 'Y') }, + { AV_PIX_FMT_YUV444P10LE, MKTAG('Y', '3', 0 , 10 ) }, + { AV_PIX_FMT_YUV444P10BE, MKTAG(10 , 0 , '3', 'Y') }, + { AV_PIX_FMT_YUV420P16LE, MKTAG('Y', '3', 11 , 16 ) }, + { AV_PIX_FMT_YUV420P16BE, MKTAG(16 , 11 , '3', 'Y') }, + { AV_PIX_FMT_YUV422P16LE, MKTAG('Y', '3', 10 , 16 ) }, + { AV_PIX_FMT_YUV422P16BE, MKTAG(16 , 10 , '3', 'Y') }, + { AV_PIX_FMT_YUV444P16LE, MKTAG('Y', '3', 0 , 16 ) }, + { AV_PIX_FMT_YUV444P16BE, MKTAG(16 , 0 , '3', 'Y') }, + { AV_PIX_FMT_YUVA420P, MKTAG('Y', '4', 11 , 8 ) }, + { AV_PIX_FMT_Y400A, MKTAG('Y', '2', 0 , 8 ) }, /* quicktime */ - { PIX_FMT_UYVY422, MKTAG('2', 'v', 'u', 'y') }, - { PIX_FMT_UYVY422, MKTAG('2', 'V', 'u', 'y') }, - { PIX_FMT_UYVY422, MKTAG('A', 'V', 'U', 'I') }, /* FIXME merge both fields */ - { PIX_FMT_YUYV422, MKTAG('y', 'u', 'v', '2') }, - { PIX_FMT_YUYV422, MKTAG('y', 'u', 'v', 's') }, - { PIX_FMT_YUYV422, MKTAG('D', 'V', 'O', 'O') }, /* Digital Voodoo SD 8 Bit */ - { PIX_FMT_RGB555LE,MKTAG('L', '5', '5', '5') }, - { PIX_FMT_RGB565LE,MKTAG('L', '5', '6', '5') }, - { PIX_FMT_RGB565BE,MKTAG('B', '5', '6', '5') }, - { PIX_FMT_BGR24, MKTAG('2', '4', 'B', 'G') }, - { PIX_FMT_BGRA, MKTAG('B', 'G', 'R', 'A') }, - { PIX_FMT_RGBA, MKTAG('R', 'G', 'B', 'A') }, - { PIX_FMT_ABGR, MKTAG('A', 'B', 'G', 'R') }, - { PIX_FMT_GRAY16BE,MKTAG('b', '1', '6', 'g') }, - { PIX_FMT_RGB48BE, MKTAG('b', '4', '8', 'r') }, + { AV_PIX_FMT_UYVY422, MKTAG('2', 'v', 'u', 'y') }, + { AV_PIX_FMT_UYVY422, MKTAG('2', 'V', 'u', 'y') }, + { AV_PIX_FMT_UYVY422, MKTAG('A', 'V', 'U', 'I') }, /* FIXME merge both fields */ + { AV_PIX_FMT_YUYV422, MKTAG('y', 'u', 'v', '2') }, + { AV_PIX_FMT_YUYV422, MKTAG('y', 'u', 'v', 's') }, + { AV_PIX_FMT_YUYV422, MKTAG('D', 'V', 'O', 'O') }, /* Digital Voodoo SD 8 Bit */ + { AV_PIX_FMT_RGB555LE,MKTAG('L', '5', '5', '5') }, + { AV_PIX_FMT_RGB565LE,MKTAG('L', '5', '6', '5') }, + { AV_PIX_FMT_RGB565BE,MKTAG('B', '5', '6', '5') }, + { AV_PIX_FMT_BGR24, MKTAG('2', '4', 'B', 'G') }, + { AV_PIX_FMT_BGRA, MKTAG('B', 'G', 'R', 'A') }, + { AV_PIX_FMT_RGBA, MKTAG('R', 'G', 'B', 'A') }, + { AV_PIX_FMT_ABGR, MKTAG('A', 'B', 'G', 'R') }, + { AV_PIX_FMT_GRAY16BE,MKTAG('b', '1', '6', 'g') }, + { AV_PIX_FMT_RGB48BE, MKTAG('b', '4', '8', 'r') }, /* special */ - { PIX_FMT_RGB565LE,MKTAG( 3 , 0 , 0 , 0 ) }, /* flipped RGB565LE */ - { PIX_FMT_YUV444P, MKTAG('Y', 'V', '2', '4') }, /* YUV444P, swapped UV */ + { AV_PIX_FMT_RGB565LE,MKTAG( 3 , 0 , 0 , 0 ) }, /* flipped RGB565LE */ + { AV_PIX_FMT_YUV444P, MKTAG('Y', 'V', '2', '4') }, /* YUV444P, swapped UV */ - { PIX_FMT_NONE, 0 }, + { AV_PIX_FMT_NONE, 0 }, }; -unsigned int avcodec_pix_fmt_to_codec_tag(enum PixelFormat fmt) +unsigned int avcodec_pix_fmt_to_codec_tag(enum AVPixelFormat fmt) { const PixelFormatTag *tags = ff_raw_pix_fmt_tags; while (tags->pix_fmt >= 0) { diff --git a/libavcodec/raw.h b/libavcodec/raw.h index 2caa3faff5..bf66671d16 100644 --- a/libavcodec/raw.h +++ b/libavcodec/raw.h @@ -30,7 +30,7 @@ #include "avcodec.h" typedef struct PixelFormatTag { - enum PixelFormat pix_fmt; + enum AVPixelFormat pix_fmt; unsigned int fourcc; } PixelFormatTag; diff --git a/libavcodec/rawdec.c b/libavcodec/rawdec.c index ee316ea723..4d12077676 100644 --- a/libavcodec/rawdec.c +++ b/libavcodec/rawdec.c @@ -39,38 +39,38 @@ typedef struct RawVideoContext { } RawVideoContext; static const PixelFormatTag pix_fmt_bps_avi[] = { - { PIX_FMT_PAL8, 4 }, - { PIX_FMT_PAL8, 8 }, - { PIX_FMT_RGB444, 12 }, - { PIX_FMT_RGB555, 15 }, - { PIX_FMT_RGB555, 16 }, - { PIX_FMT_BGR24, 24 }, - { PIX_FMT_RGB32, 32 }, - { PIX_FMT_NONE, 0 }, + { AV_PIX_FMT_PAL8, 4 }, + { AV_PIX_FMT_PAL8, 8 }, + { AV_PIX_FMT_RGB444, 12 }, + { AV_PIX_FMT_RGB555, 15 }, + { AV_PIX_FMT_RGB555, 16 }, + { AV_PIX_FMT_BGR24, 24 }, + { AV_PIX_FMT_RGB32, 32 }, + { AV_PIX_FMT_NONE, 0 }, }; static const PixelFormatTag pix_fmt_bps_mov[] = { - { PIX_FMT_MONOWHITE, 1 }, - { PIX_FMT_PAL8, 2 }, - { PIX_FMT_PAL8, 4 }, - { PIX_FMT_PAL8, 8 }, + { AV_PIX_FMT_MONOWHITE, 1 }, + { AV_PIX_FMT_PAL8, 2 }, + { AV_PIX_FMT_PAL8, 4 }, + { AV_PIX_FMT_PAL8, 8 }, // FIXME swscale does not support 16 bit in .mov, sample 16bit.mov // http://developer.apple.com/documentation/QuickTime/QTFF/QTFFChap3/qtff3.html - { PIX_FMT_RGB555BE, 16 }, - { PIX_FMT_RGB24, 24 }, - { PIX_FMT_ARGB, 32 }, - { PIX_FMT_MONOWHITE,33 }, - { PIX_FMT_NONE, 0 }, + { AV_PIX_FMT_RGB555BE, 16 }, + { AV_PIX_FMT_RGB24, 24 }, + { AV_PIX_FMT_ARGB, 32 }, + { AV_PIX_FMT_MONOWHITE,33 }, + { AV_PIX_FMT_NONE, 0 }, }; -static enum PixelFormat find_pix_fmt(const PixelFormatTag *tags, unsigned int fourcc) +static enum AVPixelFormat find_pix_fmt(const PixelFormatTag *tags, unsigned int fourcc) { while (tags->pix_fmt >= 0) { if (tags->fourcc == fourcc) return tags->pix_fmt; tags++; } - return PIX_FMT_YUV420P; + return AV_PIX_FMT_YUV420P; } static av_cold int raw_init_decoder(AVCodecContext *avctx) @@ -83,13 +83,13 @@ static av_cold int raw_init_decoder(AVCodecContext *avctx) avctx->pix_fmt = find_pix_fmt(pix_fmt_bps_avi, avctx->bits_per_coded_sample); else if (avctx->codec_tag) avctx->pix_fmt = find_pix_fmt(ff_raw_pix_fmt_tags, avctx->codec_tag); - else if (avctx->pix_fmt == PIX_FMT_NONE && avctx->bits_per_coded_sample) + else if (avctx->pix_fmt == AV_PIX_FMT_NONE && avctx->bits_per_coded_sample) avctx->pix_fmt = find_pix_fmt(pix_fmt_bps_avi, avctx->bits_per_coded_sample); ff_set_systematic_pal2(context->palette, avctx->pix_fmt); context->length = avpicture_get_size(avctx->pix_fmt, avctx->width, avctx->height); if((avctx->bits_per_coded_sample == 4 || avctx->bits_per_coded_sample == 2) && - avctx->pix_fmt==PIX_FMT_PAL8 && + avctx->pix_fmt==AV_PIX_FMT_PAL8 && (!avctx->codec_tag || avctx->codec_tag == MKTAG('r','a','w',' '))){ context->buffer = av_malloc(context->length); if (!context->buffer) @@ -130,7 +130,7 @@ static int raw_decode(AVCodecContext *avctx, frame->reordered_opaque = avctx->reordered_opaque; frame->pkt_pts = avctx->pkt->pts; - if(buf_size < context->length - (avctx->pix_fmt==PIX_FMT_PAL8 ? 256*4 : 0)) + if(buf_size < context->length - (avctx->pix_fmt==AV_PIX_FMT_PAL8 ? 256*4 : 0)) return -1; //2bpp and 4bpp raw in avi and mov (yes this is ugly ...) @@ -160,11 +160,11 @@ static int raw_decode(AVCodecContext *avctx, if ((res = avpicture_fill(picture, buf, avctx->pix_fmt, avctx->width, avctx->height)) < 0) return res; - if((avctx->pix_fmt==PIX_FMT_PAL8 && buf_size < context->length) || + if((avctx->pix_fmt==AV_PIX_FMT_PAL8 && buf_size < context->length) || (av_pix_fmt_descriptors[avctx->pix_fmt].flags & PIX_FMT_PSEUDOPAL)) { frame->data[1]= context->palette; } - if (avctx->pix_fmt == PIX_FMT_PAL8) { + if (avctx->pix_fmt == AV_PIX_FMT_PAL8) { const uint8_t *pal = av_packet_get_side_data(avpkt, AV_PKT_DATA_PALETTE, NULL); if (pal) { @@ -172,7 +172,7 @@ static int raw_decode(AVCodecContext *avctx, frame->palette_has_changed = 1; } } - if(avctx->pix_fmt==PIX_FMT_BGR24 && ((frame->linesize[0]+3)&~3)*avctx->height <= buf_size) + if(avctx->pix_fmt==AV_PIX_FMT_BGR24 && ((frame->linesize[0]+3)&~3)*avctx->height <= buf_size) frame->linesize[0] = (frame->linesize[0]+3)&~3; if(context->flip) @@ -185,7 +185,7 @@ static int raw_decode(AVCodecContext *avctx, FFSWAP(uint8_t *, picture->data[1], picture->data[2]); if(avctx->codec_tag == AV_RL32("yuv2") && - avctx->pix_fmt == PIX_FMT_YUYV422) { + avctx->pix_fmt == AV_PIX_FMT_YUYV422) { int x, y; uint8_t *line = picture->data[0]; for(y = 0; y < avctx->height; y++) { diff --git a/libavcodec/rawenc.c b/libavcodec/rawenc.c index 927d6f6633..e94b6a500b 100644 --- a/libavcodec/rawenc.c +++ b/libavcodec/rawenc.c @@ -57,7 +57,7 @@ static int raw_encode(AVCodecContext *avctx, AVPacket *pkt, return ret; if(avctx->codec_tag == AV_RL32("yuv2") && ret > 0 && - avctx->pix_fmt == PIX_FMT_YUYV422) { + avctx->pix_fmt == AV_PIX_FMT_YUYV422) { int x; for(x = 1; x < avctx->height*avctx->width*2; x += 2) pkt->data[x] ^= 0x80; diff --git a/libavcodec/rl2.c b/libavcodec/rl2.c index 57015e7de8..84434a29a1 100644 --- a/libavcodec/rl2.c +++ b/libavcodec/rl2.c @@ -134,7 +134,7 @@ static av_cold int rl2_decode_init(AVCodecContext *avctx) int back_size; int i; s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; /** parse extra data */ if(!avctx->extradata || avctx->extradata_size < EXTRADATA1_SIZE){ diff --git a/libavcodec/roqvideodec.c b/libavcodec/roqvideodec.c index a88e9f2ae9..bf5664b9e5 100644 --- a/libavcodec/roqvideodec.c +++ b/libavcodec/roqvideodec.c @@ -163,7 +163,7 @@ static av_cold int roq_decode_init(AVCodecContext *avctx) s->height = avctx->height; s->last_frame = &s->frames[0]; s->current_frame = &s->frames[1]; - avctx->pix_fmt = PIX_FMT_YUV444P; + avctx->pix_fmt = AV_PIX_FMT_YUV444P; return 0; } diff --git a/libavcodec/roqvideoenc.c b/libavcodec/roqvideoenc.c index f0b8a9807f..a5e0720889 100644 --- a/libavcodec/roqvideoenc.c +++ b/libavcodec/roqvideoenc.c @@ -1081,7 +1081,7 @@ AVCodec ff_roq_encoder = { .encode2 = roq_encode_frame, .close = roq_encode_end, .supported_framerates = (const AVRational[]){ {30,1}, {0,0} }, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV444P, - PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV444P, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("id RoQ video"), }; diff --git a/libavcodec/rpza.c b/libavcodec/rpza.c index 9337360d91..876c9b7f8f 100644 --- a/libavcodec/rpza.c +++ b/libavcodec/rpza.c @@ -236,7 +236,7 @@ static av_cold int rpza_decode_init(AVCodecContext *avctx) RpzaContext *s = avctx->priv_data; s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; s->frame.data[0] = NULL; diff --git a/libavcodec/rv10.c b/libavcodec/rv10.c index e01742ffd5..8769eae209 100644 --- a/libavcodec/rv10.c +++ b/libavcodec/rv10.c @@ -467,7 +467,7 @@ static av_cold int rv10_decode_init(AVCodecContext *avctx) av_log(avctx, AV_LOG_DEBUG, "ver:%X ver0:%X\n", rv->sub_id, avctx->extradata_size >= 4 ? ((uint32_t*)avctx->extradata)[0] : -1); } - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; if (ff_MPV_common_init(s) < 0) return -1; diff --git a/libavcodec/rv10enc.c b/libavcodec/rv10enc.c index 4fd09658d4..d3cd12c3f8 100644 --- a/libavcodec/rv10enc.c +++ b/libavcodec/rv10enc.c @@ -66,7 +66,7 @@ AVCodec ff_rv10_encoder = { .init = ff_MPV_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("RealVideo 1.0"), .priv_class = &rv10_class, }; diff --git a/libavcodec/rv20enc.c b/libavcodec/rv20enc.c index 024e36bee0..c392b53cd6 100644 --- a/libavcodec/rv20enc.c +++ b/libavcodec/rv20enc.c @@ -67,7 +67,7 @@ AVCodec ff_rv20_encoder = { .init = ff_MPV_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("RealVideo 2.0"), .priv_class = &rv20_class, }; diff --git a/libavcodec/rv34.c b/libavcodec/rv34.c index d918392442..ffc8649f12 100644 --- a/libavcodec/rv34.c +++ b/libavcodec/rv34.c @@ -1479,7 +1479,7 @@ av_cold int ff_rv34_decode_init(AVCodecContext *avctx) r->s.avctx = avctx; avctx->flags |= CODEC_FLAG_EMU_EDGE; r->s.flags |= CODEC_FLAG_EMU_EDGE; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; avctx->has_b_frames = 1; s->low_delay = 0; diff --git a/libavcodec/sgidec.c b/libavcodec/sgidec.c index 2702746148..c220452953 100644 --- a/libavcodec/sgidec.c +++ b/libavcodec/sgidec.c @@ -191,11 +191,11 @@ static int decode_frame(AVCodecContext *avctx, } if (s->depth == SGI_GRAYSCALE) { - avctx->pix_fmt = s->bytes_per_channel == 2 ? PIX_FMT_GRAY16BE : PIX_FMT_GRAY8; + avctx->pix_fmt = s->bytes_per_channel == 2 ? AV_PIX_FMT_GRAY16BE : AV_PIX_FMT_GRAY8; } else if (s->depth == SGI_RGB) { - avctx->pix_fmt = s->bytes_per_channel == 2 ? PIX_FMT_RGB48BE : PIX_FMT_RGB24; + avctx->pix_fmt = s->bytes_per_channel == 2 ? AV_PIX_FMT_RGB48BE : AV_PIX_FMT_RGB24; } else if (s->depth == SGI_RGBA && s->bytes_per_channel == 1) { - avctx->pix_fmt = PIX_FMT_RGBA; + avctx->pix_fmt = AV_PIX_FMT_RGBA; } else { av_log(avctx, AV_LOG_ERROR, "wrong picture format\n"); return -1; diff --git a/libavcodec/sgienc.c b/libavcodec/sgienc.c index 49440fa978..2aecc59872 100644 --- a/libavcodec/sgienc.c +++ b/libavcodec/sgienc.c @@ -60,15 +60,15 @@ static int encode_frame(AVCodecContext *avctx, AVPacket *pkt, height = avctx->height; switch (avctx->pix_fmt) { - case PIX_FMT_GRAY8: + case AV_PIX_FMT_GRAY8: dimension = SGI_SINGLE_CHAN; depth = SGI_GRAYSCALE; break; - case PIX_FMT_RGB24: + case AV_PIX_FMT_RGB24: dimension = SGI_MULTI_CHAN; depth = SGI_RGB; break; - case PIX_FMT_RGBA: + case AV_PIX_FMT_RGBA: dimension = SGI_MULTI_CHAN; depth = SGI_RGBA; break; @@ -177,8 +177,8 @@ AVCodec ff_sgi_encoder = { .priv_data_size = sizeof(SgiContext), .init = encode_init, .encode2 = encode_frame, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_RGB24, PIX_FMT_RGBA, PIX_FMT_GRAY8, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_RGB24, AV_PIX_FMT_RGBA, AV_PIX_FMT_GRAY8, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("SGI image"), }; diff --git a/libavcodec/smacker.c b/libavcodec/smacker.c index 72f2ed11a4..df2d4c4b17 100644 --- a/libavcodec/smacker.c +++ b/libavcodec/smacker.c @@ -515,7 +515,7 @@ static av_cold int decode_init(AVCodecContext *avctx) c->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; /* decode huffman trees from extradata */ diff --git a/libavcodec/smc.c b/libavcodec/smc.c index 25b8885ae8..38bf804c8c 100644 --- a/libavcodec/smc.c +++ b/libavcodec/smc.c @@ -414,7 +414,7 @@ static av_cold int smc_decode_init(AVCodecContext *avctx) SmcContext *s = avctx->priv_data; s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; s->frame.data[0] = NULL; diff --git a/libavcodec/snowdec.c b/libavcodec/snowdec.c index 38afdd250a..c4d290478d 100644 --- a/libavcodec/snowdec.c +++ b/libavcodec/snowdec.c @@ -356,7 +356,7 @@ static av_cold int decode_init(AVCodecContext *avctx) { int ret; - avctx->pix_fmt= PIX_FMT_YUV420P; + avctx->pix_fmt= AV_PIX_FMT_YUV420P; if ((ret = ff_snow_common_init(avctx)) < 0) { ff_snow_common_end(avctx->priv_data); diff --git a/libavcodec/snowenc.c b/libavcodec/snowenc.c index 5a4a3ace2b..29f53302b0 100644 --- a/libavcodec/snowenc.c +++ b/libavcodec/snowenc.c @@ -217,15 +217,15 @@ static av_cold int encode_init(AVCodecContext *avctx) avctx->coded_frame= &s->current_picture; switch(avctx->pix_fmt){ -// case PIX_FMT_YUV444P: -// case PIX_FMT_YUV422P: - case PIX_FMT_YUV420P: - case PIX_FMT_GRAY8: -// case PIX_FMT_YUV411P: -// case PIX_FMT_YUV410P: +// case AV_PIX_FMT_YUV444P: +// case AV_PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV420P: + case AV_PIX_FMT_GRAY8: +// case AV_PIX_FMT_YUV411P: +// case AV_PIX_FMT_YUV410P: s->colorspace_type= 0; break; -/* case PIX_FMT_RGB32: +/* case AV_PIX_FMT_RGB32: s->colorspace= 1; break;*/ default: diff --git a/libavcodec/sunrast.c b/libavcodec/sunrast.c index 43c93afeed..a10f449648 100644 --- a/libavcodec/sunrast.c +++ b/libavcodec/sunrast.c @@ -90,13 +90,13 @@ static int sunrast_decode_frame(AVCodecContext *avctx, void *data, switch (depth) { case 1: - avctx->pix_fmt = PIX_FMT_MONOWHITE; + avctx->pix_fmt = AV_PIX_FMT_MONOWHITE; break; case 8: - avctx->pix_fmt = maplength ? PIX_FMT_PAL8 : PIX_FMT_GRAY8; + avctx->pix_fmt = maplength ? AV_PIX_FMT_PAL8 : AV_PIX_FMT_GRAY8; break; case 24: - avctx->pix_fmt = (type == RT_FORMAT_RGB) ? PIX_FMT_RGB24 : PIX_FMT_BGR24; + avctx->pix_fmt = (type == RT_FORMAT_RGB) ? AV_PIX_FMT_RGB24 : AV_PIX_FMT_BGR24; break; default: av_log(avctx, AV_LOG_ERROR, "invalid depth\n"); diff --git a/libavcodec/sunrastenc.c b/libavcodec/sunrastenc.c index 8bcfdf487c..e1b52112d4 100644 --- a/libavcodec/sunrastenc.c +++ b/libavcodec/sunrastenc.c @@ -161,16 +161,16 @@ static av_cold int sunrast_encode_init(AVCodecContext *avctx) s->maplength = 0; switch (avctx->pix_fmt) { - case PIX_FMT_MONOWHITE: + case AV_PIX_FMT_MONOWHITE: s->depth = 1; break; - case PIX_FMT_PAL8 : + case AV_PIX_FMT_PAL8 : s->maptype = RMT_EQUAL_RGB; s->maplength = 3 * 256; - case PIX_FMT_GRAY8: + case AV_PIX_FMT_GRAY8: s->depth = 8; break; - case PIX_FMT_BGR24: + case AV_PIX_FMT_BGR24: s->depth = 24; break; default: @@ -220,10 +220,10 @@ AVCodec ff_sunrast_encoder = { .init = sunrast_encode_init, .encode2 = sunrast_encode_frame, .defaults = sunrast_defaults, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_BGR24, - PIX_FMT_PAL8, - PIX_FMT_GRAY8, - PIX_FMT_MONOWHITE, - PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_BGR24, + AV_PIX_FMT_PAL8, + AV_PIX_FMT_GRAY8, + AV_PIX_FMT_MONOWHITE, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Sun Rasterfile image"), }; diff --git a/libavcodec/svq1dec.c b/libavcodec/svq1dec.c index 4af8b42624..57d67bb05d 100644 --- a/libavcodec/svq1dec.c +++ b/libavcodec/svq1dec.c @@ -714,7 +714,7 @@ static av_cold int svq1_decode_init(AVCodecContext *avctx) s->width = (avctx->width+3)&~3; s->height = (avctx->height+3)&~3; s->codec_id= avctx->codec->id; - avctx->pix_fmt = PIX_FMT_YUV410P; + avctx->pix_fmt = AV_PIX_FMT_YUV410P; avctx->has_b_frames= 1; // not true, but DP frames and these behave like unidirectional b frames s->flags= avctx->flags; if (ff_MPV_common_init(s) < 0) return -1; @@ -774,6 +774,6 @@ AVCodec ff_svq1_decoder = { .decode = svq1_decode_frame, .capabilities = CODEC_CAP_DR1, .flush = ff_mpeg_flush, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV410P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV410P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Sorenson Vector Quantizer 1 / Sorenson Video 1 / SVQ1"), }; diff --git a/libavcodec/svq1enc.c b/libavcodec/svq1enc.c index d1a55bdfd6..da83754d07 100644 --- a/libavcodec/svq1enc.c +++ b/libavcodec/svq1enc.c @@ -512,7 +512,7 @@ static int svq1_encode_frame(AVCodecContext *avctx, AVPacket *pkt, return ret; } - if(avctx->pix_fmt != PIX_FMT_YUV410P){ + if(avctx->pix_fmt != AV_PIX_FMT_YUV410P){ av_log(avctx, AV_LOG_ERROR, "unsupported pixel format\n"); return -1; } @@ -587,6 +587,6 @@ AVCodec ff_svq1_encoder = { .init = svq1_encode_init, .encode2 = svq1_encode_frame, .close = svq1_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV410P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV410P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Sorenson Vector Quantizer 1 / Sorenson Video 1 / SVQ1"), }; diff --git a/libavcodec/svq3.c b/libavcodec/svq3.c index 4d0e282735..c4d5a1bc59 100644 --- a/libavcodec/svq3.c +++ b/libavcodec/svq3.c @@ -1112,5 +1112,5 @@ AVCodec ff_svq3_decoder = { .capabilities = CODEC_CAP_DRAW_HORIZ_BAND | CODEC_CAP_DR1 | CODEC_CAP_DELAY, .long_name = NULL_IF_CONFIG_SMALL("Sorenson Vector Quantizer 3 / Sorenson Video 3 / SVQ3"), - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUVJ420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_NONE }, }; diff --git a/libavcodec/targa.c b/libavcodec/targa.c index aaad2ba2d6..96d18300dd 100644 --- a/libavcodec/targa.c +++ b/libavcodec/targa.c @@ -124,19 +124,19 @@ static int decode_frame(AVCodecContext *avctx, switch(bpp){ case 8: - avctx->pix_fmt = ((compr & (~TGA_RLE)) == TGA_BW) ? PIX_FMT_GRAY8 : PIX_FMT_PAL8; + avctx->pix_fmt = ((compr & (~TGA_RLE)) == TGA_BW) ? AV_PIX_FMT_GRAY8 : AV_PIX_FMT_PAL8; break; case 15: - avctx->pix_fmt = PIX_FMT_RGB555LE; + avctx->pix_fmt = AV_PIX_FMT_RGB555LE; break; case 16: - avctx->pix_fmt = PIX_FMT_RGB555LE; + avctx->pix_fmt = AV_PIX_FMT_RGB555LE; break; case 24: - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; break; case 32: - avctx->pix_fmt = PIX_FMT_BGRA; + avctx->pix_fmt = AV_PIX_FMT_BGRA; break; default: av_log(avctx, AV_LOG_ERROR, "Bit depth %i is not supported\n", bpp); @@ -177,7 +177,7 @@ static int decode_frame(AVCodecContext *avctx, return -1; } pal_size = colors * pal_sample_size; - if(avctx->pix_fmt != PIX_FMT_PAL8)//should not occur but skip palette anyway + if(avctx->pix_fmt != AV_PIX_FMT_PAL8)//should not occur but skip palette anyway bytestream2_skip(&s->gb, pal_size); else{ int t; diff --git a/libavcodec/targaenc.c b/libavcodec/targaenc.c index 1d9b5bd595..e13545f82b 100644 --- a/libavcodec/targaenc.c +++ b/libavcodec/targaenc.c @@ -99,22 +99,22 @@ static int targa_encode_frame(AVCodecContext *avctx, AVPacket *pkt, AV_WL16(pkt->data+12, avctx->width); AV_WL16(pkt->data+14, avctx->height); /* image descriptor byte: origin is always top-left, bits 0-3 specify alpha */ - pkt->data[17] = 0x20 | (avctx->pix_fmt == PIX_FMT_BGRA ? 8 : 0); + pkt->data[17] = 0x20 | (avctx->pix_fmt == AV_PIX_FMT_BGRA ? 8 : 0); switch(avctx->pix_fmt) { - case PIX_FMT_GRAY8: + case AV_PIX_FMT_GRAY8: pkt->data[2] = TGA_BW; /* uncompressed grayscale image */ pkt->data[16] = 8; /* bpp */ break; - case PIX_FMT_RGB555LE: + case AV_PIX_FMT_RGB555LE: pkt->data[2] = TGA_RGB; /* uncompresses true-color image */ pkt->data[16] = 16; /* bpp */ break; - case PIX_FMT_BGR24: + case AV_PIX_FMT_BGR24: pkt->data[2] = TGA_RGB; /* uncompressed true-color image */ pkt->data[16] = 24; /* bpp */ break; - case PIX_FMT_BGRA: + case AV_PIX_FMT_BGRA: pkt->data[2] = TGA_RGB; /* uncompressed true-color image */ pkt->data[16] = 32; /* bpp */ break; @@ -171,9 +171,9 @@ AVCodec ff_targa_encoder = { .priv_data_size = sizeof(TargaContext), .init = targa_encode_init, .encode2 = targa_encode_frame, - .pix_fmts = (const enum PixelFormat[]){ - PIX_FMT_BGR24, PIX_FMT_BGRA, PIX_FMT_RGB555LE, PIX_FMT_GRAY8, - PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]){ + AV_PIX_FMT_BGR24, AV_PIX_FMT_BGRA, AV_PIX_FMT_RGB555LE, AV_PIX_FMT_GRAY8, + AV_PIX_FMT_NONE }, .long_name= NULL_IF_CONFIG_SMALL("Truevision Targa image"), }; diff --git a/libavcodec/tiertexseqv.c b/libavcodec/tiertexseqv.c index c488b8beb8..98f6477b76 100644 --- a/libavcodec/tiertexseqv.c +++ b/libavcodec/tiertexseqv.c @@ -214,7 +214,7 @@ static av_cold int seqvideo_decode_init(AVCodecContext *avctx) SeqVideoContext *seq = avctx->priv_data; seq->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; seq->frame.data[0] = NULL; diff --git a/libavcodec/tiff.c b/libavcodec/tiff.c index 2a48050678..a15095116c 100644 --- a/libavcodec/tiff.c +++ b/libavcodec/tiff.c @@ -250,22 +250,22 @@ static int init_image(TiffContext *s) switch (s->bpp * 10 + s->bppcount) { case 11: - s->avctx->pix_fmt = PIX_FMT_MONOBLACK; + s->avctx->pix_fmt = AV_PIX_FMT_MONOBLACK; break; case 81: - s->avctx->pix_fmt = PIX_FMT_PAL8; + s->avctx->pix_fmt = AV_PIX_FMT_PAL8; break; case 243: - s->avctx->pix_fmt = PIX_FMT_RGB24; + s->avctx->pix_fmt = AV_PIX_FMT_RGB24; break; case 161: - s->avctx->pix_fmt = s->le ? PIX_FMT_GRAY16LE : PIX_FMT_GRAY16BE; + s->avctx->pix_fmt = s->le ? AV_PIX_FMT_GRAY16LE : AV_PIX_FMT_GRAY16BE; break; case 324: - s->avctx->pix_fmt = PIX_FMT_RGBA; + s->avctx->pix_fmt = AV_PIX_FMT_RGBA; break; case 483: - s->avctx->pix_fmt = s->le ? PIX_FMT_RGB48LE : PIX_FMT_RGB48BE; + s->avctx->pix_fmt = s->le ? AV_PIX_FMT_RGB48LE : AV_PIX_FMT_RGB48BE; break; default: av_log(s->avctx, AV_LOG_ERROR, @@ -284,7 +284,7 @@ static int init_image(TiffContext *s) av_log(s->avctx, AV_LOG_ERROR, "get_buffer() failed\n"); return ret; } - if (s->avctx->pix_fmt == PIX_FMT_PAL8) { + if (s->avctx->pix_fmt == AV_PIX_FMT_PAL8) { if (s->palette_is_set) { memcpy(s->picture.data[1], s->palette, sizeof(s->palette)); } else { diff --git a/libavcodec/tiffenc.c b/libavcodec/tiffenc.c index e6fd6cf69a..9a3f17630d 100644 --- a/libavcodec/tiffenc.c +++ b/libavcodec/tiffenc.c @@ -235,11 +235,11 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, s->subsampling[1] = 1; switch (avctx->pix_fmt) { - case PIX_FMT_RGB48LE: - case PIX_FMT_GRAY16LE: - case PIX_FMT_RGB24: - case PIX_FMT_GRAY8: - case PIX_FMT_PAL8: + case AV_PIX_FMT_RGB48LE: + case AV_PIX_FMT_GRAY16LE: + case AV_PIX_FMT_RGB24: + case AV_PIX_FMT_GRAY8: + case AV_PIX_FMT_PAL8: pfd = &av_pix_fmt_descriptors[avctx->pix_fmt]; s->bpp = av_get_bits_per_pixel(pfd); if (pfd->flags & PIX_FMT_PAL) { @@ -254,21 +254,21 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, bpp_tab[i] = s->bpp / s->bpp_tab_size; } break; - case PIX_FMT_MONOBLACK: + case AV_PIX_FMT_MONOBLACK: s->bpp = 1; s->photometric_interpretation = 1; s->bpp_tab_size = 0; break; - case PIX_FMT_MONOWHITE: + case AV_PIX_FMT_MONOWHITE: s->bpp = 1; s->photometric_interpretation = 0; s->bpp_tab_size = 0; break; - case PIX_FMT_YUV420P: - case PIX_FMT_YUV422P: - case PIX_FMT_YUV444P: - case PIX_FMT_YUV410P: - case PIX_FMT_YUV411P: + case AV_PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV444P: + case AV_PIX_FMT_YUV410P: + case AV_PIX_FMT_YUV411P: s->photometric_interpretation = 6; avcodec_get_chroma_sub_sample(avctx->pix_fmt, &shift_h, &shift_v); @@ -433,7 +433,7 @@ static int encode_frame(AVCodecContext * avctx, AVPacket *pkt, add_entry(s, TIFF_SOFTWARE_NAME, TIFF_STRING, strlen(LIBAVCODEC_IDENT) + 1, LIBAVCODEC_IDENT); - if (avctx->pix_fmt == PIX_FMT_PAL8) { + if (avctx->pix_fmt == AV_PIX_FMT_PAL8) { uint16_t pal[256 * 3]; for (i = 0; i < 256; i++) { uint32_t rgb = *(uint32_t *) (p->data[1] + i * 4); @@ -496,13 +496,13 @@ AVCodec ff_tiff_encoder = { .id = AV_CODEC_ID_TIFF, .priv_data_size = sizeof(TiffEncoderContext), .encode2 = encode_frame, - .pix_fmts = (const enum PixelFormat[]) { - PIX_FMT_RGB24, PIX_FMT_RGB48LE, PIX_FMT_PAL8, - PIX_FMT_GRAY8, PIX_FMT_GRAY16LE, - PIX_FMT_MONOBLACK, PIX_FMT_MONOWHITE, - PIX_FMT_YUV420P, PIX_FMT_YUV422P, PIX_FMT_YUV444P, - PIX_FMT_YUV410P, PIX_FMT_YUV411P, - PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]) { + AV_PIX_FMT_RGB24, AV_PIX_FMT_RGB48LE, AV_PIX_FMT_PAL8, + AV_PIX_FMT_GRAY8, AV_PIX_FMT_GRAY16LE, + AV_PIX_FMT_MONOBLACK, AV_PIX_FMT_MONOWHITE, + AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV444P, + AV_PIX_FMT_YUV410P, AV_PIX_FMT_YUV411P, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("TIFF image"), .priv_class = &tiffenc_class, diff --git a/libavcodec/tmv.c b/libavcodec/tmv.c index 396dff0386..8533707f8b 100644 --- a/libavcodec/tmv.c +++ b/libavcodec/tmv.c @@ -87,7 +87,7 @@ static int tmv_decode_frame(AVCodecContext *avctx, void *data, static av_cold int tmv_decode_init(AVCodecContext *avctx) { - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/truemotion1.c b/libavcodec/truemotion1.c index 6525ddca24..c2e15894e4 100644 --- a/libavcodec/truemotion1.c +++ b/libavcodec/truemotion1.c @@ -390,10 +390,10 @@ static int truemotion1_decode_header(TrueMotion1Context *s) } if (compression_types[header.compression].algorithm == ALGO_RGB24H) { - new_pix_fmt = PIX_FMT_RGB32; + new_pix_fmt = AV_PIX_FMT_RGB32; width_shift = 1; } else - new_pix_fmt = PIX_FMT_RGB555; // RGB565 is supported as well + new_pix_fmt = AV_PIX_FMT_RGB555; // RGB565 is supported as well s->w >>= width_shift; if (av_image_check_size(s->w, s->h, 0, s->avctx) < 0) @@ -419,7 +419,7 @@ static int truemotion1_decode_header(TrueMotion1Context *s) if (compression_types[header.compression].algorithm == ALGO_RGB24H) gen_vector_table24(s, sel_vector_table); else - if (s->avctx->pix_fmt == PIX_FMT_RGB555) + if (s->avctx->pix_fmt == AV_PIX_FMT_RGB555) gen_vector_table15(s, sel_vector_table); else gen_vector_table16(s, sel_vector_table); @@ -464,9 +464,9 @@ static av_cold int truemotion1_decode_init(AVCodecContext *avctx) // FIXME: it may change ? // if (avctx->bits_per_sample == 24) -// avctx->pix_fmt = PIX_FMT_RGB24; +// avctx->pix_fmt = AV_PIX_FMT_RGB24; // else -// avctx->pix_fmt = PIX_FMT_RGB555; +// avctx->pix_fmt = AV_PIX_FMT_RGB555; s->frame.data[0] = NULL; diff --git a/libavcodec/truemotion2.c b/libavcodec/truemotion2.c index ad32612d04..b540eda3f6 100644 --- a/libavcodec/truemotion2.c +++ b/libavcodec/truemotion2.c @@ -879,7 +879,7 @@ static av_cold int decode_init(AVCodecContext *avctx){ l->avctx = avctx; l->pic.data[0]=NULL; - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; ff_dsputil_init(&l->dsp, avctx); diff --git a/libavcodec/tscc.c b/libavcodec/tscc.c index b6fdf320fc..6c40e6cfe7 100644 --- a/libavcodec/tscc.c +++ b/libavcodec/tscc.c @@ -113,7 +113,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, AVPac } /* make the palette available on the way out */ - if (c->avctx->pix_fmt == PIX_FMT_PAL8) { + if (c->avctx->pix_fmt == AV_PIX_FMT_PAL8) { const uint8_t *pal = av_packet_get_side_data(avpkt, AV_PKT_DATA_PALETTE, NULL); if (pal) { @@ -149,12 +149,12 @@ static av_cold int decode_init(AVCodecContext *avctx) // Needed if zlib unused or init aborted before inflateInit memset(&c->zstream, 0, sizeof(z_stream)); switch(avctx->bits_per_coded_sample){ - case 8: avctx->pix_fmt = PIX_FMT_PAL8; break; - case 16: avctx->pix_fmt = PIX_FMT_RGB555; break; + case 8: avctx->pix_fmt = AV_PIX_FMT_PAL8; break; + case 16: avctx->pix_fmt = AV_PIX_FMT_RGB555; break; case 24: - avctx->pix_fmt = PIX_FMT_BGR24; + avctx->pix_fmt = AV_PIX_FMT_BGR24; break; - case 32: avctx->pix_fmt = PIX_FMT_RGB32; break; + case 32: avctx->pix_fmt = AV_PIX_FMT_RGB32; break; default: av_log(avctx, AV_LOG_ERROR, "Camtasia error: unknown depth %i bpp\n", avctx->bits_per_coded_sample); return -1; } diff --git a/libavcodec/tscc2.c b/libavcodec/tscc2.c index 363f5556cf..1b2a3a4990 100644 --- a/libavcodec/tscc2.c +++ b/libavcodec/tscc2.c @@ -336,7 +336,7 @@ static av_cold int tscc2_decode_init(AVCodecContext *avctx) c->avctx = avctx; - avctx->pix_fmt = PIX_FMT_YUV444P; + avctx->pix_fmt = AV_PIX_FMT_YUV444P; if ((ret = init_vlcs(c)) < 0) { av_log(avctx, AV_LOG_ERROR, "Cannot initialise VLCs\n"); diff --git a/libavcodec/txd.c b/libavcodec/txd.c index 0eb7aaa896..67c8a093f7 100644 --- a/libavcodec/txd.c +++ b/libavcodec/txd.c @@ -68,9 +68,9 @@ static int txd_decode_frame(AVCodecContext *avctx, void *data, int *data_size, } if (depth == 8) { - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; } else if (depth == 16 || depth == 32) { - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; } else { av_log(avctx, AV_LOG_ERROR, "depth of %i is unsupported\n", depth); return -1; diff --git a/libavcodec/ulti.c b/libavcodec/ulti.c index 0431159109..41faa3de97 100644 --- a/libavcodec/ulti.c +++ b/libavcodec/ulti.c @@ -49,7 +49,7 @@ static av_cold int ulti_decode_init(AVCodecContext *avctx) s->width = avctx->width; s->height = avctx->height; s->blocks = (s->width / 8) * (s->height / 8); - avctx->pix_fmt = PIX_FMT_YUV410P; + avctx->pix_fmt = AV_PIX_FMT_YUV410P; avctx->coded_frame = &s->frame; s->ulti_codebook = ulti_codebook; diff --git a/libavcodec/utils.c b/libavcodec/utils.c index d1b9d0a3ee..c71ddbe802 100644 --- a/libavcodec/utils.c +++ b/libavcodec/utils.c @@ -165,64 +165,64 @@ void avcodec_align_dimensions2(AVCodecContext *s, int *width, int *height, int h_align = 1; switch (s->pix_fmt) { - case PIX_FMT_YUV420P: - case PIX_FMT_YUYV422: - case PIX_FMT_UYVY422: - case PIX_FMT_YUV422P: - case PIX_FMT_YUV440P: - case PIX_FMT_YUV444P: - case PIX_FMT_GBRP: - case PIX_FMT_GRAY8: - case PIX_FMT_GRAY16BE: - case PIX_FMT_GRAY16LE: - case PIX_FMT_YUVJ420P: - case PIX_FMT_YUVJ422P: - case PIX_FMT_YUVJ440P: - case PIX_FMT_YUVJ444P: - case PIX_FMT_YUVA420P: - case PIX_FMT_YUV420P9LE: - case PIX_FMT_YUV420P9BE: - case PIX_FMT_YUV420P10LE: - case PIX_FMT_YUV420P10BE: - case PIX_FMT_YUV422P9LE: - case PIX_FMT_YUV422P9BE: - case PIX_FMT_YUV422P10LE: - case PIX_FMT_YUV422P10BE: - case PIX_FMT_YUV444P9LE: - case PIX_FMT_YUV444P9BE: - case PIX_FMT_YUV444P10LE: - case PIX_FMT_YUV444P10BE: - case PIX_FMT_GBRP9LE: - case PIX_FMT_GBRP9BE: - case PIX_FMT_GBRP10LE: - case PIX_FMT_GBRP10BE: + case AV_PIX_FMT_YUV420P: + case AV_PIX_FMT_YUYV422: + case AV_PIX_FMT_UYVY422: + case AV_PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV440P: + case AV_PIX_FMT_YUV444P: + case AV_PIX_FMT_GBRP: + case AV_PIX_FMT_GRAY8: + case AV_PIX_FMT_GRAY16BE: + case AV_PIX_FMT_GRAY16LE: + case AV_PIX_FMT_YUVJ420P: + case AV_PIX_FMT_YUVJ422P: + case AV_PIX_FMT_YUVJ440P: + case AV_PIX_FMT_YUVJ444P: + case AV_PIX_FMT_YUVA420P: + case AV_PIX_FMT_YUV420P9LE: + case AV_PIX_FMT_YUV420P9BE: + case AV_PIX_FMT_YUV420P10LE: + case AV_PIX_FMT_YUV420P10BE: + case AV_PIX_FMT_YUV422P9LE: + case AV_PIX_FMT_YUV422P9BE: + case AV_PIX_FMT_YUV422P10LE: + case AV_PIX_FMT_YUV422P10BE: + case AV_PIX_FMT_YUV444P9LE: + case AV_PIX_FMT_YUV444P9BE: + case AV_PIX_FMT_YUV444P10LE: + case AV_PIX_FMT_YUV444P10BE: + case AV_PIX_FMT_GBRP9LE: + case AV_PIX_FMT_GBRP9BE: + case AV_PIX_FMT_GBRP10LE: + case AV_PIX_FMT_GBRP10BE: w_align = 16; //FIXME assume 16 pixel per macroblock h_align = 16 * 2; // interlaced needs 2 macroblocks height break; - case PIX_FMT_YUV411P: - case PIX_FMT_UYYVYY411: + case AV_PIX_FMT_YUV411P: + case AV_PIX_FMT_UYYVYY411: w_align = 32; h_align = 8; break; - case PIX_FMT_YUV410P: + case AV_PIX_FMT_YUV410P: if (s->codec_id == AV_CODEC_ID_SVQ1) { w_align = 64; h_align = 64; } - case PIX_FMT_RGB555: + case AV_PIX_FMT_RGB555: if (s->codec_id == AV_CODEC_ID_RPZA) { w_align = 4; h_align = 4; } - case PIX_FMT_PAL8: - case PIX_FMT_BGR8: - case PIX_FMT_RGB8: + case AV_PIX_FMT_PAL8: + case AV_PIX_FMT_BGR8: + case AV_PIX_FMT_RGB8: if (s->codec_id == AV_CODEC_ID_SMC) { w_align = 4; h_align = 4; } break; - case PIX_FMT_BGR24: + case AV_PIX_FMT_BGR24: if ((s->codec_id == AV_CODEC_ID_MSZH) || (s->codec_id == AV_CODEC_ID_ZLIB)) { w_align = 4; @@ -624,9 +624,9 @@ int avcodec_default_execute2(AVCodecContext *c, int (*func)(AVCodecContext *c2, return 0; } -enum PixelFormat avcodec_default_get_format(struct AVCodecContext *s, const enum PixelFormat *fmt) +enum AVPixelFormat avcodec_default_get_format(struct AVCodecContext *s, const enum AVPixelFormat *fmt) { - while (*fmt != PIX_FMT_NONE && ff_is_hwaccel_pix_fmt(*fmt)) + while (*fmt != AV_PIX_FMT_NONE && ff_is_hwaccel_pix_fmt(*fmt)) ++fmt; return fmt[0]; } @@ -809,10 +809,10 @@ int attribute_align_arg avcodec_open2(AVCodecContext *avctx, const AVCodec *code } } if (avctx->codec->pix_fmts) { - for (i = 0; avctx->codec->pix_fmts[i] != PIX_FMT_NONE; i++) + for (i = 0; avctx->codec->pix_fmts[i] != AV_PIX_FMT_NONE; i++) if (avctx->pix_fmt == avctx->codec->pix_fmts[i]) break; - if (avctx->codec->pix_fmts[i] == PIX_FMT_NONE) { + if (avctx->codec->pix_fmts[i] == AV_PIX_FMT_NONE) { av_log(avctx, AV_LOG_ERROR, "Specified pix_fmt is not supported\n"); ret = AVERROR(EINVAL); goto free_and_end; @@ -1621,7 +1621,7 @@ void avcodec_string(char *buf, int buf_size, AVCodecContext *enc, int encode) if (profile) snprintf(buf + strlen(buf), buf_size - strlen(buf), " (%s)", profile); - if (enc->pix_fmt != PIX_FMT_NONE) { + if (enc->pix_fmt != AV_PIX_FMT_NONE) { snprintf(buf + strlen(buf), buf_size - strlen(buf), ", %s", av_get_pix_fmt_name(enc->pix_fmt)); @@ -2067,7 +2067,7 @@ AVHWAccel *av_hwaccel_next(AVHWAccel *hwaccel) return hwaccel ? hwaccel->next : first_hwaccel; } -AVHWAccel *ff_find_hwaccel(enum AVCodecID codec_id, enum PixelFormat pix_fmt) +AVHWAccel *ff_find_hwaccel(enum AVCodecID codec_id, enum AVPixelFormat pix_fmt) { AVHWAccel *hwaccel = NULL; diff --git a/libavcodec/utvideodec.c b/libavcodec/utvideodec.c index 7ae2573140..c35a5692fe 100644 --- a/libavcodec/utvideodec.c +++ b/libavcodec/utvideodec.c @@ -86,7 +86,7 @@ static int decode_plane(UtvideoContext *c, int plane_no, VLC vlc; GetBitContext gb; int prev, fsym; - const int cmask = ~(!plane_no && c->avctx->pix_fmt == PIX_FMT_YUV420P); + const int cmask = ~(!plane_no && c->avctx->pix_fmt == AV_PIX_FMT_YUV420P); if (build_huff(src, &vlc, &fsym)) { av_log(c->avctx, AV_LOG_ERROR, "Cannot build Huffman codes\n"); @@ -393,8 +393,8 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, } switch (c->avctx->pix_fmt) { - case PIX_FMT_RGB24: - case PIX_FMT_RGBA: + case AV_PIX_FMT_RGB24: + case AV_PIX_FMT_RGBA: for (i = 0; i < c->planes; i++) { ret = decode_plane(c, i, c->pic.data[0] + ff_ut_rgb_order[i], c->planes, c->pic.linesize[0], avctx->width, @@ -418,7 +418,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, restore_rgb_planes(c->pic.data[0], c->planes, c->pic.linesize[0], avctx->width, avctx->height); break; - case PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV420P: for (i = 0; i < 3; i++) { ret = decode_plane(c, i, c->pic.data[i], 1, c->pic.linesize[i], avctx->width >> !!i, avctx->height >> !!i, @@ -439,7 +439,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *data_size, } } break; - case PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV422P: for (i = 0; i < 3; i++) { ret = decode_plane(c, i, c->pic.data[i], 1, c->pic.linesize[i], avctx->width >> !!i, avctx->height, @@ -507,19 +507,19 @@ static av_cold int decode_init(AVCodecContext *avctx) switch (avctx->codec_tag) { case MKTAG('U', 'L', 'R', 'G'): c->planes = 3; - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; break; case MKTAG('U', 'L', 'R', 'A'): c->planes = 4; - avctx->pix_fmt = PIX_FMT_RGBA; + avctx->pix_fmt = AV_PIX_FMT_RGBA; break; case MKTAG('U', 'L', 'Y', '0'): c->planes = 3; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; break; case MKTAG('U', 'L', 'Y', '2'): c->planes = 3; - avctx->pix_fmt = PIX_FMT_YUV422P; + avctx->pix_fmt = AV_PIX_FMT_YUV422P; break; default: av_log(avctx, AV_LOG_ERROR, "Unknown Ut Video FOURCC provided (%08X)\n", diff --git a/libavcodec/utvideoenc.c b/libavcodec/utvideoenc.c index 18bdec56a4..085c415caa 100644 --- a/libavcodec/utvideoenc.c +++ b/libavcodec/utvideoenc.c @@ -65,17 +65,17 @@ static av_cold int utvideo_encode_init(AVCodecContext *avctx) c->slice_stride = FFALIGN(avctx->width, 32); switch (avctx->pix_fmt) { - case PIX_FMT_RGB24: + case AV_PIX_FMT_RGB24: c->planes = 3; avctx->codec_tag = MKTAG('U', 'L', 'R', 'G'); original_format = UTVIDEO_RGB; break; - case PIX_FMT_RGBA: + case AV_PIX_FMT_RGBA: c->planes = 4; avctx->codec_tag = MKTAG('U', 'L', 'R', 'A'); original_format = UTVIDEO_RGBA; break; - case PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV420P: if (avctx->width & 1 || avctx->height & 1) { av_log(avctx, AV_LOG_ERROR, "4:2:0 video requires even width and height.\n"); @@ -85,7 +85,7 @@ static av_cold int utvideo_encode_init(AVCodecContext *avctx) avctx->codec_tag = MKTAG('U', 'L', 'Y', '0'); original_format = UTVIDEO_420; break; - case PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV422P: if (avctx->width & 1) { av_log(avctx, AV_LOG_ERROR, "4:2:2 video requires even width.\n"); @@ -538,14 +538,14 @@ static int utvideo_encode_frame(AVCodecContext *avctx, AVPacket *pkt, } /* In case of RGB, mangle the planes to Ut Video's format */ - if (avctx->pix_fmt == PIX_FMT_RGBA || avctx->pix_fmt == PIX_FMT_RGB24) + if (avctx->pix_fmt == AV_PIX_FMT_RGBA || avctx->pix_fmt == AV_PIX_FMT_RGB24) mangle_rgb_planes(c->slice_buffer, c->slice_stride, pic->data[0], c->planes, pic->linesize[0], width, height); /* Deal with the planes */ switch (avctx->pix_fmt) { - case PIX_FMT_RGB24: - case PIX_FMT_RGBA: + case AV_PIX_FMT_RGB24: + case AV_PIX_FMT_RGBA: for (i = 0; i < c->planes; i++) { ret = encode_plane(avctx, c->slice_buffer[i] + 2 * c->slice_stride, c->slice_buffer[i], c->slice_stride, @@ -557,7 +557,7 @@ static int utvideo_encode_frame(AVCodecContext *avctx, AVPacket *pkt, } } break; - case PIX_FMT_YUV422P: + case AV_PIX_FMT_YUV422P: for (i = 0; i < c->planes; i++) { ret = encode_plane(avctx, pic->data[i], c->slice_buffer[0], pic->linesize[i], width >> !!i, height, &pb); @@ -568,7 +568,7 @@ static int utvideo_encode_frame(AVCodecContext *avctx, AVPacket *pkt, } } break; - case PIX_FMT_YUV420P: + case AV_PIX_FMT_YUV420P: for (i = 0; i < c->planes; i++) { ret = encode_plane(avctx, pic->data[i], c->slice_buffer[0], pic->linesize[i], width >> !!i, height >> !!i, @@ -619,9 +619,9 @@ AVCodec ff_utvideo_encoder = { .init = utvideo_encode_init, .encode2 = utvideo_encode_frame, .close = utvideo_encode_close, - .pix_fmts = (const enum PixelFormat[]) { - PIX_FMT_RGB24, PIX_FMT_RGBA, PIX_FMT_YUV422P, - PIX_FMT_YUV420P, PIX_FMT_NONE + .pix_fmts = (const enum AVPixelFormat[]) { + AV_PIX_FMT_RGB24, AV_PIX_FMT_RGBA, AV_PIX_FMT_YUV422P, + AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Ut Video"), }; diff --git a/libavcodec/v210dec.c b/libavcodec/v210dec.c index dad861878a..aa08436d75 100644 --- a/libavcodec/v210dec.c +++ b/libavcodec/v210dec.c @@ -32,7 +32,7 @@ static av_cold int decode_init(AVCodecContext *avctx) av_log(avctx, AV_LOG_ERROR, "v210 needs even width\n"); return -1; } - avctx->pix_fmt = PIX_FMT_YUV422P10; + avctx->pix_fmt = AV_PIX_FMT_YUV422P10; avctx->bits_per_raw_sample = 10; avctx->coded_frame = avcodec_alloc_frame(); diff --git a/libavcodec/v210enc.c b/libavcodec/v210enc.c index fd5129f40d..ad8a4a537b 100644 --- a/libavcodec/v210enc.c +++ b/libavcodec/v210enc.c @@ -123,6 +123,6 @@ AVCodec ff_v210_encoder = { .init = encode_init, .encode2 = encode_frame, .close = encode_close, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV422P10, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV422P10, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Uncompressed 4:2:2 10-bit"), }; diff --git a/libavcodec/v210x.c b/libavcodec/v210x.c index 9fd4f7aed8..4aad304260 100644 --- a/libavcodec/v210x.c +++ b/libavcodec/v210x.c @@ -29,7 +29,7 @@ static av_cold int decode_init(AVCodecContext *avctx) av_log(avctx, AV_LOG_ERROR, "v210x needs even width\n"); return -1; } - avctx->pix_fmt = PIX_FMT_YUV422P16; + avctx->pix_fmt = AV_PIX_FMT_YUV422P16; avctx->bits_per_raw_sample= 10; avctx->coded_frame= avcodec_alloc_frame(); diff --git a/libavcodec/v410dec.c b/libavcodec/v410dec.c index 09d09d04ea..b96ba11fbd 100644 --- a/libavcodec/v410dec.c +++ b/libavcodec/v410dec.c @@ -26,7 +26,7 @@ static av_cold int v410_decode_init(AVCodecContext *avctx) { - avctx->pix_fmt = PIX_FMT_YUV444P10; + avctx->pix_fmt = AV_PIX_FMT_YUV444P10; avctx->bits_per_raw_sample = 10; if (avctx->width & 1) { diff --git a/libavcodec/v410enc.c b/libavcodec/v410enc.c index fce1f8bbc1..cc7cef7ea2 100644 --- a/libavcodec/v410enc.c +++ b/libavcodec/v410enc.c @@ -96,6 +96,6 @@ AVCodec ff_v410_encoder = { .init = v410_encode_init, .encode2 = v410_encode_frame, .close = v410_encode_close, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV444P10, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV444P10, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Uncompressed 4:4:4 10-bit"), }; diff --git a/libavcodec/vaapi_h264.c b/libavcodec/vaapi_h264.c index 3dafac51ae..4ffc7d882e 100644 --- a/libavcodec/vaapi_h264.c +++ b/libavcodec/vaapi_h264.c @@ -340,7 +340,7 @@ AVHWAccel ff_h264_vaapi_hwaccel = { .name = "h264_vaapi", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_H264, - .pix_fmt = PIX_FMT_VAAPI_VLD, + .pix_fmt = AV_PIX_FMT_VAAPI_VLD, .start_frame = start_frame, .end_frame = end_frame, .decode_slice = decode_slice, diff --git a/libavcodec/vaapi_mpeg2.c b/libavcodec/vaapi_mpeg2.c index d0de0f7f0e..cfe5d3abe0 100644 --- a/libavcodec/vaapi_mpeg2.c +++ b/libavcodec/vaapi_mpeg2.c @@ -142,7 +142,7 @@ AVHWAccel ff_mpeg2_vaapi_hwaccel = { .name = "mpeg2_vaapi", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_MPEG2VIDEO, - .pix_fmt = PIX_FMT_VAAPI_VLD, + .pix_fmt = AV_PIX_FMT_VAAPI_VLD, .start_frame = vaapi_mpeg2_start_frame, .end_frame = vaapi_mpeg2_end_frame, .decode_slice = vaapi_mpeg2_decode_slice, diff --git a/libavcodec/vaapi_mpeg4.c b/libavcodec/vaapi_mpeg4.c index 9f7d14603b..7d9ffd7fa2 100644 --- a/libavcodec/vaapi_mpeg4.c +++ b/libavcodec/vaapi_mpeg4.c @@ -154,7 +154,7 @@ AVHWAccel ff_mpeg4_vaapi_hwaccel = { .name = "mpeg4_vaapi", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_MPEG4, - .pix_fmt = PIX_FMT_VAAPI_VLD, + .pix_fmt = AV_PIX_FMT_VAAPI_VLD, .start_frame = vaapi_mpeg4_start_frame, .end_frame = vaapi_mpeg4_end_frame, .decode_slice = vaapi_mpeg4_decode_slice, @@ -166,7 +166,7 @@ AVHWAccel ff_h263_vaapi_hwaccel = { .name = "h263_vaapi", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_H263, - .pix_fmt = PIX_FMT_VAAPI_VLD, + .pix_fmt = AV_PIX_FMT_VAAPI_VLD, .start_frame = vaapi_mpeg4_start_frame, .end_frame = vaapi_mpeg4_end_frame, .decode_slice = vaapi_mpeg4_decode_slice, diff --git a/libavcodec/vaapi_vc1.c b/libavcodec/vaapi_vc1.c index e5bc6f46cc..4a98ba7e15 100644 --- a/libavcodec/vaapi_vc1.c +++ b/libavcodec/vaapi_vc1.c @@ -345,7 +345,7 @@ AVHWAccel ff_wmv3_vaapi_hwaccel = { .name = "wmv3_vaapi", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_WMV3, - .pix_fmt = PIX_FMT_VAAPI_VLD, + .pix_fmt = AV_PIX_FMT_VAAPI_VLD, .start_frame = vaapi_vc1_start_frame, .end_frame = vaapi_vc1_end_frame, .decode_slice = vaapi_vc1_decode_slice, @@ -356,7 +356,7 @@ AVHWAccel ff_vc1_vaapi_hwaccel = { .name = "vc1_vaapi", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_VC1, - .pix_fmt = PIX_FMT_VAAPI_VLD, + .pix_fmt = AV_PIX_FMT_VAAPI_VLD, .start_frame = vaapi_vc1_start_frame, .end_frame = vaapi_vc1_end_frame, .decode_slice = vaapi_vc1_decode_slice, diff --git a/libavcodec/vb.c b/libavcodec/vb.c index d84328e774..39e14a008d 100644 --- a/libavcodec/vb.c +++ b/libavcodec/vb.c @@ -246,7 +246,7 @@ static av_cold int decode_init(AVCodecContext *avctx) VBDecContext * const c = avctx->priv_data; c->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; c->frame = av_mallocz(avctx->width * avctx->height); c->prev_frame = av_mallocz(avctx->width * avctx->height); diff --git a/libavcodec/vble.c b/libavcodec/vble.c index 11b4aaabc1..9726b1089a 100644 --- a/libavcodec/vble.c +++ b/libavcodec/vble.c @@ -193,7 +193,7 @@ static av_cold int vble_decode_init(AVCodecContext *avctx) ctx->avctx = avctx; ff_dsputil_init(&ctx->dsp, avctx); - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; avctx->bits_per_raw_sample = 8; avctx->coded_frame = avcodec_alloc_frame(); diff --git a/libavcodec/vc1dec.c b/libavcodec/vc1dec.c index 310172a84d..df3a55a0ba 100644 --- a/libavcodec/vc1dec.c +++ b/libavcodec/vc1dec.c @@ -5149,7 +5149,7 @@ static av_cold int vc1_decode_init(AVCodecContext *avctx) if (!(avctx->flags & CODEC_FLAG_GRAY)) avctx->pix_fmt = avctx->get_format(avctx, avctx->codec->pix_fmts); else - avctx->pix_fmt = PIX_FMT_GRAY8; + avctx->pix_fmt = AV_PIX_FMT_GRAY8; avctx->hwaccel = ff_find_hwaccel(avctx->codec->id, avctx->pix_fmt); v->s.avctx = avctx; avctx->flags |= CODEC_FLAG_EMU_EDGE; @@ -5336,9 +5336,9 @@ static int vc1_decode_frame(AVCodecContext *avctx, void *data, if (s->avctx->codec->capabilities&CODEC_CAP_HWACCEL_VDPAU) { if (v->profile < PROFILE_ADVANCED) - avctx->pix_fmt = PIX_FMT_VDPAU_WMV3; + avctx->pix_fmt = AV_PIX_FMT_VDPAU_WMV3; else - avctx->pix_fmt = PIX_FMT_VDPAU_VC1; + avctx->pix_fmt = AV_PIX_FMT_VDPAU_VC1; } //for advanced profile we may need to parse and unescape data @@ -5722,7 +5722,7 @@ AVCodec ff_wmv3_vdpau_decoder = { .decode = vc1_decode_frame, .capabilities = CODEC_CAP_DR1 | CODEC_CAP_DELAY | CODEC_CAP_HWACCEL_VDPAU, .long_name = NULL_IF_CONFIG_SMALL("Windows Media Video 9 VDPAU"), - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_VDPAU_WMV3, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_VDPAU_WMV3, AV_PIX_FMT_NONE }, .profiles = NULL_IF_CONFIG_SMALL(profiles) }; #endif @@ -5738,7 +5738,7 @@ AVCodec ff_vc1_vdpau_decoder = { .decode = vc1_decode_frame, .capabilities = CODEC_CAP_DR1 | CODEC_CAP_DELAY | CODEC_CAP_HWACCEL_VDPAU, .long_name = NULL_IF_CONFIG_SMALL("SMPTE VC-1 VDPAU"), - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_VDPAU_VC1, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_VDPAU_VC1, AV_PIX_FMT_NONE }, .profiles = NULL_IF_CONFIG_SMALL(profiles) }; #endif diff --git a/libavcodec/vcr1.c b/libavcodec/vcr1.c index cc742f0007..df5088c9b8 100644 --- a/libavcodec/vcr1.c +++ b/libavcodec/vcr1.c @@ -47,7 +47,7 @@ static av_cold int vcr1_decode_init(AVCodecContext *avctx) { vcr1_common_init(avctx); - avctx->pix_fmt = PIX_FMT_YUV410P; + avctx->pix_fmt = AV_PIX_FMT_YUV410P; return 0; } diff --git a/libavcodec/vda_h264.c b/libavcodec/vda_h264.c index 5c8f5679f9..2a78aac61a 100644 --- a/libavcodec/vda_h264.c +++ b/libavcodec/vda_h264.c @@ -371,7 +371,7 @@ AVHWAccel ff_h264_vda_hwaccel = { .name = "h264_vda", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_H264, - .pix_fmt = PIX_FMT_VDA_VLD, + .pix_fmt = AV_PIX_FMT_VDA_VLD, .start_frame = start_frame, .decode_slice = decode_slice, .end_frame = end_frame, diff --git a/libavcodec/vmdav.c b/libavcodec/vmdav.c index fce7aec9f7..5776d92cd5 100644 --- a/libavcodec/vmdav.c +++ b/libavcodec/vmdav.c @@ -388,7 +388,7 @@ static av_cold int vmdvideo_decode_init(AVCodecContext *avctx) unsigned char *raw_palette; s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; /* make sure the VMD header made it */ if (s->avctx->extradata_size != VMD_HEADER_SIZE) { diff --git a/libavcodec/vmnc.c b/libavcodec/vmnc.c index be6c66c0bc..3153be6844 100644 --- a/libavcodec/vmnc.c +++ b/libavcodec/vmnc.c @@ -474,13 +474,13 @@ static av_cold int decode_init(AVCodecContext *avctx) switch(c->bpp){ case 8: - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; break; case 16: - avctx->pix_fmt = PIX_FMT_RGB555; + avctx->pix_fmt = AV_PIX_FMT_RGB555; break; case 32: - avctx->pix_fmt = PIX_FMT_RGB32; + avctx->pix_fmt = AV_PIX_FMT_RGB32; break; default: av_log(avctx, AV_LOG_ERROR, "Unsupported bitdepth %i\n", c->bpp); diff --git a/libavcodec/vp3.c b/libavcodec/vp3.c index 1f5f6c8982..90532cbe54 100644 --- a/libavcodec/vp3.c +++ b/libavcodec/vp3.c @@ -1673,8 +1673,8 @@ static av_cold int vp3_decode_init(AVCodecContext *avctx) s->avctx = avctx; s->width = FFALIGN(avctx->width, 16); s->height = FFALIGN(avctx->height, 16); - if (avctx->pix_fmt == PIX_FMT_NONE) - avctx->pix_fmt = PIX_FMT_YUV420P; + if (avctx->pix_fmt == AV_PIX_FMT_NONE) + avctx->pix_fmt = AV_PIX_FMT_YUV420P; avctx->chroma_sample_location = AVCHROMA_LOC_CENTER; ff_dsputil_init(&s->dsp, avctx); ff_vp3dsp_init(&s->vp3dsp, avctx->flags); @@ -2115,8 +2115,8 @@ static int vp3_init_thread_copy(AVCodecContext *avctx) } #if CONFIG_THEORA_DECODER -static const enum PixelFormat theora_pix_fmts[4] = { - PIX_FMT_YUV420P, PIX_FMT_NONE, PIX_FMT_YUV422P, PIX_FMT_YUV444P +static const enum AVPixelFormat theora_pix_fmts[4] = { + AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV444P }; static int theora_decode_header(AVCodecContext *avctx, GetBitContext *gb) diff --git a/libavcodec/vp56.c b/libavcodec/vp56.c index ac11e2e505..d67bdb6bc1 100644 --- a/libavcodec/vp56.c +++ b/libavcodec/vp56.c @@ -664,7 +664,7 @@ av_cold void ff_vp56_init(AVCodecContext *avctx, int flip, int has_alpha) int i; s->avctx = avctx; - avctx->pix_fmt = has_alpha ? PIX_FMT_YUVA420P : PIX_FMT_YUV420P; + avctx->pix_fmt = has_alpha ? AV_PIX_FMT_YUVA420P : AV_PIX_FMT_YUV420P; ff_dsputil_init(&s->dsp, avctx); ff_vp3dsp_init(&s->vp3dsp, avctx->flags); diff --git a/libavcodec/vp8.c b/libavcodec/vp8.c index e50d7c99b7..e365877bed 100644 --- a/libavcodec/vp8.c +++ b/libavcodec/vp8.c @@ -2016,7 +2016,7 @@ static av_cold int vp8_decode_init(AVCodecContext *avctx) VP8Context *s = avctx->priv_data; s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; ff_dsputil_init(&s->dsp, avctx); ff_h264_pred_init(&s->hpc, AV_CODEC_ID_VP8, 8, 1); diff --git a/libavcodec/vqavideo.c b/libavcodec/vqavideo.c index 1d88ece4a7..32b094d442 100644 --- a/libavcodec/vqavideo.c +++ b/libavcodec/vqavideo.c @@ -124,7 +124,7 @@ static av_cold int vqa_decode_init(AVCodecContext *avctx) int i, j, codebook_index; s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; /* make sure the extradata made it */ if (s->avctx->extradata_size != VQA_HEADER_SIZE) { diff --git a/libavcodec/wmv2enc.c b/libavcodec/wmv2enc.c index 7af9cf6b42..4643835145 100644 --- a/libavcodec/wmv2enc.c +++ b/libavcodec/wmv2enc.c @@ -219,6 +219,6 @@ AVCodec ff_wmv2_encoder = { .init = wmv2_encode_init, .encode2 = ff_MPV_encode_picture, .close = ff_MPV_encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_YUV420P, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_YUV420P, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Windows Media Video 8"), }; diff --git a/libavcodec/wnv1.c b/libavcodec/wnv1.c index 5b0811d196..012642d67a 100644 --- a/libavcodec/wnv1.c +++ b/libavcodec/wnv1.c @@ -135,7 +135,7 @@ static av_cold int decode_init(AVCodecContext *avctx){ static VLC_TYPE code_table[1 << CODE_VLC_BITS][2]; l->avctx = avctx; - avctx->pix_fmt = PIX_FMT_YUV422P; + avctx->pix_fmt = AV_PIX_FMT_YUV422P; code_vlc.table = code_table; code_vlc.table_allocated = 1 << CODE_VLC_BITS; diff --git a/libavcodec/xan.c b/libavcodec/xan.c index b22de0e7ab..586320b55e 100644 --- a/libavcodec/xan.c +++ b/libavcodec/xan.c @@ -79,7 +79,7 @@ static av_cold int xan_decode_init(AVCodecContext *avctx) s->avctx = avctx; s->frame_size = 0; - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; s->buffer1_size = avctx->width * avctx->height; s->buffer1 = av_malloc(s->buffer1_size); diff --git a/libavcodec/xbmenc.c b/libavcodec/xbmenc.c index 98afdcde13..7f28365d40 100644 --- a/libavcodec/xbmenc.c +++ b/libavcodec/xbmenc.c @@ -81,7 +81,7 @@ AVCodec ff_xbm_encoder = { .init = xbm_encode_init, .encode2 = xbm_encode_frame, .close = xbm_encode_close, - .pix_fmts = (const enum PixelFormat[]) { PIX_FMT_MONOWHITE, - PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_MONOWHITE, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("XBM (X BitMap) image"), }; diff --git a/libavcodec/xl.c b/libavcodec/xl.c index 7ba2413221..b63315850c 100644 --- a/libavcodec/xl.c +++ b/libavcodec/xl.c @@ -130,7 +130,7 @@ static int decode_frame(AVCodecContext *avctx, static av_cold int decode_init(AVCodecContext *avctx){ // VideoXLContext * const a = avctx->priv_data; - avctx->pix_fmt= PIX_FMT_YUV411P; + avctx->pix_fmt= AV_PIX_FMT_YUV411P; return 0; } diff --git a/libavcodec/xsubdec.c b/libavcodec/xsubdec.c index 6a1acc7b03..11e1d57fed 100644 --- a/libavcodec/xsubdec.c +++ b/libavcodec/xsubdec.c @@ -26,7 +26,7 @@ #include "bytestream.h" static av_cold int decode_init(AVCodecContext *avctx) { - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; return 0; } diff --git a/libavcodec/xwddec.c b/libavcodec/xwddec.c index bb38a1ebac..4cec6152c5 100644 --- a/libavcodec/xwddec.c +++ b/libavcodec/xwddec.c @@ -153,19 +153,19 @@ static int xwd_decode_frame(AVCodecContext *avctx, void *data, return AVERROR_PATCHWELCOME; } - avctx->pix_fmt = PIX_FMT_NONE; + avctx->pix_fmt = AV_PIX_FMT_NONE; switch (vclass) { case XWD_STATIC_GRAY: case XWD_GRAY_SCALE: if (bpp != 1) return AVERROR_INVALIDDATA; if (pixdepth == 1) - avctx->pix_fmt = PIX_FMT_MONOWHITE; + avctx->pix_fmt = AV_PIX_FMT_MONOWHITE; break; case XWD_STATIC_COLOR: case XWD_PSEUDO_COLOR: if (bpp == 8) - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; break; case XWD_TRUE_COLOR: case XWD_DIRECT_COLOR: @@ -173,24 +173,24 @@ static int xwd_decode_frame(AVCodecContext *avctx, void *data, return AVERROR_INVALIDDATA; if (bpp == 16 && pixdepth == 15) { if (rgb[0] == 0x7C00 && rgb[1] == 0x3E0 && rgb[2] == 0x1F) - avctx->pix_fmt = be ? PIX_FMT_RGB555BE : PIX_FMT_RGB555LE; + avctx->pix_fmt = be ? AV_PIX_FMT_RGB555BE : AV_PIX_FMT_RGB555LE; else if (rgb[0] == 0x1F && rgb[1] == 0x3E0 && rgb[2] == 0x7C00) - avctx->pix_fmt = be ? PIX_FMT_BGR555BE : PIX_FMT_BGR555LE; + avctx->pix_fmt = be ? AV_PIX_FMT_BGR555BE : AV_PIX_FMT_BGR555LE; } else if (bpp == 16 && pixdepth == 16) { if (rgb[0] == 0xF800 && rgb[1] == 0x7E0 && rgb[2] == 0x1F) - avctx->pix_fmt = be ? PIX_FMT_RGB565BE : PIX_FMT_RGB565LE; + avctx->pix_fmt = be ? AV_PIX_FMT_RGB565BE : AV_PIX_FMT_RGB565LE; else if (rgb[0] == 0x1F && rgb[1] == 0x7E0 && rgb[2] == 0xF800) - avctx->pix_fmt = be ? PIX_FMT_BGR565BE : PIX_FMT_BGR565LE; + avctx->pix_fmt = be ? AV_PIX_FMT_BGR565BE : AV_PIX_FMT_BGR565LE; } else if (bpp == 24) { if (rgb[0] == 0xFF0000 && rgb[1] == 0xFF00 && rgb[2] == 0xFF) - avctx->pix_fmt = be ? PIX_FMT_RGB24 : PIX_FMT_BGR24; + avctx->pix_fmt = be ? AV_PIX_FMT_RGB24 : AV_PIX_FMT_BGR24; else if (rgb[0] == 0xFF && rgb[1] == 0xFF00 && rgb[2] == 0xFF0000) - avctx->pix_fmt = be ? PIX_FMT_BGR24 : PIX_FMT_RGB24; + avctx->pix_fmt = be ? AV_PIX_FMT_BGR24 : AV_PIX_FMT_RGB24; } else if (bpp == 32) { if (rgb[0] == 0xFF0000 && rgb[1] == 0xFF00 && rgb[2] == 0xFF) - avctx->pix_fmt = be ? PIX_FMT_ARGB : PIX_FMT_BGRA; + avctx->pix_fmt = be ? AV_PIX_FMT_ARGB : AV_PIX_FMT_BGRA; else if (rgb[0] == 0xFF && rgb[1] == 0xFF00 && rgb[2] == 0xFF0000) - avctx->pix_fmt = be ? PIX_FMT_ABGR : PIX_FMT_RGBA; + avctx->pix_fmt = be ? AV_PIX_FMT_ABGR : AV_PIX_FMT_RGBA; } bytestream2_skipu(&gb, ncolors * XWD_CMAP_SIZE); break; @@ -199,7 +199,7 @@ static int xwd_decode_frame(AVCodecContext *avctx, void *data, return AVERROR_INVALIDDATA; } - if (avctx->pix_fmt == PIX_FMT_NONE) { + if (avctx->pix_fmt == AV_PIX_FMT_NONE) { av_log_ask_for_sample(avctx, "unknown file: bpp %d, pixdepth %d, vclass %d\n", bpp, pixdepth, vclass); return AVERROR_PATCHWELCOME; } @@ -216,7 +216,7 @@ static int xwd_decode_frame(AVCodecContext *avctx, void *data, p->key_frame = 1; p->pict_type = AV_PICTURE_TYPE_I; - if (avctx->pix_fmt == PIX_FMT_PAL8) { + if (avctx->pix_fmt == AV_PIX_FMT_PAL8) { uint32_t *dst = (uint32_t *)p->data[1]; uint8_t red, green, blue; diff --git a/libavcodec/xwdenc.c b/libavcodec/xwdenc.c index 8f56b1b646..a141e0839e 100644 --- a/libavcodec/xwdenc.c +++ b/libavcodec/xwdenc.c @@ -42,7 +42,7 @@ static av_cold int xwd_encode_init(AVCodecContext *avctx) static int xwd_encode_frame(AVCodecContext *avctx, AVPacket *pkt, const AVFrame *p, int *got_packet) { - enum PixelFormat pix_fmt = avctx->pix_fmt; + enum AVPixelFormat pix_fmt = avctx->pix_fmt; uint32_t pixdepth, bpp, bpad, ncolors = 0, lsize, vclass, be = 0; uint32_t rgb[3] = { 0 }, bitorder = 0; uint32_t header_size; @@ -53,15 +53,15 @@ static int xwd_encode_frame(AVCodecContext *avctx, AVPacket *pkt, if (av_pix_fmt_descriptors[pix_fmt].flags & PIX_FMT_BE) be = 1; switch (pix_fmt) { - case PIX_FMT_ARGB: - case PIX_FMT_BGRA: - case PIX_FMT_RGBA: - case PIX_FMT_ABGR: - if (pix_fmt == PIX_FMT_ARGB || - pix_fmt == PIX_FMT_ABGR) + case AV_PIX_FMT_ARGB: + case AV_PIX_FMT_BGRA: + case AV_PIX_FMT_RGBA: + case AV_PIX_FMT_ABGR: + if (pix_fmt == AV_PIX_FMT_ARGB || + pix_fmt == AV_PIX_FMT_ABGR) be = 1; - if (pix_fmt == PIX_FMT_ABGR || - pix_fmt == PIX_FMT_RGBA) { + if (pix_fmt == AV_PIX_FMT_ABGR || + pix_fmt == AV_PIX_FMT_RGBA) { rgb[0] = 0xFF; rgb[1] = 0xFF00; rgb[2] = 0xFF0000; @@ -75,9 +75,9 @@ static int xwd_encode_frame(AVCodecContext *avctx, AVPacket *pkt, vclass = XWD_TRUE_COLOR; bpad = 32; break; - case PIX_FMT_BGR24: - case PIX_FMT_RGB24: - if (pix_fmt == PIX_FMT_RGB24) + case AV_PIX_FMT_BGR24: + case AV_PIX_FMT_RGB24: + if (pix_fmt == AV_PIX_FMT_RGB24) be = 1; bpp = 24; vclass = XWD_TRUE_COLOR; @@ -86,12 +86,12 @@ static int xwd_encode_frame(AVCodecContext *avctx, AVPacket *pkt, rgb[1] = 0xFF00; rgb[2] = 0xFF; break; - case PIX_FMT_RGB565LE: - case PIX_FMT_RGB565BE: - case PIX_FMT_BGR565LE: - case PIX_FMT_BGR565BE: - if (pix_fmt == PIX_FMT_BGR565LE || - pix_fmt == PIX_FMT_BGR565BE) { + case AV_PIX_FMT_RGB565LE: + case AV_PIX_FMT_RGB565BE: + case AV_PIX_FMT_BGR565LE: + case AV_PIX_FMT_BGR565BE: + if (pix_fmt == AV_PIX_FMT_BGR565LE || + pix_fmt == AV_PIX_FMT_BGR565BE) { rgb[0] = 0x1F; rgb[1] = 0x7E0; rgb[2] = 0xF800; @@ -104,12 +104,12 @@ static int xwd_encode_frame(AVCodecContext *avctx, AVPacket *pkt, vclass = XWD_TRUE_COLOR; bpad = 16; break; - case PIX_FMT_RGB555LE: - case PIX_FMT_RGB555BE: - case PIX_FMT_BGR555LE: - case PIX_FMT_BGR555BE: - if (pix_fmt == PIX_FMT_BGR555LE || - pix_fmt == PIX_FMT_BGR555BE) { + case AV_PIX_FMT_RGB555LE: + case AV_PIX_FMT_RGB555BE: + case AV_PIX_FMT_BGR555LE: + case AV_PIX_FMT_BGR555BE: + if (pix_fmt == AV_PIX_FMT_BGR555LE || + pix_fmt == AV_PIX_FMT_BGR555BE) { rgb[0] = 0x1F; rgb[1] = 0x3E0; rgb[2] = 0x7C00; @@ -122,17 +122,17 @@ static int xwd_encode_frame(AVCodecContext *avctx, AVPacket *pkt, vclass = XWD_TRUE_COLOR; bpad = 16; break; - case PIX_FMT_RGB8: - case PIX_FMT_BGR8: - case PIX_FMT_RGB4_BYTE: - case PIX_FMT_BGR4_BYTE: - case PIX_FMT_PAL8: + case AV_PIX_FMT_RGB8: + case AV_PIX_FMT_BGR8: + case AV_PIX_FMT_RGB4_BYTE: + case AV_PIX_FMT_BGR4_BYTE: + case AV_PIX_FMT_PAL8: bpp = 8; vclass = XWD_PSEUDO_COLOR; bpad = 8; ncolors = 256; break; - case PIX_FMT_MONOWHITE: + case AV_PIX_FMT_MONOWHITE: be = 1; bitorder = 1; bpp = 1; @@ -226,26 +226,26 @@ AVCodec ff_xwd_encoder = { .init = xwd_encode_init, .encode2 = xwd_encode_frame, .close = xwd_encode_close, - .pix_fmts = (const enum PixelFormat[]) { PIX_FMT_BGRA, - PIX_FMT_RGBA, - PIX_FMT_ARGB, - PIX_FMT_ABGR, - PIX_FMT_RGB24, - PIX_FMT_BGR24, - PIX_FMT_RGB565BE, - PIX_FMT_RGB565LE, - PIX_FMT_BGR565BE, - PIX_FMT_BGR565LE, - PIX_FMT_RGB555BE, - PIX_FMT_RGB555LE, - PIX_FMT_BGR555BE, - PIX_FMT_BGR555LE, - PIX_FMT_RGB8, - PIX_FMT_BGR8, - PIX_FMT_RGB4_BYTE, - PIX_FMT_BGR4_BYTE, - PIX_FMT_PAL8, - PIX_FMT_MONOWHITE, - PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_BGRA, + AV_PIX_FMT_RGBA, + AV_PIX_FMT_ARGB, + AV_PIX_FMT_ABGR, + AV_PIX_FMT_RGB24, + AV_PIX_FMT_BGR24, + AV_PIX_FMT_RGB565BE, + AV_PIX_FMT_RGB565LE, + AV_PIX_FMT_BGR565BE, + AV_PIX_FMT_BGR565LE, + AV_PIX_FMT_RGB555BE, + AV_PIX_FMT_RGB555LE, + AV_PIX_FMT_BGR555BE, + AV_PIX_FMT_BGR555LE, + AV_PIX_FMT_RGB8, + AV_PIX_FMT_BGR8, + AV_PIX_FMT_RGB4_BYTE, + AV_PIX_FMT_BGR4_BYTE, + AV_PIX_FMT_PAL8, + AV_PIX_FMT_MONOWHITE, + AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("XWD (X Window Dump) image"), }; diff --git a/libavcodec/xxan.c b/libavcodec/xxan.c index 0c3bdb64e8..b06d314d0f 100644 --- a/libavcodec/xxan.c +++ b/libavcodec/xxan.c @@ -44,7 +44,7 @@ static av_cold int xan_decode_init(AVCodecContext *avctx) s->avctx = avctx; - avctx->pix_fmt = PIX_FMT_YUV420P; + avctx->pix_fmt = AV_PIX_FMT_YUV420P; s->buffer_size = avctx->width * avctx->height; s->y_buffer = av_malloc(s->buffer_size); diff --git a/libavcodec/yop.c b/libavcodec/yop.c index d142f3bc18..cda9cad48d 100644 --- a/libavcodec/yop.c +++ b/libavcodec/yop.c @@ -89,7 +89,7 @@ static av_cold int yop_decode_init(AVCodecContext *avctx) return -1; } - avctx->pix_fmt = PIX_FMT_PAL8; + avctx->pix_fmt = AV_PIX_FMT_PAL8; s->num_pal_colors = avctx->extradata[0]; s->first_color[0] = avctx->extradata[1]; diff --git a/libavcodec/zerocodec.c b/libavcodec/zerocodec.c index 789181cb1c..c6a9ba94a2 100644 --- a/libavcodec/zerocodec.c +++ b/libavcodec/zerocodec.c @@ -129,7 +129,7 @@ static av_cold int zerocodec_decode_init(AVCodecContext *avctx) z_stream *zstream = &zc->zstream; int zret; - avctx->pix_fmt = PIX_FMT_UYVY422; + avctx->pix_fmt = AV_PIX_FMT_UYVY422; avctx->bits_per_raw_sample = 8; zc->size = avpicture_get_size(avctx->pix_fmt, diff --git a/libavcodec/zmbv.c b/libavcodec/zmbv.c index b95a518cff..8fb1538514 100644 --- a/libavcodec/zmbv.c +++ b/libavcodec/zmbv.c @@ -627,7 +627,7 @@ static av_cold int decode_init(AVCodecContext *avctx) // Needed if zlib unused or init aborted before inflateInit memset(&c->zstream, 0, sizeof(z_stream)); - avctx->pix_fmt = PIX_FMT_RGB24; + avctx->pix_fmt = AV_PIX_FMT_RGB24; c->decomp_size = (avctx->width + 255) * 4 * (avctx->height + 64); /* Allocate decompression buffer */ diff --git a/libavcodec/zmbvenc.c b/libavcodec/zmbvenc.c index ae91c19bd8..9ac7d6d471 100644 --- a/libavcodec/zmbvenc.c +++ b/libavcodec/zmbvenc.c @@ -345,6 +345,6 @@ AVCodec ff_zmbv_encoder = { .init = encode_init, .encode2 = encode_frame, .close = encode_end, - .pix_fmts = (const enum PixelFormat[]){ PIX_FMT_PAL8, PIX_FMT_NONE }, + .pix_fmts = (const enum AVPixelFormat[]){ AV_PIX_FMT_PAL8, AV_PIX_FMT_NONE }, .long_name = NULL_IF_CONFIG_SMALL("Zip Motion Blocks Video"), }; |