/* * DPX (.dpx) image decoder * Copyright (c) 2009 Jimmy Christensen * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "libavutil/avstring.h" #include "libavutil/intreadwrite.h" #include "libavutil/intfloat.h" #include "libavutil/imgutils.h" #include "libavutil/timecode.h" #include "avcodec.h" #include "codec_internal.h" #include "decode.h" enum DPX_TRC { DPX_TRC_USER_DEFINED = 0, DPX_TRC_PRINTING_DENSITY = 1, DPX_TRC_LINEAR = 2, DPX_TRC_LOGARITHMIC = 3, DPX_TRC_UNSPECIFIED_VIDEO = 4, DPX_TRC_SMPTE_274 = 5, DPX_TRC_ITU_R_709_4 = 6, DPX_TRC_ITU_R_601_625 = 7, DPX_TRC_ITU_R_601_525 = 8, DPX_TRC_SMPTE_170 = 9, DPX_TRC_ITU_R_624_4_PAL = 10, DPX_TRC_Z_LINEAR = 11, DPX_TRC_Z_HOMOGENEOUS = 12, }; enum DPX_COL_SPEC { DPX_COL_SPEC_USER_DEFINED = 0, DPX_COL_SPEC_PRINTING_DENSITY = 1, /* 2 = N/A */ /* 3 = N/A */ DPX_COL_SPEC_UNSPECIFIED_VIDEO = 4, DPX_COL_SPEC_SMPTE_274 = 5, DPX_COL_SPEC_ITU_R_709_4 = 6, DPX_COL_SPEC_ITU_R_601_625 = 7, DPX_COL_SPEC_ITU_R_601_525 = 8, DPX_COL_SPEC_SMPTE_170 = 9, DPX_COL_SPEC_ITU_R_624_4_PAL = 10, /* 11 = N/A */ /* 12 = N/A */ }; static unsigned int read16(const uint8_t **ptr, int is_big) { unsigned int temp; if (is_big) { temp = AV_RB16(*ptr); } else { temp = AV_RL16(*ptr); } *ptr += 2; return temp; } static unsigned int read32(const uint8_t **ptr, int is_big) { unsigned int temp; if (is_big) { temp = AV_RB32(*ptr); } else { temp = AV_RL32(*ptr); } *ptr += 4; return temp; } static uint16_t read10in32_gray(const uint8_t **ptr, uint32_t *lbuf, int *n_datum, int is_big, int shift) { uint16_t temp; if (*n_datum) (*n_datum)--; else { *lbuf = read32(ptr, is_big); *n_datum = 2; } temp = *lbuf >> shift & 0x3FF; *lbuf = *lbuf >> 10; return temp; } static uint16_t read10in32(const uint8_t **ptr, uint32_t *lbuf, int *n_datum, int is_big, int shift) { if (*n_datum) (*n_datum)--; else { *lbuf = read32(ptr, is_big); *n_datum = 2; } *lbuf = *lbuf << 10 | *lbuf >> shift & 0x3FFFFF; return *lbuf & 0x3FF; } static uint16_t read12in32(const uint8_t **ptr, uint32_t *lbuf, int *n_datum, int is_big) { if (*n_datum) (*n_datum)--; else { *lbuf = read32(ptr, is_big); *n_datum = 7; } switch (*n_datum){ case 7: return *lbuf & 0xFFF; case 6: return (*lbuf >> 12) & 0xFFF; case 5: { uint32_t c = *lbuf >> 24; *lbuf = read32(ptr, is_big); c |= *lbuf << 8; return c & 0xFFF; } case 4: return (*lbuf >> 4) & 0xFFF; case 3: return (*lbuf >> 16) & 0xFFF; case 2: { uint32_t c = *lbuf >> 28; *lbuf = read32(ptr, is_big); c |= *lbuf << 4; return c & 0xFFF; } case 1: return (*lbuf >> 8) & 0xFFF; default: return *lbuf >> 20; } } static int decode_frame(AVCodecContext *avctx, AVFrame *p, int *got_frame, AVPacket *avpkt) { const uint8_t *buf = avpkt->data; int buf_size = avpkt->size; uint8_t *ptr[AV_NUM_DATA_POINTERS]; uint32_t header_version, version = 0; char creator[101] = { 0 }; char input_device[33] = { 0 }; unsigned int offset; int magic_num, endian; int x, y, stride, i, j, ret; int w, h, bits_per_color, descriptor, elements, packing; int yuv, color_trc, color_spec; int encoding, need_align = 0, unpadded_10bit = 0; unsigned int rgbBuffer = 0; int n_datum = 0; if (avpkt->size <= 1634) { av_log(avctx, AV_LOG_ERROR, "Packet too small for DPX header\n"); return AVERROR_INVALIDDATA; } magic_num = AV_RB32(buf); buf += 4; /* Check if the files "magic number" is "SDPX" which means it uses * big-endian or XPDS which is for little-endian files */ if (magic_num == AV_RL32("SDPX")) { endian = 0; } else if (magic_num == AV_RB32("SDPX")) { endian = 1; } else { av_log(avctx, AV_LOG_ERROR, "DPX marker not found\n"); return AVERROR_INVALIDDATA; } offset = read32(&buf, endian); if (avpkt->size <= offset) { av_log(avctx, AV_LOG_ERROR, "Invalid data start offset\n"); return AVERROR_INVALIDDATA; } header_version = read32(&buf, 0); if (header_version == MKTAG('V','1','.','0')) version = 1; if (header_version == MKTAG('V','2','.','0')) version = 2; if (!version) av_log(avctx, AV_LOG_WARNING, "Unknown header format version %s.\n", av_fourcc2str(header_version)); // Check encryption buf = avpkt->data + 660; ret = read32(&buf, endian); if (ret != 0xFFFFFFFF) { avpriv_report_missing_feature(avctx, "Encryption"); av_log(avctx, AV_LOG_WARNING, "The image is encrypted and may " "not properly decode.\n"); } // Need to end in 0x304 offset from start of file buf = avpkt->data + 0x304; w = read32(&buf, endian); h = read32(&buf, endian); if ((ret = ff_set_dimensions(avctx, w, h)) < 0) return ret; // Need to end in 0x320 to read the descriptor buf += 20; descriptor = buf[0]; color_trc = buf[1]; color_spec = buf[2]; // Need to end in 0x323 to read the bits per color buf += 3; avctx->bits_per_raw_sample = bits_per_color = buf[0]; buf++; packing = read16(&buf, endian); encoding = read16(&buf, endian); if (encoding) { avpriv_report_missing_feature(avctx, "Encoding %d", encoding); return AVERROR_PATCHWELCOME; } if (bits_per_color > 31) return AVERROR_INVALIDDATA; buf += 820; avctx->sample_aspect_ratio.num = read32(&buf, endian); avctx->sample_aspect_ratio.den = read32(&buf, endian); if (avctx->sample_aspect_ratio.num > 0 && avctx->sample_aspect_ratio.den > 0) av_reduce(&avctx->sample_aspect_ratio.num, &avctx->sample_aspect_ratio.den, avctx->sample_aspect_ratio.num, avctx->sample_aspect_ratio.den, 0x10000); else avctx->sample_aspect_ratio = (AVRational){ 0, 1 }; /* preferred frame rate from Motion-picture film header */ if (offset >= 1724 + 4) { buf = avpkt->data + 1724; i = read32(&buf, endian); if(i && i != 0xFFFFFFFF) { AVRational q = av_d2q(av_int2float(i), 4096); if (q.num > 0 && q.den > 0) avctx->framerate = q; } } /* alternative frame rate from television header */ if (offset >= 1940 + 4 && !(avctx->framerate.num && avctx->framerate.den)) { buf = avpkt->data + 1940; i = read32(&buf, endian); if(i && i != 0xFFFFFFFF) { AVRational q = av_d2q(av_int2float(i), 4096); if (q.num > 0 && q.den > 0) avctx->framerate = q; } } /* SMPTE TC from television header */ if (offset >= 1920 + 4) { uint32_t tc; uint32_t *tc_sd; char tcbuf[AV_TIMECODE_STR_SIZE]; buf = avpkt->data + 1920; // read32 to native endian, av_bswap32 to opposite of native for // compatibility with av_timecode_make_smpte_tc_string2 etc tc = av_bswap32(read32(&buf, endian)); if (i != 0xFFFFFFFF) { AVFrameSideData *tcside = av_frame_new_side_data(p, AV_FRAME_DATA_S12M_TIMECODE, sizeof(uint32_t) * 4); if (!tcside) return AVERROR(ENOMEM); tc_sd = (uint32_t*)tcside->data; tc_sd[0] = 1; tc_sd[1] = tc; av_timecode_make_smpte_tc_string2(tcbuf, avctx->framerate, tc_sd[1], 0, 0); av_dict_set(&p->metadata, "timecode", tcbuf, 0); } } /* color range from television header */ if (offset >= 1964 + 4) { buf = avpkt->data + 1952; i = read32(&buf, endian); buf = avpkt->data + 1964; j = read32(&buf, endian); if (i != 0xFFFFFFFF && j != 0xFFFFFFFF) { float minCV, maxCV; minCV = av_int2float(i); maxCV = av_int2float(j); if (bits_per_color >= 1 && minCV == 0.0f && maxCV == ((1U<color_range = AVCOL_RANGE_JPEG; } else if (bits_per_color >= 8 && minCV == (1 <<(bits_per_color - 4)) && maxCV == (235<<(bits_per_color - 8))) { avctx->color_range = AVCOL_RANGE_MPEG; } } } switch (descriptor) { case 1: // R case 2: // G case 3: // B case 4: // A case 6: // Y elements = 1; yuv = 1; break; case 50: // RGB elements = 3; yuv = 0; break; case 52: // ABGR case 51: // RGBA elements = 4; yuv = 0; break; case 100: // UYVY422 elements = 2; yuv = 1; break; case 102: // UYV444 elements = 3; yuv = 1; break; case 103: // UYVA4444 elements = 4; yuv = 1; break; default: avpriv_report_missing_feature(avctx, "Descriptor %d", descriptor); return AVERROR_PATCHWELCOME; } switch (bits_per_color) { case 8: stride = avctx->width * elements; break; case 10: if (!packing) { av_log(avctx, AV_LOG_ERROR, "Packing to 32bit required\n"); return -1; } stride = (avctx->width * elements + 2) / 3 * 4; break; case 12: stride = avctx->width * elements; if (packing) { stride *= 2; } else { stride *= 3; if (stride % 8) { stride /= 8; stride++; stride *= 8; } stride /= 2; } break; case 16: stride = 2 * avctx->width * elements; break; case 32: stride = 4 * avctx->width * elements; break; case 1: case 64: avpriv_report_missing_feature(avctx, "Depth %d", bits_per_color); return AVERROR_PATCHWELCOME; default: return AVERROR_INVALIDDATA; } switch (color_trc) { case DPX_TRC_LINEAR: avctx->color_trc = AVCOL_TRC_LINEAR; break; case DPX_TRC_SMPTE_274: case DPX_TRC_ITU_R_709_4: avctx->color_trc = AVCOL_TRC_BT709; break; case DPX_TRC_ITU_R_601_625: case DPX_TRC_ITU_R_601_525: case DPX_TRC_SMPTE_170: avctx->color_trc = AVCOL_TRC_SMPTE170M; break; case DPX_TRC_ITU_R_624_4_PAL: avctx->color_trc = AVCOL_TRC_GAMMA28; break; case DPX_TRC_USER_DEFINED: case DPX_TRC_UNSPECIFIED_VIDEO: /* Nothing to do */ break; default: av_log(avctx, AV_LOG_VERBOSE, "Cannot map DPX transfer characteristic " "%d to color_trc.\n", color_trc); break; } switch (color_spec) { case DPX_COL_SPEC_SMPTE_274: case DPX_COL_SPEC_ITU_R_709_4: avctx->color_primaries = AVCOL_PRI_BT709; break; case DPX_COL_SPEC_ITU_R_601_625: case DPX_COL_SPEC_ITU_R_624_4_PAL: avctx->color_primaries = AVCOL_PRI_BT470BG; break; case DPX_COL_SPEC_ITU_R_601_525: case DPX_COL_SPEC_SMPTE_170: avctx->color_primaries = AVCOL_PRI_SMPTE170M; break; case DPX_COL_SPEC_USER_DEFINED: case DPX_COL_SPEC_UNSPECIFIED_VIDEO: /* Nothing to do */ break; default: av_log(avctx, AV_LOG_VERBOSE, "Cannot map DPX color specification " "%d to color_primaries.\n", color_spec); break; } if (yuv) { switch (color_spec) { case DPX_COL_SPEC_SMPTE_274: case DPX_COL_SPEC_ITU_R_709_4: avctx->colorspace = AVCOL_SPC_BT709; break; case DPX_COL_SPEC_ITU_R_601_625: case DPX_COL_SPEC_ITU_R_624_4_PAL: avctx->colorspace = AVCOL_SPC_BT470BG; break; case DPX_COL_SPEC_ITU_R_601_525: case DPX_COL_SPEC_SMPTE_170: avctx->colorspace = AVCOL_SPC_SMPTE170M; break; case DPX_COL_SPEC_USER_DEFINED: case DPX_COL_SPEC_UNSPECIFIED_VIDEO: /* Nothing to do */ break; default: av_log(avctx, AV_LOG_INFO, "Cannot map DPX color specification " "%d to colorspace.\n", color_spec); break; } } else { avctx->colorspace = AVCOL_SPC_RGB; } av_strlcpy(creator, avpkt->data + 160, 100); creator[100] = '\0'; av_dict_set(&p->metadata, "Creator", creator, 0); av_strlcpy(input_device, avpkt->data + 1556, 32); input_device[32] = '\0'; av_dict_set(&p->metadata, "Input Device", input_device, 0); // Some devices do not pad 10bit samples to whole 32bit words per row if (!memcmp(input_device, "Scanity", 7) || !memcmp(creator, "Lasergraphics Inc.", 18)) { if (bits_per_color == 10) unpadded_10bit = 1; } // Table 3c: Runs will always break at scan line boundaries. Packing // will always break to the next 32-bit word at scan-line boundaries. // Unfortunately, the encoder produced invalid files, so attempt // to detect it // Also handle special case with unpadded content need_align = FFALIGN(stride, 4); if (need_align*avctx->height + (int64_t)offset > avpkt->size && (!unpadded_10bit || (avctx->width * avctx->height * elements + 2) / 3 * 4 + (int64_t)offset > avpkt->size)) { // Alignment seems unappliable, try without if (stride*avctx->height + (int64_t)offset > avpkt->size || unpadded_10bit) { av_log(avctx, AV_LOG_ERROR, "Overread buffer. Invalid header?\n"); return AVERROR_INVALIDDATA; } else { av_log(avctx, AV_LOG_INFO, "Decoding DPX without scanline " "alignment.\n"); need_align = 0; } } else { need_align -= stride; stride = FFALIGN(stride, 4); } switch (1000 * descriptor + 10 * bits_per_color + endian) { case 1081: case 1080: case 2081: case 2080: case 3081: case 3080: case 4081: case 4080: case 6081: case 6080: avctx->pix_fmt = AV_PIX_FMT_GRAY8; break; case 6121: case 6120: avctx->pix_fmt = AV_PIX_FMT_GRAY12; break; case 1320: case 2320: case 3320: case 4320: case 6320: avctx->pix_fmt = AV_PIX_FMT_GRAYF32LE; break; case 1321: case 2321: case 3321: case 4321: case 6321: avctx->pix_fmt = AV_PIX_FMT_GRAYF32BE; break; case 50081: case 50080: avctx->pix_fmt = AV_PIX_FMT_RGB24; break; case 52081: case 52080: avctx->pix_fmt = AV_PIX_FMT_ABGR; break; case 51081: case 51080: avctx->pix_fmt = AV_PIX_FMT_RGBA; break; case 50100: case 50101: avctx->pix_fmt = AV_PIX_FMT_GBRP10; break; case 51100: case 51101: avctx->pix_fmt = AV_PIX_FMT_GBRAP10; break; case 50120: case 50121: avctx->pix_fmt = AV_PIX_FMT_GBRP12; break; case 51120: case 51121: avctx->pix_fmt = AV_PIX_FMT_GBRAP12; break; case 6100: case 6101: avctx->pix_fmt = AV_PIX_FMT_GRAY10; break; case 6161: avctx->pix_fmt = AV_PIX_FMT_GRAY16BE; break; case 6160: avctx->pix_fmt = AV_PIX_FMT_GRAY16LE; break; case 50161: avctx->pix_fmt = AV_PIX_FMT_RGB48BE; break; case 50160: avctx->pix_fmt = AV_PIX_FMT_RGB48LE; break; case 51161: avctx->pix_fmt = AV_PIX_FMT_RGBA64BE; break; case 51160: avctx->pix_fmt = AV_PIX_FMT_RGBA64LE; break; case 50320: avctx->pix_fmt = AV_PIX_FMT_GBRPF32LE; break; case 50321: avctx->pix_fmt = AV_PIX_FMT_GBRPF32BE; break; case 51320: avctx->pix_fmt = AV_PIX_FMT_GBRAPF32LE; break; case 51321: avctx->pix_fmt = AV_PIX_FMT_GBRAPF32BE; break; case 100081: avctx->pix_fmt = AV_PIX_FMT_UYVY422; break; case 102081: avctx->pix_fmt = AV_PIX_FMT_YUV444P; break; case 103081: avctx->pix_fmt = AV_PIX_FMT_YUVA444P; break; default: av_log(avctx, AV_LOG_ERROR, "Unsupported format %d\n", 1000 * descriptor + 10 * bits_per_color + endian); return AVERROR_PATCHWELCOME; } ff_set_sar(avctx, avctx->sample_aspect_ratio); if ((ret = ff_get_buffer(avctx, p, 0)) < 0) return ret; // Move pointer to offset from start of file buf = avpkt->data + offset; for (i=0; idata[i]; switch (bits_per_color) { case 10: for (x = 0; x < avctx->height; x++) { uint16_t *dst[4] = {(uint16_t*)ptr[0], (uint16_t*)ptr[1], (uint16_t*)ptr[2], (uint16_t*)ptr[3]}; int shift = elements > 1 ? packing == 1 ? 22 : 20 : packing == 1 ? 2 : 0; for (y = 0; y < avctx->width; y++) { if (elements >= 3) *dst[2]++ = read10in32(&buf, &rgbBuffer, &n_datum, endian, shift); if (elements == 1) *dst[0]++ = read10in32_gray(&buf, &rgbBuffer, &n_datum, endian, shift); else *dst[0]++ = read10in32(&buf, &rgbBuffer, &n_datum, endian, shift); if (elements >= 2) *dst[1]++ = read10in32(&buf, &rgbBuffer, &n_datum, endian, shift); if (elements == 4) *dst[3]++ = read10in32(&buf, &rgbBuffer, &n_datum, endian, shift); } if (!unpadded_10bit) n_datum = 0; for (i = 0; i < elements; i++) ptr[i] += p->linesize[i]; } break; case 12: for (x = 0; x < avctx->height; x++) { uint16_t *dst[4] = {(uint16_t*)ptr[0], (uint16_t*)ptr[1], (uint16_t*)ptr[2], (uint16_t*)ptr[3]}; int shift = packing == 1 ? 4 : 0; for (y = 0; y < avctx->width; y++) { if (packing) { if (elements >= 3) *dst[2]++ = read16(&buf, endian) >> shift & 0xFFF; *dst[0]++ = read16(&buf, endian) >> shift & 0xFFF; if (elements >= 2) *dst[1]++ = read16(&buf, endian) >> shift & 0xFFF; if (elements == 4) *dst[3]++ = read16(&buf, endian) >> shift & 0xFFF; } else { if (elements >= 3) *dst[2]++ = read12in32(&buf, &rgbBuffer, &n_datum, endian); *dst[0]++ = read12in32(&buf, &rgbBuffer, &n_datum, endian); if (elements >= 2) *dst[1]++ = read12in32(&buf, &rgbBuffer, &n_datum, endian); if (elements == 4) *dst[3]++ = read12in32(&buf, &rgbBuffer, &n_datum, endian); } } n_datum = 0; for (i = 0; i < elements; i++) ptr[i] += p->linesize[i]; // Jump to next aligned position buf += need_align; } break; case 32: if (elements == 1) { av_image_copy_plane(ptr[0], p->linesize[0], buf, stride, elements * avctx->width * 4, avctx->height); } else { for (y = 0; y < avctx->height; y++) { ptr[0] = p->data[0] + y * p->linesize[0]; ptr[1] = p->data[1] + y * p->linesize[1]; ptr[2] = p->data[2] + y * p->linesize[2]; ptr[3] = p->data[3] + y * p->linesize[3]; for (x = 0; x < avctx->width; x++) { AV_WN32(ptr[2], AV_RN32(buf)); AV_WN32(ptr[0], AV_RN32(buf + 4)); AV_WN32(ptr[1], AV_RN32(buf + 8)); if (avctx->pix_fmt == AV_PIX_FMT_GBRAPF32BE || avctx->pix_fmt == AV_PIX_FMT_GBRAPF32LE) { AV_WN32(ptr[3], AV_RN32(buf + 12)); buf += 4; ptr[3] += 4; } buf += 12; ptr[2] += 4; ptr[0] += 4; ptr[1] += 4; } } } break; case 16: elements *= 2; case 8: if ( avctx->pix_fmt == AV_PIX_FMT_YUVA444P || avctx->pix_fmt == AV_PIX_FMT_YUV444P) { for (x = 0; x < avctx->height; x++) { ptr[0] = p->data[0] + x * p->linesize[0]; ptr[1] = p->data[1] + x * p->linesize[1]; ptr[2] = p->data[2] + x * p->linesize[2]; ptr[3] = p->data[3] + x * p->linesize[3]; for (y = 0; y < avctx->width; y++) { *ptr[1]++ = *buf++; *ptr[0]++ = *buf++; *ptr[2]++ = *buf++; if (avctx->pix_fmt == AV_PIX_FMT_YUVA444P) *ptr[3]++ = *buf++; } } } else { av_image_copy_plane(ptr[0], p->linesize[0], buf, stride, elements * avctx->width, avctx->height); } break; } *got_frame = 1; return buf_size; } const FFCodec ff_dpx_decoder = { .p.name = "dpx", CODEC_LONG_NAME("DPX (Digital Picture Exchange) image"), .p.type = AVMEDIA_TYPE_VIDEO, .p.id = AV_CODEC_ID_DPX, FF_CODEC_DECODE_CB(decode_frame), .p.capabilities = AV_CODEC_CAP_DR1, };