| /* |
| * DPX (.dpx) image decoder |
| * Copyright (c) 2009 Jimmy Christensen |
| * |
| * This file is part of FFmpeg. |
| * |
| * FFmpeg is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2.1 of the License, or (at your option) any later version. |
| * |
| * FFmpeg is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with FFmpeg; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| |
| #include "libavutil/avstring.h" |
| #include "libavutil/intreadwrite.h" |
| #include "libavutil/intfloat.h" |
| #include "libavutil/imgutils.h" |
| #include "libavutil/timecode.h" |
| #include "bytestream.h" |
| #include "avcodec.h" |
| #include "internal.h" |
| |
| enum DPX_TRC { |
| DPX_TRC_USER_DEFINED = 0, |
| DPX_TRC_PRINTING_DENSITY = 1, |
| DPX_TRC_LINEAR = 2, |
| DPX_TRC_LOGARITHMIC = 3, |
| DPX_TRC_UNSPECIFIED_VIDEO = 4, |
| DPX_TRC_SMPTE_274 = 5, |
| DPX_TRC_ITU_R_709_4 = 6, |
| DPX_TRC_ITU_R_601_625 = 7, |
| DPX_TRC_ITU_R_601_525 = 8, |
| DPX_TRC_SMPTE_170 = 9, |
| DPX_TRC_ITU_R_624_4_PAL = 10, |
| DPX_TRC_Z_LINEAR = 11, |
| DPX_TRC_Z_HOMOGENEOUS = 12, |
| }; |
| |
| enum DPX_COL_SPEC { |
| DPX_COL_SPEC_USER_DEFINED = 0, |
| DPX_COL_SPEC_PRINTING_DENSITY = 1, |
| /* 2 = N/A */ |
| /* 3 = N/A */ |
| DPX_COL_SPEC_UNSPECIFIED_VIDEO = 4, |
| DPX_COL_SPEC_SMPTE_274 = 5, |
| DPX_COL_SPEC_ITU_R_709_4 = 6, |
| DPX_COL_SPEC_ITU_R_601_625 = 7, |
| DPX_COL_SPEC_ITU_R_601_525 = 8, |
| DPX_COL_SPEC_SMPTE_170 = 9, |
| DPX_COL_SPEC_ITU_R_624_4_PAL = 10, |
| /* 11 = N/A */ |
| /* 12 = N/A */ |
| }; |
| |
| static unsigned int read16(const uint8_t **ptr, int is_big) |
| { |
| unsigned int temp; |
| if (is_big) { |
| temp = AV_RB16(*ptr); |
| } else { |
| temp = AV_RL16(*ptr); |
| } |
| *ptr += 2; |
| return temp; |
| } |
| |
| static unsigned int read32(const uint8_t **ptr, int is_big) |
| { |
| unsigned int temp; |
| if (is_big) { |
| temp = AV_RB32(*ptr); |
| } else { |
| temp = AV_RL32(*ptr); |
| } |
| *ptr += 4; |
| return temp; |
| } |
| |
| static uint16_t read10in32_gray(const uint8_t **ptr, uint32_t *lbuf, |
| int *n_datum, int is_big, int shift) |
| { |
| uint16_t temp; |
| |
| if (*n_datum) |
| (*n_datum)--; |
| else { |
| *lbuf = read32(ptr, is_big); |
| *n_datum = 2; |
| } |
| |
| temp = *lbuf >> shift & 0x3FF; |
| *lbuf = *lbuf >> 10; |
| |
| return temp; |
| } |
| |
| static uint16_t read10in32(const uint8_t **ptr, uint32_t *lbuf, |
| int *n_datum, int is_big, int shift) |
| { |
| if (*n_datum) |
| (*n_datum)--; |
| else { |
| *lbuf = read32(ptr, is_big); |
| *n_datum = 2; |
| } |
| |
| *lbuf = *lbuf << 10 | *lbuf >> shift & 0x3FFFFF; |
| |
| return *lbuf & 0x3FF; |
| } |
| |
| static uint16_t read12in32(const uint8_t **ptr, uint32_t *lbuf, |
| int *n_datum, int is_big) |
| { |
| if (*n_datum) |
| (*n_datum)--; |
| else { |
| *lbuf = read32(ptr, is_big); |
| *n_datum = 7; |
| } |
| |
| switch (*n_datum){ |
| case 7: return *lbuf & 0xFFF; |
| case 6: return (*lbuf >> 12) & 0xFFF; |
| case 5: { |
| uint32_t c = *lbuf >> 24; |
| *lbuf = read32(ptr, is_big); |
| c |= *lbuf << 8; |
| return c & 0xFFF; |
| } |
| case 4: return (*lbuf >> 4) & 0xFFF; |
| case 3: return (*lbuf >> 16) & 0xFFF; |
| case 2: { |
| uint32_t c = *lbuf >> 28; |
| *lbuf = read32(ptr, is_big); |
| c |= *lbuf << 4; |
| return c & 0xFFF; |
| } |
| case 1: return (*lbuf >> 8) & 0xFFF; |
| default: return *lbuf >> 20; |
| } |
| } |
| |
| static int decode_frame(AVCodecContext *avctx, |
| void *data, |
| int *got_frame, |
| AVPacket *avpkt) |
| { |
| const uint8_t *buf = avpkt->data; |
| int buf_size = avpkt->size; |
| AVFrame *const p = data; |
| uint8_t *ptr[AV_NUM_DATA_POINTERS]; |
| uint32_t header_version, version = 0; |
| char creator[101] = { 0 }; |
| char input_device[33] = { 0 }; |
| |
| unsigned int offset; |
| int magic_num, endian; |
| int x, y, stride, i, j, ret; |
| int w, h, bits_per_color, descriptor, elements, packing; |
| int yuv, color_trc, color_spec; |
| int encoding, need_align = 0, unpadded_10bit = 0; |
| |
| unsigned int rgbBuffer = 0; |
| int n_datum = 0; |
| |
| if (avpkt->size <= 1634) { |
| av_log(avctx, AV_LOG_ERROR, "Packet too small for DPX header\n"); |
| return AVERROR_INVALIDDATA; |
| } |
| |
| magic_num = AV_RB32(buf); |
| buf += 4; |
| |
| /* Check if the files "magic number" is "SDPX" which means it uses |
| * big-endian or XPDS which is for little-endian files */ |
| if (magic_num == AV_RL32("SDPX")) { |
| endian = 0; |
| } else if (magic_num == AV_RB32("SDPX")) { |
| endian = 1; |
| } else { |
| av_log(avctx, AV_LOG_ERROR, "DPX marker not found\n"); |
| return AVERROR_INVALIDDATA; |
| } |
| |
| offset = read32(&buf, endian); |
| if (avpkt->size <= offset) { |
| av_log(avctx, AV_LOG_ERROR, "Invalid data start offset\n"); |
| return AVERROR_INVALIDDATA; |
| } |
| |
| header_version = read32(&buf, 0); |
| if (header_version == MKTAG('V','1','.','0')) |
| version = 1; |
| if (header_version == MKTAG('V','2','.','0')) |
| version = 2; |
| if (!version) |
| av_log(avctx, AV_LOG_WARNING, "Unknown header format version %s.\n", |
| av_fourcc2str(header_version)); |
| |
| // Check encryption |
| buf = avpkt->data + 660; |
| ret = read32(&buf, endian); |
| if (ret != 0xFFFFFFFF) { |
| avpriv_report_missing_feature(avctx, "Encryption"); |
| av_log(avctx, AV_LOG_WARNING, "The image is encrypted and may " |
| "not properly decode.\n"); |
| } |
| |
| // Need to end in 0x304 offset from start of file |
| buf = avpkt->data + 0x304; |
| w = read32(&buf, endian); |
| h = read32(&buf, endian); |
| |
| if ((ret = ff_set_dimensions(avctx, w, h)) < 0) |
| return ret; |
| |
| // Need to end in 0x320 to read the descriptor |
| buf += 20; |
| descriptor = buf[0]; |
| color_trc = buf[1]; |
| color_spec = buf[2]; |
| |
| // Need to end in 0x323 to read the bits per color |
| buf += 3; |
| avctx->bits_per_raw_sample = |
| bits_per_color = buf[0]; |
| buf++; |
| packing = read16(&buf, endian); |
| encoding = read16(&buf, endian); |
| |
| if (encoding) { |
| avpriv_report_missing_feature(avctx, "Encoding %d", encoding); |
| return AVERROR_PATCHWELCOME; |
| } |
| |
| buf += 820; |
| avctx->sample_aspect_ratio.num = read32(&buf, endian); |
| avctx->sample_aspect_ratio.den = read32(&buf, endian); |
| if (avctx->sample_aspect_ratio.num > 0 && avctx->sample_aspect_ratio.den > 0) |
| av_reduce(&avctx->sample_aspect_ratio.num, &avctx->sample_aspect_ratio.den, |
| avctx->sample_aspect_ratio.num, avctx->sample_aspect_ratio.den, |
| 0x10000); |
| else |
| avctx->sample_aspect_ratio = (AVRational){ 0, 1 }; |
| |
| /* preferred frame rate from Motion-picture film header */ |
| if (offset >= 1724 + 4) { |
| buf = avpkt->data + 1724; |
| i = read32(&buf, endian); |
| if(i && i != 0xFFFFFFFF) { |
| AVRational q = av_d2q(av_int2float(i), 4096); |
| if (q.num > 0 && q.den > 0) |
| avctx->framerate = q; |
| } |
| } |
| |
| /* alternative frame rate from television header */ |
| if (offset >= 1940 + 4 && |
| !(avctx->framerate.num && avctx->framerate.den)) { |
| buf = avpkt->data + 1940; |
| i = read32(&buf, endian); |
| if(i && i != 0xFFFFFFFF) { |
| AVRational q = av_d2q(av_int2float(i), 4096); |
| if (q.num > 0 && q.den > 0) |
| avctx->framerate = q; |
| } |
| } |
| |
| /* SMPTE TC from television header */ |
| if (offset >= 1920 + 4) { |
| uint32_t tc; |
| uint32_t *tc_sd; |
| char tcbuf[AV_TIMECODE_STR_SIZE]; |
| |
| buf = avpkt->data + 1920; |
| // read32 to native endian, av_bswap32 to opposite of native for |
| // compatibility with av_timecode_make_smpte_tc_string2 etc |
| tc = av_bswap32(read32(&buf, endian)); |
| |
| if (i != 0xFFFFFFFF) { |
| AVFrameSideData *tcside = |
| av_frame_new_side_data(p, AV_FRAME_DATA_S12M_TIMECODE, |
| sizeof(uint32_t) * 4); |
| if (!tcside) |
| return AVERROR(ENOMEM); |
| |
| tc_sd = (uint32_t*)tcside->data; |
| tc_sd[0] = 1; |
| tc_sd[1] = tc; |
| |
| av_timecode_make_smpte_tc_string2(tcbuf, avctx->framerate, |
| tc_sd[1], 0, 0); |
| av_dict_set(&p->metadata, "timecode", tcbuf, 0); |
| } |
| } |
| |
| /* color range from television header */ |
| if (offset >= 1964 + 4) { |
| buf = avpkt->data + 1952; |
| i = read32(&buf, endian); |
| |
| buf = avpkt->data + 1964; |
| j = read32(&buf, endian); |
| |
| if (i != 0xFFFFFFFF && j != 0xFFFFFFFF) { |
| float minCV, maxCV; |
| minCV = av_int2float(i); |
| maxCV = av_int2float(j); |
| if (bits_per_color >= 1 && |
| minCV == 0.0f && maxCV == ((1<<bits_per_color) - 1)) { |
| avctx->color_range = AVCOL_RANGE_JPEG; |
| } else if (bits_per_color >= 8 && |
| minCV == (1 <<(bits_per_color - 4)) && |
| maxCV == (235<<(bits_per_color - 8))) { |
| avctx->color_range = AVCOL_RANGE_MPEG; |
| } |
| } |
| } |
| |
| switch (descriptor) { |
| case 1: // R |
| case 2: // G |
| case 3: // B |
| case 4: // A |
| case 6: // Y |
| elements = 1; |
| yuv = 1; |
| break; |
| case 50: // RGB |
| elements = 3; |
| yuv = 0; |
| break; |
| case 52: // ABGR |
| case 51: // RGBA |
| elements = 4; |
| yuv = 0; |
| break; |
| case 100: // UYVY422 |
| elements = 2; |
| yuv = 1; |
| break; |
| case 102: // UYV444 |
| elements = 3; |
| yuv = 1; |
| break; |
| case 103: // UYVA4444 |
| elements = 4; |
| yuv = 1; |
| break; |
| default: |
| avpriv_report_missing_feature(avctx, "Descriptor %d", descriptor); |
| return AVERROR_PATCHWELCOME; |
| } |
| |
| switch (bits_per_color) { |
| case 8: |
| stride = avctx->width * elements; |
| break; |
| case 10: |
| if (!packing) { |
| av_log(avctx, AV_LOG_ERROR, "Packing to 32bit required\n"); |
| return -1; |
| } |
| stride = (avctx->width * elements + 2) / 3 * 4; |
| break; |
| case 12: |
| stride = avctx->width * elements; |
| if (packing) { |
| stride *= 2; |
| } else { |
| stride *= 3; |
| if (stride % 8) { |
| stride /= 8; |
| stride++; |
| stride *= 8; |
| } |
| stride /= 2; |
| } |
| break; |
| case 16: |
| stride = 2 * avctx->width * elements; |
| break; |
| case 32: |
| stride = 4 * avctx->width * elements; |
| break; |
| case 1: |
| case 64: |
| avpriv_report_missing_feature(avctx, "Depth %d", bits_per_color); |
| return AVERROR_PATCHWELCOME; |
| default: |
| return AVERROR_INVALIDDATA; |
| } |
| |
| switch (color_trc) { |
| case DPX_TRC_LINEAR: |
| avctx->color_trc = AVCOL_TRC_LINEAR; |
| break; |
| case DPX_TRC_SMPTE_274: |
| case DPX_TRC_ITU_R_709_4: |
| avctx->color_trc = AVCOL_TRC_BT709; |
| break; |
| case DPX_TRC_ITU_R_601_625: |
| case DPX_TRC_ITU_R_601_525: |
| case DPX_TRC_SMPTE_170: |
| avctx->color_trc = AVCOL_TRC_SMPTE170M; |
| break; |
| case DPX_TRC_ITU_R_624_4_PAL: |
| avctx->color_trc = AVCOL_TRC_GAMMA28; |
| break; |
| case DPX_TRC_USER_DEFINED: |
| case DPX_TRC_UNSPECIFIED_VIDEO: |
| /* Nothing to do */ |
| break; |
| default: |
| av_log(avctx, AV_LOG_VERBOSE, "Cannot map DPX transfer characteristic " |
| "%d to color_trc.\n", color_trc); |
| break; |
| } |
| |
| switch (color_spec) { |
| case DPX_COL_SPEC_SMPTE_274: |
| case DPX_COL_SPEC_ITU_R_709_4: |
| avctx->color_primaries = AVCOL_PRI_BT709; |
| break; |
| case DPX_COL_SPEC_ITU_R_601_625: |
| case DPX_COL_SPEC_ITU_R_624_4_PAL: |
| avctx->color_primaries = AVCOL_PRI_BT470BG; |
| break; |
| case DPX_COL_SPEC_ITU_R_601_525: |
| case DPX_COL_SPEC_SMPTE_170: |
| avctx->color_primaries = AVCOL_PRI_SMPTE170M; |
| break; |
| case DPX_COL_SPEC_USER_DEFINED: |
| case DPX_COL_SPEC_UNSPECIFIED_VIDEO: |
| /* Nothing to do */ |
| break; |
| default: |
| av_log(avctx, AV_LOG_VERBOSE, "Cannot map DPX color specification " |
| "%d to color_primaries.\n", color_spec); |
| break; |
| } |
| |
| if (yuv) { |
| switch (color_spec) { |
| case DPX_COL_SPEC_SMPTE_274: |
| case DPX_COL_SPEC_ITU_R_709_4: |
| avctx->colorspace = AVCOL_SPC_BT709; |
| break; |
| case DPX_COL_SPEC_ITU_R_601_625: |
| case DPX_COL_SPEC_ITU_R_624_4_PAL: |
| avctx->colorspace = AVCOL_SPC_BT470BG; |
| break; |
| case DPX_COL_SPEC_ITU_R_601_525: |
| case DPX_COL_SPEC_SMPTE_170: |
| avctx->colorspace = AVCOL_SPC_SMPTE170M; |
| break; |
| case DPX_COL_SPEC_USER_DEFINED: |
| case DPX_COL_SPEC_UNSPECIFIED_VIDEO: |
| /* Nothing to do */ |
| break; |
| default: |
| av_log(avctx, AV_LOG_INFO, "Cannot map DPX color specification " |
| "%d to colorspace.\n", color_spec); |
| break; |
| } |
| } else { |
| avctx->colorspace = AVCOL_SPC_RGB; |
| } |
| |
| // Table 3c: Runs will always break at scan line boundaries. Packing |
| // will always break to the next 32-bit word at scan-line boundaries. |
| // Unfortunately, the encoder produced invalid files, so attempt |
| // to detect it |
| need_align = FFALIGN(stride, 4); |
| if (need_align*avctx->height + (int64_t)offset > avpkt->size) { |
| // Alignment seems unappliable, try without |
| if (stride*avctx->height + (int64_t)offset > avpkt->size) { |
| av_log(avctx, AV_LOG_ERROR, "Overread buffer. Invalid header?\n"); |
| return AVERROR_INVALIDDATA; |
| } else { |
| av_log(avctx, AV_LOG_INFO, "Decoding DPX without scanline " |
| "alignment.\n"); |
| need_align = 0; |
| } |
| } else { |
| need_align -= stride; |
| stride = FFALIGN(stride, 4); |
| } |
| |
| switch (1000 * descriptor + 10 * bits_per_color + endian) { |
| case 1081: |
| case 1080: |
| case 2081: |
| case 2080: |
| case 3081: |
| case 3080: |
| case 4081: |
| case 4080: |
| case 6081: |
| case 6080: |
| avctx->pix_fmt = AV_PIX_FMT_GRAY8; |
| break; |
| case 6121: |
| case 6120: |
| avctx->pix_fmt = AV_PIX_FMT_GRAY12; |
| break; |
| case 1320: |
| case 2320: |
| case 3320: |
| case 4320: |
| case 6320: |
| avctx->pix_fmt = AV_PIX_FMT_GRAYF32LE; |
| break; |
| case 1321: |
| case 2321: |
| case 3321: |
| case 4321: |
| case 6321: |
| avctx->pix_fmt = AV_PIX_FMT_GRAYF32BE; |
| break; |
| case 50081: |
| case 50080: |
| avctx->pix_fmt = AV_PIX_FMT_RGB24; |
| break; |
| case 52081: |
| case 52080: |
| avctx->pix_fmt = AV_PIX_FMT_ABGR; |
| break; |
| case 51081: |
| case 51080: |
| avctx->pix_fmt = AV_PIX_FMT_RGBA; |
| break; |
| case 50100: |
| case 50101: |
| avctx->pix_fmt = AV_PIX_FMT_GBRP10; |
| break; |
| case 51100: |
| case 51101: |
| avctx->pix_fmt = AV_PIX_FMT_GBRAP10; |
| break; |
| case 50120: |
| case 50121: |
| avctx->pix_fmt = AV_PIX_FMT_GBRP12; |
| break; |
| case 51120: |
| case 51121: |
| avctx->pix_fmt = AV_PIX_FMT_GBRAP12; |
| break; |
| case 6100: |
| case 6101: |
| avctx->pix_fmt = AV_PIX_FMT_GRAY10; |
| break; |
| case 6161: |
| avctx->pix_fmt = AV_PIX_FMT_GRAY16BE; |
| break; |
| case 6160: |
| avctx->pix_fmt = AV_PIX_FMT_GRAY16LE; |
| break; |
| case 50161: |
| avctx->pix_fmt = AV_PIX_FMT_RGB48BE; |
| break; |
| case 50160: |
| avctx->pix_fmt = AV_PIX_FMT_RGB48LE; |
| break; |
| case 51161: |
| avctx->pix_fmt = AV_PIX_FMT_RGBA64BE; |
| break; |
| case 51160: |
| avctx->pix_fmt = AV_PIX_FMT_RGBA64LE; |
| break; |
| case 50320: |
| avctx->pix_fmt = AV_PIX_FMT_GBRPF32LE; |
| break; |
| case 50321: |
| avctx->pix_fmt = AV_PIX_FMT_GBRPF32BE; |
| break; |
| case 51320: |
| avctx->pix_fmt = AV_PIX_FMT_GBRAPF32LE; |
| break; |
| case 51321: |
| avctx->pix_fmt = AV_PIX_FMT_GBRAPF32BE; |
| break; |
| case 100081: |
| avctx->pix_fmt = AV_PIX_FMT_UYVY422; |
| break; |
| case 102081: |
| avctx->pix_fmt = AV_PIX_FMT_YUV444P; |
| break; |
| case 103081: |
| avctx->pix_fmt = AV_PIX_FMT_YUVA444P; |
| break; |
| default: |
| av_log(avctx, AV_LOG_ERROR, "Unsupported format %d\n", |
| 1000 * descriptor + 10 * bits_per_color + endian); |
| return AVERROR_PATCHWELCOME; |
| } |
| |
| ff_set_sar(avctx, avctx->sample_aspect_ratio); |
| |
| if ((ret = ff_get_buffer(avctx, p, 0)) < 0) |
| return ret; |
| |
| av_strlcpy(creator, avpkt->data + 160, 100); |
| creator[100] = '\0'; |
| av_dict_set(&p->metadata, "Creator", creator, 0); |
| |
| av_strlcpy(input_device, avpkt->data + 1556, 32); |
| input_device[32] = '\0'; |
| av_dict_set(&p->metadata, "Input Device", input_device, 0); |
| |
| // Some devices do not pad 10bit samples to whole 32bit words per row |
| if (!memcmp(input_device, "Scanity", 7) || |
| !memcmp(creator, "Lasergraphics Inc.", 18)) { |
| unpadded_10bit = 1; |
| } |
| |
| // Move pointer to offset from start of file |
| buf = avpkt->data + offset; |
| |
| for (i=0; i<AV_NUM_DATA_POINTERS; i++) |
| ptr[i] = p->data[i]; |
| |
| switch (bits_per_color) { |
| case 10: |
| for (x = 0; x < avctx->height; x++) { |
| uint16_t *dst[4] = {(uint16_t*)ptr[0], |
| (uint16_t*)ptr[1], |
| (uint16_t*)ptr[2], |
| (uint16_t*)ptr[3]}; |
| int shift = elements > 1 ? packing == 1 ? 22 : 20 : packing == 1 ? 2 : 0; |
| for (y = 0; y < avctx->width; y++) { |
| if (elements >= 3) |
| *dst[2]++ = read10in32(&buf, &rgbBuffer, |
| &n_datum, endian, shift); |
| if (elements == 1) |
| *dst[0]++ = read10in32_gray(&buf, &rgbBuffer, |
| &n_datum, endian, shift); |
| else |
| *dst[0]++ = read10in32(&buf, &rgbBuffer, |
| &n_datum, endian, shift); |
| if (elements >= 2) |
| *dst[1]++ = read10in32(&buf, &rgbBuffer, |
| &n_datum, endian, shift); |
| if (elements == 4) |
| *dst[3]++ = |
| read10in32(&buf, &rgbBuffer, |
| &n_datum, endian, shift); |
| } |
| if (!unpadded_10bit) |
| n_datum = 0; |
| for (i = 0; i < elements; i++) |
| ptr[i] += p->linesize[i]; |
| } |
| break; |
| case 12: |
| for (x = 0; x < avctx->height; x++) { |
| uint16_t *dst[4] = {(uint16_t*)ptr[0], |
| (uint16_t*)ptr[1], |
| (uint16_t*)ptr[2], |
| (uint16_t*)ptr[3]}; |
| int shift = packing == 1 ? 4 : 0; |
| for (y = 0; y < avctx->width; y++) { |
| if (packing) { |
| if (elements >= 3) |
| *dst[2]++ = read16(&buf, endian) >> shift & 0xFFF; |
| *dst[0]++ = read16(&buf, endian) >> shift & 0xFFF; |
| if (elements >= 2) |
| *dst[1]++ = read16(&buf, endian) >> shift & 0xFFF; |
| if (elements == 4) |
| *dst[3]++ = read16(&buf, endian) >> shift & 0xFFF; |
| } else { |
| if (elements >= 3) |
| *dst[2]++ = read12in32(&buf, &rgbBuffer, |
| &n_datum, endian); |
| *dst[0]++ = read12in32(&buf, &rgbBuffer, |
| &n_datum, endian); |
| if (elements >= 2) |
| *dst[1]++ = read12in32(&buf, &rgbBuffer, |
| &n_datum, endian); |
| if (elements == 4) |
| *dst[3]++ = read12in32(&buf, &rgbBuffer, |
| &n_datum, endian); |
| } |
| } |
| n_datum = 0; |
| for (i = 0; i < elements; i++) |
| ptr[i] += p->linesize[i]; |
| // Jump to next aligned position |
| buf += need_align; |
| } |
| break; |
| case 32: |
| if (elements == 1) { |
| av_image_copy_plane(ptr[0], p->linesize[0], |
| buf, stride, |
| elements * avctx->width * 4, avctx->height); |
| } else { |
| for (y = 0; y < avctx->height; y++) { |
| ptr[0] = p->data[0] + y * p->linesize[0]; |
| ptr[1] = p->data[1] + y * p->linesize[1]; |
| ptr[2] = p->data[2] + y * p->linesize[2]; |
| ptr[3] = p->data[3] + y * p->linesize[3]; |
| for (x = 0; x < avctx->width; x++) { |
| AV_WN32(ptr[2], AV_RN32(buf)); |
| AV_WN32(ptr[0], AV_RN32(buf + 4)); |
| AV_WN32(ptr[1], AV_RN32(buf + 8)); |
| if (avctx->pix_fmt == AV_PIX_FMT_GBRAPF32BE || |
| avctx->pix_fmt == AV_PIX_FMT_GBRAPF32LE) { |
| AV_WN32(ptr[3], AV_RN32(buf + 12)); |
| buf += 4; |
| ptr[3] += 4; |
| } |
| |
| buf += 12; |
| ptr[2] += 4; |
| ptr[0] += 4; |
| ptr[1] += 4; |
| } |
| } |
| } |
| break; |
| case 16: |
| elements *= 2; |
| case 8: |
| if ( avctx->pix_fmt == AV_PIX_FMT_YUVA444P |
| || avctx->pix_fmt == AV_PIX_FMT_YUV444P) { |
| for (x = 0; x < avctx->height; x++) { |
| ptr[0] = p->data[0] + x * p->linesize[0]; |
| ptr[1] = p->data[1] + x * p->linesize[1]; |
| ptr[2] = p->data[2] + x * p->linesize[2]; |
| ptr[3] = p->data[3] + x * p->linesize[3]; |
| for (y = 0; y < avctx->width; y++) { |
| *ptr[1]++ = *buf++; |
| *ptr[0]++ = *buf++; |
| *ptr[2]++ = *buf++; |
| if (avctx->pix_fmt == AV_PIX_FMT_YUVA444P) |
| *ptr[3]++ = *buf++; |
| } |
| } |
| } else { |
| av_image_copy_plane(ptr[0], p->linesize[0], |
| buf, stride, |
| elements * avctx->width, avctx->height); |
| } |
| break; |
| } |
| |
| *got_frame = 1; |
| |
| return buf_size; |
| } |
| |
| AVCodec ff_dpx_decoder = { |
| .name = "dpx", |
| .long_name = NULL_IF_CONFIG_SMALL("DPX (Digital Picture Exchange) image"), |
| .type = AVMEDIA_TYPE_VIDEO, |
| .id = AV_CODEC_ID_DPX, |
| .decode = decode_frame, |
| .capabilities = AV_CODEC_CAP_DR1, |
| }; |