| /* |
| * ColorMatrix v2.2 for Avisynth 2.5.x |
| * |
| * Copyright (C) 2006-2007 Kevin Stone |
| * |
| * ColorMatrix 1.x is Copyright (C) Wilbert Dijkhof |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms of the GNU General Public License as published by the |
| * Free Software Foundation; either version 2 of the License, or (at your |
| * option) any later version. |
| * |
| * This program is distributed in the hope that it will be useful, but |
| * OUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY |
| * or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public |
| * License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; if not, write to the Free Software Foundation, |
| * Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| |
| /** |
| * @file |
| * ColorMatrix 2.0 is based on the original ColorMatrix filter by Wilbert |
| * Dijkhof. It adds the ability to convert between any of: Rec.709, FCC, |
| * Rec.601, and SMPTE 240M. It also makes pre and post clipping optional, |
| * adds an option to use scaled or non-scaled coefficients, and more... |
| */ |
| |
| #include <float.h> |
| #include "avfilter.h" |
| #include "formats.h" |
| #include "internal.h" |
| #include "video.h" |
| #include "libavutil/opt.h" |
| #include "libavutil/pixdesc.h" |
| #include "libavutil/avstring.h" |
| |
| #define NS(n) ((n) < 0 ? (int)((n)*65536.0-0.5+DBL_EPSILON) : (int)((n)*65536.0+0.5)) |
| #define CB(n) av_clip_uint8(n) |
| |
| static const double yuv_coeff_luma[5][3] = { |
| { +0.7152, +0.0722, +0.2126 }, // Rec.709 (0) |
| { +0.5900, +0.1100, +0.3000 }, // FCC (1) |
| { +0.5870, +0.1140, +0.2990 }, // Rec.601 (ITU-R BT.470-2/SMPTE 170M) (2) |
| { +0.7010, +0.0870, +0.2120 }, // SMPTE 240M (3) |
| { +0.6780, +0.0593, +0.2627 }, // Rec.2020 (4) |
| }; |
| |
| enum ColorMode { |
| COLOR_MODE_NONE = -1, |
| COLOR_MODE_BT709, |
| COLOR_MODE_FCC, |
| COLOR_MODE_BT601, |
| COLOR_MODE_SMPTE240M, |
| COLOR_MODE_BT2020, |
| COLOR_MODE_COUNT |
| }; |
| |
| typedef struct ColorMatrixContext { |
| const AVClass *class; |
| int yuv_convert[25][3][3]; |
| int interlaced; |
| int source, dest; ///< ColorMode |
| int mode; |
| int hsub, vsub; |
| } ColorMatrixContext; |
| |
| typedef struct ThreadData { |
| AVFrame *dst; |
| const AVFrame *src; |
| int c2; |
| int c3; |
| int c4; |
| int c5; |
| int c6; |
| int c7; |
| } ThreadData; |
| |
| #define OFFSET(x) offsetof(ColorMatrixContext, x) |
| #define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM |
| |
| static const AVOption colormatrix_options[] = { |
| { "src", "set source color matrix", OFFSET(source), AV_OPT_TYPE_INT, {.i64=COLOR_MODE_NONE}, COLOR_MODE_NONE, COLOR_MODE_COUNT-1, .flags=FLAGS, .unit="color_mode" }, |
| { "dst", "set destination color matrix", OFFSET(dest), AV_OPT_TYPE_INT, {.i64=COLOR_MODE_NONE}, COLOR_MODE_NONE, COLOR_MODE_COUNT-1, .flags=FLAGS, .unit="color_mode" }, |
| { "bt709", "set BT.709 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT709}, .flags=FLAGS, .unit="color_mode" }, |
| { "fcc", "set FCC colorspace ", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_FCC}, .flags=FLAGS, .unit="color_mode" }, |
| { "bt601", "set BT.601 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT601}, .flags=FLAGS, .unit="color_mode" }, |
| { "bt470", "set BT.470 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT601}, .flags=FLAGS, .unit="color_mode" }, |
| { "bt470bg", "set BT.470 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT601}, .flags=FLAGS, .unit="color_mode" }, |
| { "smpte170m", "set SMTPE-170M colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT601}, .flags=FLAGS, .unit="color_mode" }, |
| { "smpte240m", "set SMPTE-240M colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_SMPTE240M}, .flags=FLAGS, .unit="color_mode" }, |
| { "bt2020", "set BT.2020 colorspace", 0, AV_OPT_TYPE_CONST, {.i64=COLOR_MODE_BT2020}, .flags=FLAGS, .unit="color_mode" }, |
| { NULL } |
| }; |
| |
| AVFILTER_DEFINE_CLASS(colormatrix); |
| |
| #define ma m[0][0] |
| #define mb m[0][1] |
| #define mc m[0][2] |
| #define md m[1][0] |
| #define me m[1][1] |
| #define mf m[1][2] |
| #define mg m[2][0] |
| #define mh m[2][1] |
| #define mi m[2][2] |
| |
| #define ima im[0][0] |
| #define imb im[0][1] |
| #define imc im[0][2] |
| #define imd im[1][0] |
| #define ime im[1][1] |
| #define imf im[1][2] |
| #define img im[2][0] |
| #define imh im[2][1] |
| #define imi im[2][2] |
| |
| static void inverse3x3(double im[3][3], double m[3][3]) |
| { |
| double det = ma * (me * mi - mf * mh) - mb * (md * mi - mf * mg) + mc * (md * mh - me * mg); |
| det = 1.0 / det; |
| ima = det * (me * mi - mf * mh); |
| imb = det * (mc * mh - mb * mi); |
| imc = det * (mb * mf - mc * me); |
| imd = det * (mf * mg - md * mi); |
| ime = det * (ma * mi - mc * mg); |
| imf = det * (mc * md - ma * mf); |
| img = det * (md * mh - me * mg); |
| imh = det * (mb * mg - ma * mh); |
| imi = det * (ma * me - mb * md); |
| } |
| |
| static void solve_coefficients(double cm[3][3], double rgb[3][3], double yuv[3][3]) |
| { |
| int i, j; |
| for (i = 0; i < 3; i++) |
| for (j = 0; j < 3; j++) |
| cm[i][j] = yuv[i][0] * rgb[0][j] + yuv[i][1] * rgb[1][j] + yuv[i][2] * rgb[2][j]; |
| } |
| |
| static void calc_coefficients(AVFilterContext *ctx) |
| { |
| ColorMatrixContext *color = ctx->priv; |
| double yuv_coeff[5][3][3]; |
| double rgb_coeffd[5][3][3]; |
| double yuv_convertd[25][3][3]; |
| double bscale, rscale; |
| int v = 0; |
| int i, j, k; |
| for (i = 0; i < 5; i++) { |
| yuv_coeff[i][0][0] = yuv_coeff_luma[i][0]; |
| yuv_coeff[i][0][1] = yuv_coeff_luma[i][1]; |
| yuv_coeff[i][0][2] = yuv_coeff_luma[i][2]; |
| bscale = 0.5 / (yuv_coeff[i][0][1] - 1.0); |
| rscale = 0.5 / (yuv_coeff[i][0][2] - 1.0); |
| yuv_coeff[i][1][0] = bscale * yuv_coeff[i][0][0]; |
| yuv_coeff[i][1][1] = 0.5; |
| yuv_coeff[i][1][2] = bscale * yuv_coeff[i][0][2]; |
| yuv_coeff[i][2][0] = rscale * yuv_coeff[i][0][0]; |
| yuv_coeff[i][2][1] = rscale * yuv_coeff[i][0][1]; |
| yuv_coeff[i][2][2] = 0.5; |
| } |
| for (i = 0; i < 5; i++) |
| inverse3x3(rgb_coeffd[i], yuv_coeff[i]); |
| for (i = 0; i < 5; i++) { |
| for (j = 0; j < 5; j++) { |
| solve_coefficients(yuv_convertd[v], rgb_coeffd[i], yuv_coeff[j]); |
| for (k = 0; k < 3; k++) { |
| color->yuv_convert[v][k][0] = NS(yuv_convertd[v][k][0]); |
| color->yuv_convert[v][k][1] = NS(yuv_convertd[v][k][1]); |
| color->yuv_convert[v][k][2] = NS(yuv_convertd[v][k][2]); |
| } |
| if (color->yuv_convert[v][0][0] != 65536 || color->yuv_convert[v][1][0] != 0 || |
| color->yuv_convert[v][2][0] != 0) { |
| av_log(ctx, AV_LOG_ERROR, "error calculating conversion coefficients\n"); |
| } |
| v++; |
| } |
| } |
| } |
| |
| static const char * const color_modes[] = {"bt709", "fcc", "bt601", "smpte240m", "bt2020"}; |
| |
| static av_cold int init(AVFilterContext *ctx) |
| { |
| ColorMatrixContext *color = ctx->priv; |
| |
| if (color->dest == COLOR_MODE_NONE) { |
| av_log(ctx, AV_LOG_ERROR, "Unspecified destination color space\n"); |
| return AVERROR(EINVAL); |
| } |
| |
| if (color->source == color->dest) { |
| av_log(ctx, AV_LOG_ERROR, "Source and destination color space must not be identical\n"); |
| return AVERROR(EINVAL); |
| } |
| |
| calc_coefficients(ctx); |
| |
| return 0; |
| } |
| |
| static int process_slice_uyvy422(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) |
| { |
| const ThreadData *td = arg; |
| const AVFrame *src = td->src; |
| AVFrame *dst = td->dst; |
| const int height = src->height; |
| const int width = src->width*2; |
| const int src_pitch = src->linesize[0]; |
| const int dst_pitch = dst->linesize[0]; |
| const int slice_start = (height * jobnr ) / nb_jobs; |
| const int slice_end = (height * (jobnr+1)) / nb_jobs; |
| const unsigned char *srcp = src->data[0] + slice_start * src_pitch; |
| unsigned char *dstp = dst->data[0] + slice_start * dst_pitch; |
| const int c2 = td->c2; |
| const int c3 = td->c3; |
| const int c4 = td->c4; |
| const int c5 = td->c5; |
| const int c6 = td->c6; |
| const int c7 = td->c7; |
| int x, y; |
| |
| for (y = slice_start; y < slice_end; y++) { |
| for (x = 0; x < width; x += 4) { |
| const int u = srcp[x + 0] - 128; |
| const int v = srcp[x + 2] - 128; |
| const int uvval = c2 * u + c3 * v + 1081344; |
| dstp[x + 0] = CB((c4 * u + c5 * v + 8421376) >> 16); |
| dstp[x + 1] = CB((65536 * (srcp[x + 1] - 16) + uvval) >> 16); |
| dstp[x + 2] = CB((c6 * u + c7 * v + 8421376) >> 16); |
| dstp[x + 3] = CB((65536 * (srcp[x + 3] - 16) + uvval) >> 16); |
| } |
| srcp += src_pitch; |
| dstp += dst_pitch; |
| } |
| |
| return 0; |
| } |
| |
| static int process_slice_yuv444p(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) |
| { |
| const ThreadData *td = arg; |
| const AVFrame *src = td->src; |
| AVFrame *dst = td->dst; |
| const int height = src->height; |
| const int width = src->width; |
| const int slice_start = (height * jobnr ) / nb_jobs; |
| const int slice_end = (height * (jobnr+1)) / nb_jobs; |
| const int src_pitchY = src->linesize[0]; |
| const int src_pitchUV = src->linesize[1]; |
| const unsigned char *srcpU = src->data[1] + slice_start * src_pitchUV; |
| const unsigned char *srcpV = src->data[2] + slice_start * src_pitchUV; |
| const unsigned char *srcpY = src->data[0] + slice_start * src_pitchY; |
| const int dst_pitchY = dst->linesize[0]; |
| const int dst_pitchUV = dst->linesize[1]; |
| unsigned char *dstpU = dst->data[1] + slice_start * dst_pitchUV; |
| unsigned char *dstpV = dst->data[2] + slice_start * dst_pitchUV; |
| unsigned char *dstpY = dst->data[0] + slice_start * dst_pitchY; |
| const int c2 = td->c2; |
| const int c3 = td->c3; |
| const int c4 = td->c4; |
| const int c5 = td->c5; |
| const int c6 = td->c6; |
| const int c7 = td->c7; |
| int x, y; |
| |
| for (y = slice_start; y < slice_end; y++) { |
| for (x = 0; x < width; x++) { |
| const int u = srcpU[x] - 128; |
| const int v = srcpV[x] - 128; |
| const int uvval = c2 * u + c3 * v + 1081344; |
| dstpY[x] = CB((65536 * (srcpY[x] - 16) + uvval) >> 16); |
| dstpU[x] = CB((c4 * u + c5 * v + 8421376) >> 16); |
| dstpV[x] = CB((c6 * u + c7 * v + 8421376) >> 16); |
| } |
| srcpY += src_pitchY; |
| dstpY += dst_pitchY; |
| srcpU += src_pitchUV; |
| srcpV += src_pitchUV; |
| dstpU += dst_pitchUV; |
| dstpV += dst_pitchUV; |
| } |
| |
| return 0; |
| } |
| |
| static int process_slice_yuv422p(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) |
| { |
| const ThreadData *td = arg; |
| const AVFrame *src = td->src; |
| AVFrame *dst = td->dst; |
| const int height = src->height; |
| const int width = src->width; |
| const int slice_start = (height * jobnr ) / nb_jobs; |
| const int slice_end = (height * (jobnr+1)) / nb_jobs; |
| const int src_pitchY = src->linesize[0]; |
| const int src_pitchUV = src->linesize[1]; |
| const unsigned char *srcpU = src->data[1] + slice_start * src_pitchUV; |
| const unsigned char *srcpV = src->data[2] + slice_start * src_pitchUV; |
| const unsigned char *srcpY = src->data[0] + slice_start * src_pitchY; |
| const int dst_pitchY = dst->linesize[0]; |
| const int dst_pitchUV = dst->linesize[1]; |
| unsigned char *dstpU = dst->data[1] + slice_start * dst_pitchUV; |
| unsigned char *dstpV = dst->data[2] + slice_start * dst_pitchUV; |
| unsigned char *dstpY = dst->data[0] + slice_start * dst_pitchY; |
| const int c2 = td->c2; |
| const int c3 = td->c3; |
| const int c4 = td->c4; |
| const int c5 = td->c5; |
| const int c6 = td->c6; |
| const int c7 = td->c7; |
| int x, y; |
| |
| for (y = slice_start; y < slice_end; y++) { |
| for (x = 0; x < width; x += 2) { |
| const int u = srcpU[x >> 1] - 128; |
| const int v = srcpV[x >> 1] - 128; |
| const int uvval = c2 * u + c3 * v + 1081344; |
| dstpY[x + 0] = CB((65536 * (srcpY[x + 0] - 16) + uvval) >> 16); |
| dstpY[x + 1] = CB((65536 * (srcpY[x + 1] - 16) + uvval) >> 16); |
| dstpU[x >> 1] = CB((c4 * u + c5 * v + 8421376) >> 16); |
| dstpV[x >> 1] = CB((c6 * u + c7 * v + 8421376) >> 16); |
| } |
| srcpY += src_pitchY; |
| dstpY += dst_pitchY; |
| srcpU += src_pitchUV; |
| srcpV += src_pitchUV; |
| dstpU += dst_pitchUV; |
| dstpV += dst_pitchUV; |
| } |
| |
| return 0; |
| } |
| |
| static int process_slice_yuv420p(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) |
| { |
| const ThreadData *td = arg; |
| const AVFrame *src = td->src; |
| AVFrame *dst = td->dst; |
| const int height = FFALIGN(src->height, 2) >> 1; |
| const int width = src->width; |
| const int slice_start = ((height * jobnr ) / nb_jobs) << 1; |
| const int slice_end = ((height * (jobnr+1)) / nb_jobs) << 1; |
| const int src_pitchY = src->linesize[0]; |
| const int src_pitchUV = src->linesize[1]; |
| const int dst_pitchY = dst->linesize[0]; |
| const int dst_pitchUV = dst->linesize[1]; |
| const unsigned char *srcpY = src->data[0] + src_pitchY * slice_start; |
| const unsigned char *srcpU = src->data[1] + src_pitchUV * (slice_start >> 1); |
| const unsigned char *srcpV = src->data[2] + src_pitchUV * (slice_start >> 1); |
| const unsigned char *srcpN = src->data[0] + src_pitchY * (slice_start + 1); |
| unsigned char *dstpU = dst->data[1] + dst_pitchUV * (slice_start >> 1); |
| unsigned char *dstpV = dst->data[2] + dst_pitchUV * (slice_start >> 1); |
| unsigned char *dstpY = dst->data[0] + dst_pitchY * slice_start; |
| unsigned char *dstpN = dst->data[0] + dst_pitchY * (slice_start + 1); |
| const int c2 = td->c2; |
| const int c3 = td->c3; |
| const int c4 = td->c4; |
| const int c5 = td->c5; |
| const int c6 = td->c6; |
| const int c7 = td->c7; |
| int x, y; |
| |
| for (y = slice_start; y < slice_end; y += 2) { |
| for (x = 0; x < width; x += 2) { |
| const int u = srcpU[x >> 1] - 128; |
| const int v = srcpV[x >> 1] - 128; |
| const int uvval = c2 * u + c3 * v + 1081344; |
| dstpY[x + 0] = CB((65536 * (srcpY[x + 0] - 16) + uvval) >> 16); |
| dstpY[x + 1] = CB((65536 * (srcpY[x + 1] - 16) + uvval) >> 16); |
| dstpN[x + 0] = CB((65536 * (srcpN[x + 0] - 16) + uvval) >> 16); |
| dstpN[x + 1] = CB((65536 * (srcpN[x + 1] - 16) + uvval) >> 16); |
| dstpU[x >> 1] = CB((c4 * u + c5 * v + 8421376) >> 16); |
| dstpV[x >> 1] = CB((c6 * u + c7 * v + 8421376) >> 16); |
| } |
| srcpY += src_pitchY << 1; |
| dstpY += dst_pitchY << 1; |
| srcpN += src_pitchY << 1; |
| dstpN += dst_pitchY << 1; |
| srcpU += src_pitchUV; |
| srcpV += src_pitchUV; |
| dstpU += dst_pitchUV; |
| dstpV += dst_pitchUV; |
| } |
| |
| return 0; |
| } |
| |
| static int config_input(AVFilterLink *inlink) |
| { |
| AVFilterContext *ctx = inlink->dst; |
| ColorMatrixContext *color = ctx->priv; |
| const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(inlink->format); |
| |
| color->hsub = pix_desc->log2_chroma_w; |
| color->vsub = pix_desc->log2_chroma_h; |
| |
| av_log(ctx, AV_LOG_VERBOSE, "%s -> %s\n", |
| color_modes[color->source], color_modes[color->dest]); |
| |
| return 0; |
| } |
| |
| static int query_formats(AVFilterContext *ctx) |
| { |
| static const enum AVPixelFormat pix_fmts[] = { |
| AV_PIX_FMT_YUV444P, |
| AV_PIX_FMT_YUV422P, |
| AV_PIX_FMT_YUV420P, |
| AV_PIX_FMT_UYVY422, |
| AV_PIX_FMT_NONE |
| }; |
| AVFilterFormats *fmts_list = ff_make_format_list(pix_fmts); |
| if (!fmts_list) |
| return AVERROR(ENOMEM); |
| return ff_set_common_formats(ctx, fmts_list); |
| } |
| |
| static int filter_frame(AVFilterLink *link, AVFrame *in) |
| { |
| AVFilterContext *ctx = link->dst; |
| ColorMatrixContext *color = ctx->priv; |
| AVFilterLink *outlink = ctx->outputs[0]; |
| AVFrame *out; |
| ThreadData td = {0}; |
| |
| out = ff_get_video_buffer(outlink, outlink->w, outlink->h); |
| if (!out) { |
| av_frame_free(&in); |
| return AVERROR(ENOMEM); |
| } |
| av_frame_copy_props(out, in); |
| |
| if (color->source == COLOR_MODE_NONE) { |
| enum AVColorSpace cs = in->colorspace; |
| enum ColorMode source; |
| |
| switch(cs) { |
| case AVCOL_SPC_BT709 : source = COLOR_MODE_BT709 ; break; |
| case AVCOL_SPC_FCC : source = COLOR_MODE_FCC ; break; |
| case AVCOL_SPC_SMPTE240M : source = COLOR_MODE_SMPTE240M ; break; |
| case AVCOL_SPC_BT470BG : source = COLOR_MODE_BT601 ; break; |
| case AVCOL_SPC_SMPTE170M : source = COLOR_MODE_BT601 ; break; |
| case AVCOL_SPC_BT2020_NCL: source = COLOR_MODE_BT2020 ; break; |
| case AVCOL_SPC_BT2020_CL : source = COLOR_MODE_BT2020 ; break; |
| default : |
| av_log(ctx, AV_LOG_ERROR, "Input frame does not specify a supported colorspace, and none has been specified as source either\n"); |
| av_frame_free(&out); |
| return AVERROR(EINVAL); |
| } |
| color->mode = source * 5 + color->dest; |
| } else |
| color->mode = color->source * 5 + color->dest; |
| |
| switch(color->dest) { |
| case COLOR_MODE_BT709 : out->colorspace = AVCOL_SPC_BT709 ; break; |
| case COLOR_MODE_FCC : out->colorspace = AVCOL_SPC_FCC ; break; |
| case COLOR_MODE_SMPTE240M: out->colorspace = AVCOL_SPC_SMPTE240M ; break; |
| case COLOR_MODE_BT601 : out->colorspace = AVCOL_SPC_BT470BG ; break; |
| case COLOR_MODE_BT2020 : out->colorspace = AVCOL_SPC_BT2020_NCL; break; |
| } |
| |
| td.src = in; |
| td.dst = out; |
| td.c2 = color->yuv_convert[color->mode][0][1]; |
| td.c3 = color->yuv_convert[color->mode][0][2]; |
| td.c4 = color->yuv_convert[color->mode][1][1]; |
| td.c5 = color->yuv_convert[color->mode][1][2]; |
| td.c6 = color->yuv_convert[color->mode][2][1]; |
| td.c7 = color->yuv_convert[color->mode][2][2]; |
| |
| if (in->format == AV_PIX_FMT_YUV444P) |
| ctx->internal->execute(ctx, process_slice_yuv444p, &td, NULL, |
| FFMIN(in->height, ff_filter_get_nb_threads(ctx))); |
| else if (in->format == AV_PIX_FMT_YUV422P) |
| ctx->internal->execute(ctx, process_slice_yuv422p, &td, NULL, |
| FFMIN(in->height, ff_filter_get_nb_threads(ctx))); |
| else if (in->format == AV_PIX_FMT_YUV420P) |
| ctx->internal->execute(ctx, process_slice_yuv420p, &td, NULL, |
| FFMIN(in->height / 2, ff_filter_get_nb_threads(ctx))); |
| else |
| ctx->internal->execute(ctx, process_slice_uyvy422, &td, NULL, |
| FFMIN(in->height, ff_filter_get_nb_threads(ctx))); |
| |
| av_frame_free(&in); |
| return ff_filter_frame(outlink, out); |
| } |
| |
| static const AVFilterPad colormatrix_inputs[] = { |
| { |
| .name = "default", |
| .type = AVMEDIA_TYPE_VIDEO, |
| .config_props = config_input, |
| .filter_frame = filter_frame, |
| }, |
| { NULL } |
| }; |
| |
| static const AVFilterPad colormatrix_outputs[] = { |
| { |
| .name = "default", |
| .type = AVMEDIA_TYPE_VIDEO, |
| }, |
| { NULL } |
| }; |
| |
| AVFilter ff_vf_colormatrix = { |
| .name = "colormatrix", |
| .description = NULL_IF_CONFIG_SMALL("Convert color matrix."), |
| .priv_size = sizeof(ColorMatrixContext), |
| .init = init, |
| .query_formats = query_formats, |
| .inputs = colormatrix_inputs, |
| .outputs = colormatrix_outputs, |
| .priv_class = &colormatrix_class, |
| .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC | AVFILTER_FLAG_SLICE_THREADS, |
| }; |