/* * Copyright (C) 2012 British Broadcasting Corporation, All Rights Reserved * Author of de-interlace algorithm: Jim Easterbrook for BBC R&D * Based on the process described by Martin Weston for BBC R&D * Author of FFmpeg filter: Mark Himsley for BBC Broadcast Systems Development * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "libavutil/common.h" #include "libavutil/imgutils.h" #include "libavutil/opt.h" #include "libavutil/pixdesc.h" #include "avfilter.h" #include "formats.h" #include "internal.h" #include "video.h" #include "w3fdif.h" typedef struct W3FDIFContext { const AVClass *class; int filter; ///< 0 is simple, 1 is more complex int deint; ///< which frames to deinterlace int linesize[4]; ///< bytes of pixel data per line for each plane int planeheight[4]; ///< height of each plane int field; ///< which field are we on, 0 or 1 int eof; int nb_planes; AVFrame *prev, *cur, *next; ///< previous, current, next frames int32_t **work_line; ///< lines we are calculating int nb_threads; int max; W3FDIFDSPContext dsp; } W3FDIFContext; #define OFFSET(x) offsetof(W3FDIFContext, x) #define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM #define CONST(name, help, val, unit) { name, help, 0, AV_OPT_TYPE_CONST, {.i64=val}, 0, 0, FLAGS, unit } static const AVOption w3fdif_options[] = { { "filter", "specify the filter", OFFSET(filter), AV_OPT_TYPE_INT, {.i64=1}, 0, 1, FLAGS, "filter" }, CONST("simple", NULL, 0, "filter"), CONST("complex", NULL, 1, "filter"), { "deint", "specify which frames to deinterlace", OFFSET(deint), AV_OPT_TYPE_INT, {.i64=0}, 0, 1, FLAGS, "deint" }, CONST("all", "deinterlace all frames", 0, "deint"), CONST("interlaced", "only deinterlace frames marked as interlaced", 1, "deint"), { NULL } }; AVFILTER_DEFINE_CLASS(w3fdif); static int query_formats(AVFilterContext *ctx) { static const enum AVPixelFormat pix_fmts[] = { AV_PIX_FMT_YUV410P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ411P, AV_PIX_FMT_YUVA420P, AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUVA444P, AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRAP, AV_PIX_FMT_GRAY8, AV_PIX_FMT_YUV420P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV444P9, AV_PIX_FMT_YUV420P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV444P10, AV_PIX_FMT_YUV420P12, AV_PIX_FMT_YUV422P12, AV_PIX_FMT_YUV444P12, AV_PIX_FMT_YUV420P14, AV_PIX_FMT_YUV422P14, AV_PIX_FMT_YUV444P14, AV_PIX_FMT_GBRP9, AV_PIX_FMT_GBRP10, AV_PIX_FMT_GBRP12, AV_PIX_FMT_GBRP14, AV_PIX_FMT_NONE }; AVFilterFormats *fmts_list = ff_make_format_list(pix_fmts); if (!fmts_list) return AVERROR(ENOMEM); return ff_set_common_formats(ctx, fmts_list); } static void filter_simple_low(int32_t *work_line, uint8_t *in_lines_cur[2], const int16_t *coef, int linesize) { int i; for (i = 0; i < linesize; i++) { *work_line = *in_lines_cur[0]++ * coef[0]; *work_line++ += *in_lines_cur[1]++ * coef[1]; } } static void filter_complex_low(int32_t *work_line, uint8_t *in_lines_cur[4], const int16_t *coef, int linesize) { int i; for (i = 0; i < linesize; i++) { *work_line = *in_lines_cur[0]++ * coef[0]; *work_line += *in_lines_cur[1]++ * coef[1]; *work_line += *in_lines_cur[2]++ * coef[2]; *work_line++ += *in_lines_cur[3]++ * coef[3]; } } static void filter_simple_high(int32_t *work_line, uint8_t *in_lines_cur[3], uint8_t *in_lines_adj[3], const int16_t *coef, int linesize) { int i; for (i = 0; i < linesize; i++) { *work_line += *in_lines_cur[0]++ * coef[0]; *work_line += *in_lines_adj[0]++ * coef[0]; *work_line += *in_lines_cur[1]++ * coef[1]; *work_line += *in_lines_adj[1]++ * coef[1]; *work_line += *in_lines_cur[2]++ * coef[2]; *work_line++ += *in_lines_adj[2]++ * coef[2]; } } static void filter_complex_high(int32_t *work_line, uint8_t *in_lines_cur[5], uint8_t *in_lines_adj[5], const int16_t *coef, int linesize) { int i; for (i = 0; i < linesize; i++) { *work_line += *in_lines_cur[0]++ * coef[0]; *work_line += *in_lines_adj[0]++ * coef[0]; *work_line += *in_lines_cur[1]++ * coef[1]; *work_line += *in_lines_adj[1]++ * coef[1]; *work_line += *in_lines_cur[2]++ * coef[2]; *work_line += *in_lines_adj[2]++ * coef[2]; *work_line += *in_lines_cur[3]++ * coef[3]; *work_line += *in_lines_adj[3]++ * coef[3]; *work_line += *in_lines_cur[4]++ * coef[4]; *work_line++ += *in_lines_adj[4]++ * coef[4]; } } static void filter_scale(uint8_t *out_pixel, const int32_t *work_pixel, int linesize, int max) { int j; for (j = 0; j < linesize; j++, out_pixel++, work_pixel++) *out_pixel = av_clip(*work_pixel, 0, 255 * 256 * 128) >> 15; } static void filter16_simple_low(int32_t *work_line, uint8_t *in_lines_cur8[2], const int16_t *coef, int linesize) { uint16_t *in_lines_cur[2] = { (uint16_t *)in_lines_cur8[0], (uint16_t *)in_lines_cur8[1] }; int i; linesize /= 2; for (i = 0; i < linesize; i++) { *work_line = *in_lines_cur[0]++ * coef[0]; *work_line++ += *in_lines_cur[1]++ * coef[1]; } } static void filter16_complex_low(int32_t *work_line, uint8_t *in_lines_cur8[4], const int16_t *coef, int linesize) { uint16_t *in_lines_cur[4] = { (uint16_t *)in_lines_cur8[0], (uint16_t *)in_lines_cur8[1], (uint16_t *)in_lines_cur8[2], (uint16_t *)in_lines_cur8[3] }; int i; linesize /= 2; for (i = 0; i < linesize; i++) { *work_line = *in_lines_cur[0]++ * coef[0]; *work_line += *in_lines_cur[1]++ * coef[1]; *work_line += *in_lines_cur[2]++ * coef[2]; *work_line++ += *in_lines_cur[3]++ * coef[3]; } } static void filter16_simple_high(int32_t *work_line, uint8_t *in_lines_cur8[3], uint8_t *in_lines_adj8[3], const int16_t *coef, int linesize) { uint16_t *in_lines_cur[3] = { (uint16_t *)in_lines_cur8[0], (uint16_t *)in_lines_cur8[1], (uint16_t *)in_lines_cur8[2] }; uint16_t *in_lines_adj[3] = { (uint16_t *)in_lines_adj8[0], (uint16_t *)in_lines_adj8[1], (uint16_t *)in_lines_adj8[2] }; int i; linesize /= 2; for (i = 0; i < linesize; i++) { *work_line += *in_lines_cur[0]++ * coef[0]; *work_line += *in_lines_adj[0]++ * coef[0]; *work_line += *in_lines_cur[1]++ * coef[1]; *work_line += *in_lines_adj[1]++ * coef[1]; *work_line += *in_lines_cur[2]++ * coef[2]; *work_line++ += *in_lines_adj[2]++ * coef[2]; } } static void filter16_complex_high(int32_t *work_line, uint8_t *in_lines_cur8[5], uint8_t *in_lines_adj8[5], const int16_t *coef, int linesize) { uint16_t *in_lines_cur[5] = { (uint16_t *)in_lines_cur8[0], (uint16_t *)in_lines_cur8[1], (uint16_t *)in_lines_cur8[2], (uint16_t *)in_lines_cur8[3], (uint16_t *)in_lines_cur8[4] }; uint16_t *in_lines_adj[5] = { (uint16_t *)in_lines_adj8[0], (uint16_t *)in_lines_adj8[1], (uint16_t *)in_lines_adj8[2], (uint16_t *)in_lines_adj8[3], (uint16_t *)in_lines_adj8[4] }; int i; linesize /= 2; for (i = 0; i < linesize; i++) { *work_line += *in_lines_cur[0]++ * coef[0]; *work_line += *in_lines_adj[0]++ * coef[0]; *work_line += *in_lines_cur[1]++ * coef[1]; *work_line += *in_lines_adj[1]++ * coef[1]; *work_line += *in_lines_cur[2]++ * coef[2]; *work_line += *in_lines_adj[2]++ * coef[2]; *work_line += *in_lines_cur[3]++ * coef[3]; *work_line += *in_lines_adj[3]++ * coef[3]; *work_line += *in_lines_cur[4]++ * coef[4]; *work_line++ += *in_lines_adj[4]++ * coef[4]; } } static void filter16_scale(uint8_t *out_pixel8, const int32_t *work_pixel, int linesize, int max) { uint16_t *out_pixel = (uint16_t *)out_pixel8; int j; linesize /= 2; for (j = 0; j < linesize; j++, out_pixel++, work_pixel++) *out_pixel = av_clip(*work_pixel, 0, max) >> 15; } static int config_input(AVFilterLink *inlink) { AVFilterContext *ctx = inlink->dst; W3FDIFContext *s = ctx->priv; const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format); int ret, i, depth; if ((ret = av_image_fill_linesizes(s->linesize, inlink->format, inlink->w)) < 0) return ret; s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h); s->planeheight[0] = s->planeheight[3] = inlink->h; if (inlink->h < 3) { av_log(ctx, AV_LOG_ERROR, "Video of less than 3 lines is not supported\n"); return AVERROR(EINVAL); } s->nb_planes = av_pix_fmt_count_planes(inlink->format); s->nb_threads = ff_filter_get_nb_threads(ctx); s->work_line = av_calloc(s->nb_threads, sizeof(*s->work_line)); if (!s->work_line) return AVERROR(ENOMEM); for (i = 0; i < s->nb_threads; i++) { s->work_line[i] = av_calloc(FFALIGN(s->linesize[0], 32), sizeof(*s->work_line[0])); if (!s->work_line[i]) return AVERROR(ENOMEM); } depth = desc->comp[0].depth; s->max = ((1 << depth) - 1) * 256 * 128; if (depth <= 8) { s->dsp.filter_simple_low = filter_simple_low; s->dsp.filter_complex_low = filter_complex_low; s->dsp.filter_simple_high = filter_simple_high; s->dsp.filter_complex_high = filter_complex_high; s->dsp.filter_scale = filter_scale; } else { s->dsp.filter_simple_low = filter16_simple_low; s->dsp.filter_complex_low = filter16_complex_low; s->dsp.filter_simple_high = filter16_simple_high; s->dsp.filter_complex_high = filter16_complex_high; s->dsp.filter_scale = filter16_scale; } if (ARCH_X86) ff_w3fdif_init_x86(&s->dsp, depth); return 0; } static int config_output(AVFilterLink *outlink) { AVFilterLink *inlink = outlink->src->inputs[0]; outlink->time_base.num = inlink->time_base.num; outlink->time_base.den = inlink->time_base.den * 2; outlink->frame_rate.num = inlink->frame_rate.num * 2; outlink->frame_rate.den = inlink->frame_rate.den; return 0; } /* * Filter coefficients from PH-2071, scaled by 256 * 128. * Each set of coefficients has a set for low-frequencies and high-frequencies. * n_coef_lf[] and n_coef_hf[] are the number of coefs for simple and more-complex. * It is important for later that n_coef_lf[] is even and n_coef_hf[] is odd. * coef_lf[][] and coef_hf[][] are the coefficients for low-frequencies * and high-frequencies for simple and more-complex mode. */ static const int8_t n_coef_lf[2] = { 2, 4 }; static const int16_t coef_lf[2][4] = {{ 16384, 16384, 0, 0}, { -852, 17236, 17236, -852}}; static const int8_t n_coef_hf[2] = { 3, 5 }; static const int16_t coef_hf[2][5] = {{ -2048, 4096, -2048, 0, 0}, { 1016, -3801, 5570, -3801, 1016}}; typedef struct ThreadData { AVFrame *out, *cur, *adj; int plane; } ThreadData; static int deinterlace_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs) { W3FDIFContext *s = ctx->priv; ThreadData *td = arg; AVFrame *out = td->out; AVFrame *cur = td->cur; AVFrame *adj = td->adj; const int plane = td->plane; const int filter = s->filter; uint8_t *in_line, *in_lines_cur[5], *in_lines_adj[5]; uint8_t *out_line, *out_pixel; int32_t *work_line, *work_pixel; uint8_t *cur_data = cur->data[plane]; uint8_t *adj_data = adj->data[plane]; uint8_t *dst_data = out->data[plane]; const int linesize = s->linesize[plane]; const int height = s->planeheight[plane]; const int cur_line_stride = cur->linesize[plane]; const int adj_line_stride = adj->linesize[plane]; const int dst_line_stride = out->linesize[plane]; const int start = (height * jobnr) / nb_jobs; const int end = (height * (jobnr+1)) / nb_jobs; const int max = s->max; int j, y_in, y_out; /* copy unchanged the lines of the field */ y_out = start + ((s->field == cur->top_field_first) ^ (start & 1)); in_line = cur_data + (y_out * cur_line_stride); out_line = dst_data + (y_out * dst_line_stride); while (y_out < end) { memcpy(out_line, in_line, linesize); y_out += 2; in_line += cur_line_stride * 2; out_line += dst_line_stride * 2; } /* interpolate other lines of the field */ y_out = start + ((s->field != cur->top_field_first) ^ (start & 1)); out_line = dst_data + (y_out * dst_line_stride); while (y_out < end) { /* get low vertical frequencies from current field */ for (j = 0; j < n_coef_lf[filter]; j++) { y_in = (y_out + 1) + (j * 2) - n_coef_lf[filter]; while (y_in < 0) y_in += 2; while (y_in >= height) y_in -= 2; in_lines_cur[j] = cur_data + (y_in * cur_line_stride); } work_line = s->work_line[jobnr]; switch (n_coef_lf[filter]) { case 2: s->dsp.filter_simple_low(work_line, in_lines_cur, coef_lf[filter], linesize); break; case 4: s->dsp.filter_complex_low(work_line, in_lines_cur, coef_lf[filter], linesize); } /* get high vertical frequencies from adjacent fields */ for (j = 0; j < n_coef_hf[filter]; j++) { y_in = (y_out + 1) + (j * 2) - n_coef_hf[filter]; while (y_in < 0) y_in += 2; while (y_in >= height) y_in -= 2; in_lines_cur[j] = cur_data + (y_in * cur_line_stride); in_lines_adj[j] = adj_data + (y_in * adj_line_stride); } work_line = s->work_line[jobnr]; switch (n_coef_hf[filter]) { case 3: s->dsp.filter_simple_high(work_line, in_lines_cur, in_lines_adj, coef_hf[filter], linesize); break; case 5: s->dsp.filter_complex_high(work_line, in_lines_cur, in_lines_adj, coef_hf[filter], linesize); } /* save scaled result to the output frame, scaling down by 256 * 128 */ work_pixel = s->work_line[jobnr]; out_pixel = out_line; s->dsp.filter_scale(out_pixel, work_pixel, linesize, max); /* move on to next line */ y_out += 2; out_line += dst_line_stride * 2; } return 0; } static int filter(AVFilterContext *ctx, int is_second) { W3FDIFContext *s = ctx->priv; AVFilterLink *outlink = ctx->outputs[0]; AVFrame *out, *adj; ThreadData td; int plane; out = ff_get_video_buffer(outlink, outlink->w, outlink->h); if (!out) return AVERROR(ENOMEM); av_frame_copy_props(out, s->cur); out->interlaced_frame = 0; if (!is_second) { if (out->pts != AV_NOPTS_VALUE) out->pts *= 2; } else { int64_t cur_pts = s->cur->pts; int64_t next_pts = s->next->pts; if (next_pts != AV_NOPTS_VALUE && cur_pts != AV_NOPTS_VALUE) { out->pts = cur_pts + next_pts; } else { out->pts = AV_NOPTS_VALUE; } } adj = s->field ? s->next : s->prev; td.out = out; td.cur = s->cur; td.adj = adj; for (plane = 0; plane < s->nb_planes; plane++) { td.plane = plane; ctx->internal->execute(ctx, deinterlace_slice, &td, NULL, FFMIN(s->planeheight[plane], s->nb_threads)); } s->field = !s->field; return ff_filter_frame(outlink, out); } static int filter_frame(AVFilterLink *inlink, AVFrame *frame) { AVFilterContext *ctx = inlink->dst; W3FDIFContext *s = ctx->priv; int ret; av_frame_free(&s->prev); s->prev = s->cur; s->cur = s->next; s->next = frame; if (!s->cur) { s->cur = av_frame_clone(s->next); if (!s->cur) return AVERROR(ENOMEM); } if ((s->deint && !s->cur->interlaced_frame) || ctx->is_disabled) { AVFrame *out = av_frame_clone(s->cur); if (!out) return AVERROR(ENOMEM); av_frame_free(&s->prev); if (out->pts != AV_NOPTS_VALUE) out->pts *= 2; return ff_filter_frame(ctx->outputs[0], out); } if (!s->prev) return 0; ret = filter(ctx, 0); if (ret < 0) return ret; return filter(ctx, 1); } static int request_frame(AVFilterLink *outlink) { AVFilterContext *ctx = outlink->src; W3FDIFContext *s = ctx->priv; int ret; if (s->eof) return AVERROR_EOF; ret = ff_request_frame(ctx->inputs[0]); if (ret == AVERROR_EOF && s->cur) { AVFrame *next = av_frame_clone(s->next); if (!next) return AVERROR(ENOMEM); next->pts = s->next->pts * 2 - s->cur->pts; filter_frame(ctx->inputs[0], next); s->eof = 1; } else if (ret < 0) { return ret; } return 0; } static av_cold void uninit(AVFilterContext *ctx) { W3FDIFContext *s = ctx->priv; int i; av_frame_free(&s->prev); av_frame_free(&s->cur ); av_frame_free(&s->next); for (i = 0; i < s->nb_threads; i++) av_freep(&s->work_line[i]); av_freep(&s->work_line); } static const AVFilterPad w3fdif_inputs[] = { { .name = "default", .type = AVMEDIA_TYPE_VIDEO, .filter_frame = filter_frame, .config_props = config_input, }, { NULL } }; static const AVFilterPad w3fdif_outputs[] = { { .name = "default", .type = AVMEDIA_TYPE_VIDEO, .config_props = config_output, .request_frame = request_frame, }, { NULL } }; AVFilter ff_vf_w3fdif = { .name = "w3fdif", .description = NULL_IF_CONFIG_SMALL("Apply Martin Weston three field deinterlace."), .priv_size = sizeof(W3FDIFContext), .priv_class = &w3fdif_class, .uninit = uninit, .query_formats = query_formats, .inputs = w3fdif_inputs, .outputs = w3fdif_outputs, .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_INTERNAL | AVFILTER_FLAG_SLICE_THREADS, };