Commit 19804024 authored by Nicolas George's avatar Nicolas George

lavfi/vf_overlay: move to framesync2.

parent 0ae8df41
......@@ -245,7 +245,7 @@ OBJS-$(CONFIG_OCR_FILTER) += vf_ocr.o
OBJS-$(CONFIG_OCV_FILTER) += vf_libopencv.o
OBJS-$(CONFIG_OPENCL) += deshake_opencl.o unsharp_opencl.o
OBJS-$(CONFIG_OSCILLOSCOPE_FILTER) += vf_datascope.o
OBJS-$(CONFIG_OVERLAY_FILTER) += vf_overlay.o dualinput.o framesync.o
OBJS-$(CONFIG_OVERLAY_FILTER) += vf_overlay.o framesync2.o
OBJS-$(CONFIG_OWDENOISE_FILTER) += vf_owdenoise.o
OBJS-$(CONFIG_PAD_FILTER) += vf_pad.o
OBJS-$(CONFIG_PALETTEGEN_FILTER) += vf_palettegen.o
......
......@@ -36,8 +36,8 @@
#include "libavutil/opt.h"
#include "libavutil/timestamp.h"
#include "internal.h"
#include "dualinput.h"
#include "drawutils.h"
#include "framesync2.h"
#include "video.h"
static const char *const var_names[] = {
......@@ -121,7 +121,7 @@ typedef struct OverlayContext {
int format; ///< OverlayFormat
int eval_mode; ///< EvalMode
FFDualInputContext dinput;
FFFrameSync fs;
int main_pix_step[4]; ///< steps per pixel for each plane of the main output
int overlay_pix_step[4]; ///< steps per pixel for each plane of the overlay
......@@ -132,6 +132,8 @@ typedef struct OverlayContext {
char *x_expr, *y_expr;
int eof_action; ///< action to take on EOF from source
int opt_shortest;
int opt_repeatlast;
AVExpr *x_pexpr, *y_pexpr;
......@@ -142,7 +144,7 @@ static av_cold void uninit(AVFilterContext *ctx)
{
OverlayContext *s = ctx->priv;
ff_dualinput_uninit(&s->dinput);
ff_framesync2_uninit(&s->fs);
av_expr_free(s->x_pexpr); s->x_pexpr = NULL;
av_expr_free(s->y_pexpr); s->y_pexpr = NULL;
}
......@@ -390,14 +392,20 @@ static int config_output(AVFilterLink *outlink)
OverlayContext *s = ctx->priv;
int ret;
if ((ret = ff_dualinput_init(ctx, &s->dinput)) < 0)
if ((ret = ff_framesync2_init_dualinput(&s->fs, ctx)) < 0)
return ret;
if (s->opt_shortest)
s->fs.in[0].after = s->fs.in[1].after = EXT_STOP;
if (!s->opt_repeatlast) {
s->fs.in[1].after = EXT_NULL;
s->fs.in[1].sync = 0;
}
outlink->w = ctx->inputs[MAIN]->w;
outlink->h = ctx->inputs[MAIN]->h;
outlink->time_base = ctx->inputs[MAIN]->time_base;
return 0;
return ff_framesync2_configure(&s->fs);
}
// divide by 255 and round to nearest
......@@ -770,11 +778,19 @@ static int config_input_main(AVFilterLink *inlink)
return 0;
}
static AVFrame *do_blend(AVFilterContext *ctx, AVFrame *mainpic,
const AVFrame *second)
static int do_blend(FFFrameSync *fs)
{
AVFilterContext *ctx = fs->parent;
AVFrame *mainpic, *second;
OverlayContext *s = ctx->priv;
AVFilterLink *inlink = ctx->inputs[0];
int ret;
ret = ff_framesync2_dualinput_get_writable(fs, &mainpic, &second);
if (ret < 0)
return ret;
if (!second)
return ff_filter_frame(ctx->outputs[0], mainpic);
if (s->eval_mode == EVAL_MODE_FRAME) {
int64_t pos = mainpic->pkt_pos;
......@@ -799,39 +815,32 @@ static AVFrame *do_blend(AVFilterContext *ctx, AVFrame *mainpic,
if (s->x < mainpic->width && s->x + second->width >= 0 ||
s->y < mainpic->height && s->y + second->height >= 0)
s->blend_image(ctx, mainpic, second, s->x, s->y);
return mainpic;
}
static int filter_frame(AVFilterLink *inlink, AVFrame *inpicref)
{
OverlayContext *s = inlink->dst->priv;
av_log(inlink->dst, AV_LOG_DEBUG, "Incoming frame (time:%s) from link #%d\n", av_ts2timestr(inpicref->pts, &inlink->time_base), FF_INLINK_IDX(inlink));
return ff_dualinput_filter_frame(&s->dinput, inlink, inpicref);
}
static int request_frame(AVFilterLink *outlink)
{
OverlayContext *s = outlink->src->priv;
return ff_dualinput_request_frame(&s->dinput, outlink);
return ff_filter_frame(ctx->outputs[0], mainpic);
}
static av_cold int init(AVFilterContext *ctx)
{
OverlayContext *s = ctx->priv;
if (!s->dinput.repeatlast || s->eof_action == EOF_ACTION_PASS) {
s->dinput.repeatlast = 0;
if (!s->opt_repeatlast || s->eof_action == EOF_ACTION_PASS) {
s->opt_repeatlast = 0;
s->eof_action = EOF_ACTION_PASS;
}
if (s->dinput.shortest || s->eof_action == EOF_ACTION_ENDALL) {
s->dinput.shortest = 1;
if (s->opt_shortest || s->eof_action == EOF_ACTION_ENDALL) {
s->opt_shortest = 1;
s->eof_action = EOF_ACTION_ENDALL;
}
s->dinput.process = do_blend;
s->fs.on_event = do_blend;
return 0;
}
static int activate(AVFilterContext *ctx)
{
OverlayContext *s = ctx->priv;
return ff_framesync2_activate(&s->fs);
}
#define OFFSET(x) offsetof(OverlayContext, x)
#define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
......@@ -847,7 +856,7 @@ static const AVOption overlay_options[] = {
{ "eval", "specify when to evaluate expressions", OFFSET(eval_mode), AV_OPT_TYPE_INT, {.i64 = EVAL_MODE_FRAME}, 0, EVAL_MODE_NB-1, FLAGS, "eval" },
{ "init", "eval expressions once during initialization", 0, AV_OPT_TYPE_CONST, {.i64=EVAL_MODE_INIT}, .flags = FLAGS, .unit = "eval" },
{ "frame", "eval expressions per-frame", 0, AV_OPT_TYPE_CONST, {.i64=EVAL_MODE_FRAME}, .flags = FLAGS, .unit = "eval" },
{ "shortest", "force termination when the shortest input terminates", OFFSET(dinput.shortest), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS },
{ "shortest", "force termination when the shortest input terminates", OFFSET(opt_shortest), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS },
{ "format", "set output format", OFFSET(format), AV_OPT_TYPE_INT, {.i64=OVERLAY_FORMAT_YUV420}, 0, OVERLAY_FORMAT_NB-1, FLAGS, "format" },
{ "yuv420", "", 0, AV_OPT_TYPE_CONST, {.i64=OVERLAY_FORMAT_YUV420}, .flags = FLAGS, .unit = "format" },
{ "yuv422", "", 0, AV_OPT_TYPE_CONST, {.i64=OVERLAY_FORMAT_YUV422}, .flags = FLAGS, .unit = "format" },
......@@ -855,7 +864,7 @@ static const AVOption overlay_options[] = {
{ "rgb", "", 0, AV_OPT_TYPE_CONST, {.i64=OVERLAY_FORMAT_RGB}, .flags = FLAGS, .unit = "format" },
{ "gbrp", "", 0, AV_OPT_TYPE_CONST, {.i64=OVERLAY_FORMAT_GBRP}, .flags = FLAGS, .unit = "format" },
{ "auto", "", 0, AV_OPT_TYPE_CONST, {.i64=OVERLAY_FORMAT_AUTO}, .flags = FLAGS, .unit = "format" },
{ "repeatlast", "repeat overlay of the last overlay frame", OFFSET(dinput.repeatlast), AV_OPT_TYPE_BOOL, {.i64=1}, 0, 1, FLAGS },
{ "repeatlast", "repeat overlay of the last overlay frame", OFFSET(opt_repeatlast), AV_OPT_TYPE_BOOL, {.i64=1}, 0, 1, FLAGS },
{ NULL }
};
......@@ -866,14 +875,11 @@ static const AVFilterPad avfilter_vf_overlay_inputs[] = {
.name = "main",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_input_main,
.filter_frame = filter_frame,
.needs_writable = 1,
},
{
.name = "overlay",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_input_overlay,
.filter_frame = filter_frame,
},
{ NULL }
};
......@@ -883,7 +889,6 @@ static const AVFilterPad avfilter_vf_overlay_outputs[] = {
.name = "default",
.type = AVMEDIA_TYPE_VIDEO,
.config_props = config_output,
.request_frame = request_frame,
},
{ NULL }
};
......@@ -896,6 +901,7 @@ AVFilter ff_vf_overlay = {
.priv_size = sizeof(OverlayContext),
.priv_class = &overlay_class,
.query_formats = query_formats,
.activate = activate,
.process_command = process_command,
.inputs = avfilter_vf_overlay_inputs,
.outputs = avfilter_vf_overlay_outputs,
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment