Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Contribute to GitLab
Sign in / Register
Toggle navigation
F
ffmpeg.wasm-core
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Linshizhi
ffmpeg.wasm-core
Commits
5d774301
Commit
5d774301
authored
Aug 24, 2016
by
Paul B Mahol
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
avfilter: add lut2 filter
parent
01aee814
Hide whitespace changes
Inline
Side-by-side
Showing
6 changed files
with
415 additions
and
1 deletion
+415
-1
Changelog
Changelog
+1
-0
filters.texi
doc/filters.texi
+38
-0
Makefile
libavfilter/Makefile
+1
-0
allfilters.c
libavfilter/allfilters.c
+1
-0
version.h
libavfilter/version.h
+1
-1
vf_lut2.c
libavfilter/vf_lut2.c
+373
-0
No files found.
Changelog
View file @
5d774301
...
@@ -20,6 +20,7 @@ version <next>:
...
@@ -20,6 +20,7 @@ version <next>:
- fifo muxer
- fifo muxer
- maskedclamp filter
- maskedclamp filter
- hysteresis filter
- hysteresis filter
- lut2 filter
version 3.1:
version 3.1:
...
...
doc/filters.texi
View file @
5d774301
...
@@ -9253,6 +9253,44 @@ lutyuv=u='(val-maxval/2)*2+maxval/2':v='(val-maxval/2)*2+maxval/2'
...
@@ -9253,6 +9253,44 @@ lutyuv=u='(val-maxval/2)*2+maxval/2':v='(val-maxval/2)*2+maxval/2'
@end example
@end example
@end itemize
@end itemize
@section lut2
Compute and apply a lookup table from two video inputs.
This filter accepts the following parameters:
@table @option
@item c0
set first pixel component expression
@item c1
set second pixel component expression
@item c2
set third pixel component expression
@item c3
set fourth pixel component expression, corresponds to the alpha component
@end table
Each of them specifies the expression to use for computing the lookup table for
the corresponding pixel component values.
The exact component associated to each of the @var{c*} options depends on the
format in inputs.
The expressions can contain the following constants:
@table @option
@item w
@item h
The input width and height.
@item x
The first input value for the pixel component.
@item y
The second input value for the pixel component.
@end table
All expressions default to "x".
@section maskedclamp
@section maskedclamp
Clamp the first input stream with the second input and third input stream.
Clamp the first input stream with the second input and third input stream.
...
...
libavfilter/Makefile
View file @
5d774301
...
@@ -205,6 +205,7 @@ OBJS-$(CONFIG_KERNDEINT_FILTER) += vf_kerndeint.o
...
@@ -205,6 +205,7 @@ OBJS-$(CONFIG_KERNDEINT_FILTER) += vf_kerndeint.o
OBJS-$(CONFIG_LENSCORRECTION_FILTER)
+=
vf_lenscorrection.o
OBJS-$(CONFIG_LENSCORRECTION_FILTER)
+=
vf_lenscorrection.o
OBJS-$(CONFIG_LOOP_FILTER)
+=
f_loop.o
OBJS-$(CONFIG_LOOP_FILTER)
+=
f_loop.o
OBJS-$(CONFIG_LUT_FILTER)
+=
vf_lut.o
OBJS-$(CONFIG_LUT_FILTER)
+=
vf_lut.o
OBJS-$(CONFIG_LUT2_FILTER)
+=
vf_lut2.o
framesync.o
OBJS-$(CONFIG_LUT3D_FILTER)
+=
vf_lut3d.o
OBJS-$(CONFIG_LUT3D_FILTER)
+=
vf_lut3d.o
OBJS-$(CONFIG_LUTRGB_FILTER)
+=
vf_lut.o
OBJS-$(CONFIG_LUTRGB_FILTER)
+=
vf_lut.o
OBJS-$(CONFIG_LUTYUV_FILTER)
+=
vf_lut.o
OBJS-$(CONFIG_LUTYUV_FILTER)
+=
vf_lut.o
...
...
libavfilter/allfilters.c
View file @
5d774301
...
@@ -222,6 +222,7 @@ void avfilter_register_all(void)
...
@@ -222,6 +222,7 @@ void avfilter_register_all(void)
REGISTER_FILTER
(
LENSCORRECTION
,
lenscorrection
,
vf
);
REGISTER_FILTER
(
LENSCORRECTION
,
lenscorrection
,
vf
);
REGISTER_FILTER
(
LOOP
,
loop
,
vf
);
REGISTER_FILTER
(
LOOP
,
loop
,
vf
);
REGISTER_FILTER
(
LUT
,
lut
,
vf
);
REGISTER_FILTER
(
LUT
,
lut
,
vf
);
REGISTER_FILTER
(
LUT2
,
lut2
,
vf
);
REGISTER_FILTER
(
LUT3D
,
lut3d
,
vf
);
REGISTER_FILTER
(
LUT3D
,
lut3d
,
vf
);
REGISTER_FILTER
(
LUTRGB
,
lutrgb
,
vf
);
REGISTER_FILTER
(
LUTRGB
,
lutrgb
,
vf
);
REGISTER_FILTER
(
LUTYUV
,
lutyuv
,
vf
);
REGISTER_FILTER
(
LUTYUV
,
lutyuv
,
vf
);
...
...
libavfilter/version.h
View file @
5d774301
...
@@ -30,7 +30,7 @@
...
@@ -30,7 +30,7 @@
#include "libavutil/version.h"
#include "libavutil/version.h"
#define LIBAVFILTER_VERSION_MAJOR 6
#define LIBAVFILTER_VERSION_MAJOR 6
#define LIBAVFILTER_VERSION_MINOR 5
4
#define LIBAVFILTER_VERSION_MINOR 5
5
#define LIBAVFILTER_VERSION_MICRO 100
#define LIBAVFILTER_VERSION_MICRO 100
#define LIBAVFILTER_VERSION_INT AV_VERSION_INT(LIBAVFILTER_VERSION_MAJOR, \
#define LIBAVFILTER_VERSION_INT AV_VERSION_INT(LIBAVFILTER_VERSION_MAJOR, \
...
...
libavfilter/vf_lut2.c
0 → 100644
View file @
5d774301
/*
* Copyright (c) 2016 Paul B Mahol
*
* This file is part of FFmpeg.
*
* FFmpeg is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* FFmpeg is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with FFmpeg; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
*/
#include "libavutil/attributes.h"
#include "libavutil/common.h"
#include "libavutil/eval.h"
#include "libavutil/opt.h"
#include "libavutil/pixdesc.h"
#include "avfilter.h"
#include "drawutils.h"
#include "formats.h"
#include "internal.h"
#include "video.h"
#include "framesync.h"
static
const
char
*
const
var_names
[]
=
{
"w"
,
///< width of the input video
"h"
,
///< height of the input video
"x"
,
///< input value for the pixel from input #1
"y"
,
///< input value for the pixel from input #2
NULL
};
enum
var_name
{
VAR_W
,
VAR_H
,
VAR_X
,
VAR_Y
,
VAR_VARS_NB
};
typedef
struct
LUT2Context
{
const
AVClass
*
class
;
char
*
comp_expr_str
[
4
];
AVExpr
*
comp_expr
[
4
];
double
var_values
[
VAR_VARS_NB
];
uint16_t
*
lut
[
4
];
///< lookup table for each component
int
width
[
4
],
height
[
4
];
int
nb_planes
;
int
depth
,
depthx
,
depthy
;
void
(
*
lut2
)(
struct
LUT2Context
*
s
,
AVFrame
*
dst
,
AVFrame
*
srcx
,
AVFrame
*
srcy
);
FFFrameSync
fs
;
}
LUT2Context
;
#define OFFSET(x) offsetof(LUT2Context, x)
#define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
static
const
AVOption
lut2_options
[]
=
{
{
"c0"
,
"set component #0 expression"
,
OFFSET
(
comp_expr_str
[
0
]),
AV_OPT_TYPE_STRING
,
{
.
str
=
"x"
},
.
flags
=
FLAGS
},
{
"c1"
,
"set component #1 expression"
,
OFFSET
(
comp_expr_str
[
1
]),
AV_OPT_TYPE_STRING
,
{
.
str
=
"x"
},
.
flags
=
FLAGS
},
{
"c2"
,
"set component #2 expression"
,
OFFSET
(
comp_expr_str
[
2
]),
AV_OPT_TYPE_STRING
,
{
.
str
=
"x"
},
.
flags
=
FLAGS
},
{
"c3"
,
"set component #3 expression"
,
OFFSET
(
comp_expr_str
[
3
]),
AV_OPT_TYPE_STRING
,
{
.
str
=
"x"
},
.
flags
=
FLAGS
},
{
NULL
}
};
static
av_cold
void
uninit
(
AVFilterContext
*
ctx
)
{
LUT2Context
*
s
=
ctx
->
priv
;
int
i
;
for
(
i
=
0
;
i
<
4
;
i
++
)
{
av_expr_free
(
s
->
comp_expr
[
i
]);
s
->
comp_expr
[
i
]
=
NULL
;
av_freep
(
&
s
->
comp_expr_str
[
i
]);
av_freep
(
&
s
->
lut
[
i
]);
}
}
static
int
query_formats
(
AVFilterContext
*
ctx
)
{
static
const
enum
AVPixelFormat
pix_fmts
[]
=
{
AV_PIX_FMT_YUVA444P
,
AV_PIX_FMT_YUV444P
,
AV_PIX_FMT_YUV440P
,
AV_PIX_FMT_YUVJ444P
,
AV_PIX_FMT_YUVJ440P
,
AV_PIX_FMT_YUVA422P
,
AV_PIX_FMT_YUV422P
,
AV_PIX_FMT_YUVA420P
,
AV_PIX_FMT_YUV420P
,
AV_PIX_FMT_YUVJ422P
,
AV_PIX_FMT_YUVJ420P
,
AV_PIX_FMT_YUVJ411P
,
AV_PIX_FMT_YUV411P
,
AV_PIX_FMT_YUV410P
,
AV_PIX_FMT_YUV420P9
,
AV_PIX_FMT_YUV422P9
,
AV_PIX_FMT_YUV444P9
,
AV_PIX_FMT_YUV420P10
,
AV_PIX_FMT_YUV422P10
,
AV_PIX_FMT_YUV444P10
,
AV_PIX_FMT_YUV420P12
,
AV_PIX_FMT_YUV422P12
,
AV_PIX_FMT_YUV444P12
,
AV_PIX_FMT_YUV440P12
,
AV_PIX_FMT_YUVA420P9
,
AV_PIX_FMT_YUVA422P9
,
AV_PIX_FMT_YUVA444P9
,
AV_PIX_FMT_YUVA420P10
,
AV_PIX_FMT_YUVA422P10
,
AV_PIX_FMT_YUVA444P10
,
AV_PIX_FMT_GBRP
,
AV_PIX_FMT_GBRP9
,
AV_PIX_FMT_GBRP10
,
AV_PIX_FMT_GBRP12
,
AV_PIX_FMT_GBRAP
,
AV_PIX_FMT_GBRAP12
,
AV_PIX_FMT_GRAY8
,
AV_PIX_FMT_NONE
};
return
ff_set_common_formats
(
ctx
,
ff_make_format_list
(
pix_fmts
));
}
static
int
config_inputx
(
AVFilterLink
*
inlink
)
{
AVFilterContext
*
ctx
=
inlink
->
dst
;
LUT2Context
*
s
=
ctx
->
priv
;
const
AVPixFmtDescriptor
*
desc
=
av_pix_fmt_desc_get
(
inlink
->
format
);
int
hsub
=
desc
->
log2_chroma_w
;
int
vsub
=
desc
->
log2_chroma_h
;
s
->
nb_planes
=
av_pix_fmt_count_planes
(
inlink
->
format
);
s
->
height
[
1
]
=
s
->
height
[
2
]
=
AV_CEIL_RSHIFT
(
inlink
->
h
,
vsub
);
s
->
height
[
0
]
=
s
->
height
[
3
]
=
inlink
->
h
;
s
->
width
[
1
]
=
s
->
width
[
2
]
=
AV_CEIL_RSHIFT
(
inlink
->
w
,
hsub
);
s
->
width
[
0
]
=
s
->
width
[
3
]
=
inlink
->
w
;
s
->
var_values
[
VAR_W
]
=
inlink
->
w
;
s
->
var_values
[
VAR_H
]
=
inlink
->
h
;
s
->
depthx
=
desc
->
comp
[
0
].
depth
;
return
0
;
}
static
int
config_inputy
(
AVFilterLink
*
inlink
)
{
AVFilterContext
*
ctx
=
inlink
->
dst
;
LUT2Context
*
s
=
ctx
->
priv
;
const
AVPixFmtDescriptor
*
desc
=
av_pix_fmt_desc_get
(
inlink
->
format
);
s
->
depthy
=
desc
->
comp
[
0
].
depth
;
return
0
;
}
static
void
lut2_8bit
(
struct
LUT2Context
*
s
,
AVFrame
*
out
,
AVFrame
*
srcx
,
AVFrame
*
srcy
)
{
int
p
,
y
,
x
;
for
(
p
=
0
;
p
<
s
->
nb_planes
;
p
++
)
{
const
uint16_t
*
lut
=
s
->
lut
[
p
];
const
uint8_t
*
srcxx
,
*
srcyy
;
uint8_t
*
dst
;
dst
=
out
->
data
[
p
];
srcxx
=
srcx
->
data
[
p
];
srcyy
=
srcy
->
data
[
p
];
for
(
y
=
0
;
y
<
s
->
height
[
p
];
y
++
)
{
for
(
x
=
0
;
x
<
s
->
width
[
p
];
x
++
)
{
dst
[
x
]
=
lut
[(
srcyy
[
x
]
<<
s
->
depthx
)
|
srcxx
[
x
]];
}
dst
+=
out
->
linesize
[
p
];
srcxx
+=
srcx
->
linesize
[
p
];
srcyy
+=
srcy
->
linesize
[
p
];
}
}
}
static
void
lut2_16bit
(
struct
LUT2Context
*
s
,
AVFrame
*
out
,
AVFrame
*
srcx
,
AVFrame
*
srcy
)
{
int
p
,
y
,
x
;
for
(
p
=
0
;
p
<
s
->
nb_planes
;
p
++
)
{
const
uint16_t
*
lut
=
s
->
lut
[
p
];
const
uint16_t
*
srcxx
,
*
srcyy
;
uint16_t
*
dst
;
dst
=
(
uint16_t
*
)
out
->
data
[
p
];
srcxx
=
(
uint16_t
*
)
srcx
->
data
[
p
];
srcyy
=
(
uint16_t
*
)
srcy
->
data
[
p
];
for
(
y
=
0
;
y
<
s
->
height
[
p
];
y
++
)
{
for
(
x
=
0
;
x
<
s
->
width
[
p
];
x
++
)
{
dst
[
x
]
=
lut
[(
srcyy
[
x
]
<<
s
->
depthx
)
|
srcxx
[
x
]];
}
dst
+=
out
->
linesize
[
p
]
/
2
;
srcxx
+=
srcx
->
linesize
[
p
]
/
2
;
srcyy
+=
srcy
->
linesize
[
p
]
/
2
;
}
}
}
static
int
process_frame
(
FFFrameSync
*
fs
)
{
AVFilterContext
*
ctx
=
fs
->
parent
;
LUT2Context
*
s
=
fs
->
opaque
;
AVFilterLink
*
outlink
=
ctx
->
outputs
[
0
];
AVFrame
*
out
,
*
srcx
,
*
srcy
;
int
ret
;
if
((
ret
=
ff_framesync_get_frame
(
&
s
->
fs
,
0
,
&
srcx
,
0
))
<
0
||
(
ret
=
ff_framesync_get_frame
(
&
s
->
fs
,
1
,
&
srcy
,
0
))
<
0
)
return
ret
;
if
(
ctx
->
is_disabled
)
{
out
=
av_frame_clone
(
srcx
);
if
(
!
out
)
return
AVERROR
(
ENOMEM
);
}
else
{
out
=
ff_get_video_buffer
(
outlink
,
outlink
->
w
,
outlink
->
h
);
if
(
!
out
)
return
AVERROR
(
ENOMEM
);
av_frame_copy_props
(
out
,
srcx
);
s
->
lut2
(
s
,
out
,
srcx
,
srcy
);
}
out
->
pts
=
av_rescale_q
(
s
->
fs
.
pts
,
s
->
fs
.
time_base
,
outlink
->
time_base
);
return
ff_filter_frame
(
outlink
,
out
);
}
static
int
config_output
(
AVFilterLink
*
outlink
)
{
AVFilterContext
*
ctx
=
outlink
->
src
;
LUT2Context
*
s
=
ctx
->
priv
;
AVFilterLink
*
srcx
=
ctx
->
inputs
[
0
];
AVFilterLink
*
srcy
=
ctx
->
inputs
[
1
];
FFFrameSyncIn
*
in
;
int
p
,
ret
;
s
->
depth
=
s
->
depthx
+
s
->
depthy
;
if
(
srcx
->
format
!=
srcy
->
format
)
{
av_log
(
ctx
,
AV_LOG_ERROR
,
"inputs must be of same pixel format
\n
"
);
return
AVERROR
(
EINVAL
);
}
if
(
srcx
->
w
!=
srcy
->
w
||
srcx
->
h
!=
srcy
->
h
||
srcx
->
sample_aspect_ratio
.
num
!=
srcy
->
sample_aspect_ratio
.
num
||
srcx
->
sample_aspect_ratio
.
den
!=
srcy
->
sample_aspect_ratio
.
den
)
{
av_log
(
ctx
,
AV_LOG_ERROR
,
"First input link %s parameters "
"(size %dx%d, SAR %d:%d) do not match the corresponding "
"second input link %s parameters (%dx%d, SAR %d:%d)
\n
"
,
ctx
->
input_pads
[
0
].
name
,
srcx
->
w
,
srcx
->
h
,
srcx
->
sample_aspect_ratio
.
num
,
srcx
->
sample_aspect_ratio
.
den
,
ctx
->
input_pads
[
1
].
name
,
srcy
->
w
,
srcy
->
h
,
srcy
->
sample_aspect_ratio
.
num
,
srcy
->
sample_aspect_ratio
.
den
);
return
AVERROR
(
EINVAL
);
}
outlink
->
w
=
srcx
->
w
;
outlink
->
h
=
srcx
->
h
;
outlink
->
time_base
=
srcx
->
time_base
;
outlink
->
sample_aspect_ratio
=
srcx
->
sample_aspect_ratio
;
outlink
->
frame_rate
=
srcx
->
frame_rate
;
if
((
ret
=
ff_framesync_init
(
&
s
->
fs
,
ctx
,
2
))
<
0
)
return
ret
;
in
=
s
->
fs
.
in
;
in
[
0
].
time_base
=
srcx
->
time_base
;
in
[
1
].
time_base
=
srcy
->
time_base
;
in
[
0
].
sync
=
1
;
in
[
0
].
before
=
EXT_STOP
;
in
[
0
].
after
=
EXT_INFINITY
;
in
[
1
].
sync
=
1
;
in
[
1
].
before
=
EXT_STOP
;
in
[
1
].
after
=
EXT_INFINITY
;
s
->
fs
.
opaque
=
s
;
s
->
fs
.
on_event
=
process_frame
;
s
->
lut2
=
s
->
depth
>
16
?
lut2_16bit
:
lut2_8bit
;
for
(
p
=
0
;
p
<
s
->
nb_planes
;
p
++
)
{
s
->
lut
[
p
]
=
av_malloc_array
(
1
<<
s
->
depth
,
sizeof
(
uint16_t
));
if
(
!
s
->
lut
[
p
])
return
AVERROR
(
ENOMEM
);
}
for
(
p
=
0
;
p
<
s
->
nb_planes
;
p
++
)
{
double
res
;
int
x
,
y
;
/* create the parsed expression */
av_expr_free
(
s
->
comp_expr
[
p
]);
s
->
comp_expr
[
p
]
=
NULL
;
ret
=
av_expr_parse
(
&
s
->
comp_expr
[
p
],
s
->
comp_expr_str
[
p
],
var_names
,
NULL
,
NULL
,
NULL
,
NULL
,
0
,
ctx
);
if
(
ret
<
0
)
{
av_log
(
ctx
,
AV_LOG_ERROR
,
"Error when parsing the expression '%s' for the component %d.
\n
"
,
s
->
comp_expr_str
[
p
],
p
);
return
AVERROR
(
EINVAL
);
}
/* compute the lut */
for
(
y
=
0
;
y
<
(
1
<<
s
->
depthx
);
y
++
)
{
s
->
var_values
[
VAR_Y
]
=
y
;
for
(
x
=
0
;
x
<
(
1
<<
s
->
depthx
);
x
++
)
{
s
->
var_values
[
VAR_X
]
=
x
;
res
=
av_expr_eval
(
s
->
comp_expr
[
p
],
s
->
var_values
,
s
);
if
(
isnan
(
res
))
{
av_log
(
ctx
,
AV_LOG_ERROR
,
"Error when evaluating the expression '%s' for the values %d and %d for the component %d.
\n
"
,
s
->
comp_expr_str
[
p
],
x
,
y
,
p
);
return
AVERROR
(
EINVAL
);
}
s
->
lut
[
p
][(
y
<<
s
->
depthx
)
+
x
]
=
res
;
}
}
}
return
ff_framesync_configure
(
&
s
->
fs
);
}
static
int
filter_frame
(
AVFilterLink
*
inlink
,
AVFrame
*
buf
)
{
LUT2Context
*
s
=
inlink
->
dst
->
priv
;
return
ff_framesync_filter_frame
(
&
s
->
fs
,
inlink
,
buf
);
}
static
int
request_frame
(
AVFilterLink
*
outlink
)
{
LUT2Context
*
s
=
outlink
->
src
->
priv
;
return
ff_framesync_request_frame
(
&
s
->
fs
,
outlink
);
}
static
const
AVFilterPad
inputs
[]
=
{
{
.
name
=
"srcx"
,
.
type
=
AVMEDIA_TYPE_VIDEO
,
.
filter_frame
=
filter_frame
,
.
config_props
=
config_inputx
,
},
{
.
name
=
"srcy"
,
.
type
=
AVMEDIA_TYPE_VIDEO
,
.
filter_frame
=
filter_frame
,
.
config_props
=
config_inputy
,
},
{
NULL
}
};
static
const
AVFilterPad
outputs
[]
=
{
{
.
name
=
"default"
,
.
type
=
AVMEDIA_TYPE_VIDEO
,
.
config_props
=
config_output
,
.
request_frame
=
request_frame
,
},
{
NULL
}
};
AVFILTER_DEFINE_CLASS
(
lut2
);
AVFilter
ff_vf_lut2
=
{
.
name
=
"lut2"
,
.
description
=
NULL_IF_CONFIG_SMALL
(
"Compute and apply a lookup table from two video inputs."
),
.
priv_size
=
sizeof
(
LUT2Context
),
.
priv_class
=
&
lut2_class
,
.
uninit
=
uninit
,
.
query_formats
=
query_formats
,
.
inputs
=
inputs
,
.
outputs
=
outputs
,
.
flags
=
AVFILTER_FLAG_SUPPORT_TIMELINE_INTERNAL
,
};
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment