Go to the documentation of this file.
36 #define OFFSET(x) offsetof(TInterlaceContext, x)
37 #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
77 #define FULL_SCALE_YUVJ_FORMATS \
78 AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P
113 ptrdiff_t mref, ptrdiff_t pref,
int clip_max)
115 const uint8_t *srcp_above = srcp + mref;
116 const uint8_t *srcp_below = srcp + pref;
122 dstp[
i] = (1 + srcp[
i] + srcp[
i] + srcp_above[
i] + srcp_below[
i]) >> 2;
127 ptrdiff_t mref, ptrdiff_t pref,
int clip_max)
129 uint16_t *dstp = (uint16_t *)dst8;
130 const uint16_t *srcp = (
const uint16_t *)src8;
131 const uint16_t *srcp_above = srcp + mref / 2;
132 const uint16_t *srcp_below = srcp + pref / 2;
145 ptrdiff_t mref, ptrdiff_t pref,
int clip_max)
147 const uint8_t *srcp_above = srcp + mref;
148 const uint8_t *srcp_below = srcp + pref;
149 const uint8_t *srcp_above2 = srcp + mref * 2;
150 const uint8_t *srcp_below2 = srcp + pref * 2;
151 int i, src_x, src_ab;
156 src_x = srcp[
i] << 1;
157 src_ab = srcp_above[
i] + srcp_below[
i];
159 - srcp_above2[
i] - srcp_below2[
i]) >> 3);
163 if (src_ab > src_x) {
164 if (dstp[
i] < srcp[
i])
166 }
else if (dstp[
i] > srcp[
i])
172 ptrdiff_t mref, ptrdiff_t pref,
int clip_max)
174 uint16_t *dstp = (uint16_t *)dst8;
175 const uint16_t *srcp = (
const uint16_t *)src8;
176 const uint16_t *srcp_above = srcp + mref / 2;
177 const uint16_t *srcp_below = srcp + pref / 2;
178 const uint16_t *srcp_above2 = srcp + mref;
179 const uint16_t *srcp_below2 = srcp + pref;
180 int i, dst_le, src_le, src_x, src_ab;
188 dst_le =
av_clip((4 + ((src_le + src_x + src_ab) << 1)
194 if (src_ab > src_x) {
199 }
else if (dst_le > src_le) {
223 tinterlace->
vsub =
desc->log2_chroma_h;
232 uint8_t black[4] = { 0, 0, 0, 16 };
239 outlink->
w, outlink->
h, outlink->
format, 16);
299 #define FIELD_UPPER 0
300 #define FIELD_LOWER 1
301 #define FIELD_UPPER_AND_LOWER 2
314 uint8_t *dst[4],
int dst_linesize[4],
322 int plane, vsub =
desc->log2_chroma_h;
326 for (plane = 0; plane <
desc->nb_components; plane++) {
327 int lines = plane == 1 || plane == 2 ?
AV_CEIL_RSHIFT(src_h, vsub) : src_h;
331 int srcp_linesize = src_linesize[plane] * k;
332 int dstp_linesize = dst_linesize[plane] * (
interleave ? 2 : 1);
333 int clip_max = (1 << tinterlace->
csp->
comp[plane].
depth) - 1;
337 srcp += src_linesize[plane];
339 dstp += dst_linesize[plane];
345 for (
h = lines;
h > 0;
h--) {
346 ptrdiff_t pref = src_linesize[plane];
347 ptrdiff_t mref = -pref;
348 if (
h >= (lines - x)) mref = 0;
349 else if (
h <= (1 + x)) pref = 0;
351 tinterlace->
lowpass_line(dstp, cols, srcp, mref, pref, clip_max);
352 dstp += dstp_linesize;
353 srcp += srcp_linesize;
372 tinterlace->
cur = tinterlace->
next;
373 tinterlace->
next = picref;
375 cur = tinterlace->
cur;
376 next = tinterlace->
next;
378 if (!tinterlace->
cur)
381 switch (tinterlace->
mode) {
390 out->height = outlink->
h;
391 out->interlaced_frame = 1;
392 out->top_field_first = 1;
423 out->height = outlink->
h;
445 "video is already interlaced, adjusting framerate only\n");
458 out->interlaced_frame = 1;
459 out->top_field_first = tff;
469 (
const uint8_t **)next->data, next->linesize,
480 out->interlaced_frame = 1;
489 tff = next->top_field_first;
494 out->interlaced_frame = 1;
495 out->top_field_first = !tff;
498 out->pts = cur->
pts + next->pts;
509 (
const uint8_t **)next->data, next->linesize,
559 .
name =
"tinterlace",
566 .priv_class = &tinterlace_class,
579 .priv_class = &interlace_class,
AVFrame * ff_get_video_buffer(AVFilterLink *link, int w, int h)
Request a picture buffer with a specific set of permissions.
static int filter_frame(AVFilterLink *inlink, AVFrame *picref)
#define AV_LOG_WARNING
Something somehow does not look correct.
AVPixelFormat
Pixel format.
static av_cold int init(AVCodecContext *avctx)
static int init_interlace(AVFilterContext *ctx)
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
#define TINTERLACE_FLAG_BYPASS_IL
static const AVRational standard_tbs[]
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
The exact code depends on how similar the blocks are and how related they are to the and needs to apply these operations to the correct inlink or outlink if there are several Macros are available to factor that when no extra processing is inlink
void(* lowpass_line)(uint8_t *dstp, ptrdiff_t width, const uint8_t *srcp, ptrdiff_t mref, ptrdiff_t pref, int clip_max)
static void lowpass_line_complex_c(uint8_t *dstp, ptrdiff_t width, const uint8_t *srcp, ptrdiff_t mref, ptrdiff_t pref, int clip_max)
#define FIELD_UPPER_AND_LOWER
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
This structure describes decoded (raw) audio or video data.
int64_t pts
Presentation timestamp in time_base units (time when frame should be shown to user).
int depth
Number of bits in the component.
#define AV_LOG_VERBOSE
Detailed information.
@ AV_PIX_FMT_YUV440P
planar YUV 4:4:0 (1 Cr & Cb sample per 1x2 Y samples)
uint8_t * black_data[4]
buffer used to fill padded lines
const char * name
Filter name.
A link between two filters.
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
void av_image_copy_plane(uint8_t *dst, int dst_linesize, const uint8_t *src, int src_linesize, int bytewidth, int height)
Copy image plane from src to dst.
int vsub
chroma vertical subsampling
AVRational preout_time_base
@ AV_PIX_FMT_YUV420P12LE
planar YUV 4:2:0,18bpp, (1 Cr & Cb sample per 2x2 Y samples), little-endian
static int query_formats(AVFilterContext *ctx)
static int config_out_props(AVFilterLink *outlink)
union FFDrawColor::@204 comp[MAX_PLANES]
static const AVFilterPad tinterlace_inputs[]
A filter pad used for either input or output.
@ AV_PIX_FMT_YUV420P10LE
planar YUV 4:2:0, 15bpp, (1 Cr & Cb sample per 2x2 Y samples), little-endian
static av_cold void uninit(AVFilterContext *ctx)
@ AV_PIX_FMT_YUV444P12LE
planar YUV 4:4:4,36bpp, (1 Cr & Cb sample per 1x1 Y samples), little-endian
#define FF_ARRAY_ELEMS(a)
int flags
flags affecting interlacing algorithm
const AVPixFmtDescriptor * csp
@ AV_PIX_FMT_YUVA420P
planar YUV 4:2:0, 20bpp, (1 Cr & Cb sample per 2x2 Y & A samples)
#define AV_CEIL_RSHIFT(a, b)
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample format(the sample packing is implied by the sample format) and sample rate. The lists are not just lists
AVRational sample_aspect_ratio
agreed upon sample aspect ratio
#define av_assert0(cond)
assert() equivalent, that is always enabled.
static const AVFilterPad outputs[]
AVRational frame_rate
Frame rate of the stream on the link, or 1/0 if unknown or variable; if left to 0/0,...
static enum AVPixelFormat pix_fmts[]
AVFrame * av_frame_clone(const AVFrame *src)
Create a new frame that references the same data as src.
int64_t av_rescale_q(int64_t a, AVRational bq, AVRational cq)
Rescale a 64-bit integer by 2 rational numbers.
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this field
int ff_draw_init(FFDrawContext *draw, enum AVPixelFormat format, unsigned flags)
Init a draw context.
void ff_tinterlace_init_x86(TInterlaceContext *interlace)
@ AV_PIX_FMT_YUV420P
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
@ AV_PIX_FMT_YUV444P10LE
planar YUV 4:4:4, 30bpp, (1 Cr & Cb sample per 1x1 Y samples), little-endian
@ AV_PIX_FMT_YUVA422P10LE
planar YUV 4:2:2 30bpp, (1 Cr & Cb sample per 2x1 Y & A samples, little-endian)
#define TINTERLACE_FLAG_EXACT_TB
int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
Copy only "metadata" fields from src to dst.
static const AVOption tinterlace_options[]
Rational number (pair of numerator and denominator).
AVFilterLink ** inputs
array of pointers to input links
int mode
TInterlaceMode, interlace mode selected.
@ AV_PIX_FMT_YUV440P10LE
planar YUV 4:4:0,20bpp, (1 Cr & Cb sample per 1x2 Y samples), little-endian
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several inputs
@ AV_PIX_FMT_GRAY8
Y , 8bpp.
int av_image_alloc(uint8_t *pointers[4], int linesizes[4], int w, int h, enum AVPixelFormat pix_fmt, int align)
Allocate an image with size w and h and pixel format pix_fmt, and fill pointers and linesizes accordi...
static void interleave(uint8_t *dst, uint8_t *src, int w, int h, int dst_linesize, int src_linesize, enum FilterMode mode, int swap)
@ AV_PIX_FMT_YUV440P12LE
planar YUV 4:4:0,24bpp, (1 Cr & Cb sample per 1x2 Y samples), little-endian
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
@ AV_PIX_FMT_YUV422P10LE
planar YUV 4:2:2, 20bpp, (1 Cr & Cb sample per 2x1 Y samples), little-endian
int format
agreed upon media format
#define TINTERLACE_FLAG_CVLPF
static AVRational av_make_q(int num, int den)
Create an AVRational.
#define AV_NOPTS_VALUE
Undefined timestamp value.
void ff_fill_rectangle(FFDrawContext *draw, FFDrawColor *color, uint8_t *dst[], int dst_linesize[], int dst_x, int dst_y, int w, int h)
Fill a rectangle with an uniform color.
AVFilterContext * src
source filter
AVFilter ff_vf_tinterlace
@ AV_PIX_FMT_YUVA444P
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
#define FULL_SCALE_YUVJ_FORMATS
static void copy_picture_field(TInterlaceContext *tinterlace, uint8_t *dst[4], int dst_linesize[4], const uint8_t *src[4], int src_linesize[4], enum AVPixelFormat format, int w, int src_h, int src_field, int interleave, int dst_field, int flags)
Copy picture field from src to dst.
@ AV_PIX_FMT_YUVA420P10LE
planar YUV 4:2:0 25bpp, (1 Cr & Cb sample per 2x2 Y & A samples, little-endian)
int interlaced_frame
The content of the picture is interlaced.
int w
agreed upon image width
const char * name
Pad name.
void ff_draw_color(FFDrawContext *draw, FFDrawColor *color, const uint8_t rgba[4])
Prepare a color.
static int av_cmp_q(AVRational a, AVRational b)
Compare two rationals.
int lowpass
legacy interlace filter lowpass mode
AVRational sample_aspect_ratio
Sample aspect ratio for the video frame, 0/1 if unknown/unspecified.
int h
agreed upon image height
AVComponentDescriptor comp[4]
Parameters that describe how pixels are packed.
AVRational time_base
Define the time base used by the PTS of the frames/samples which will pass through this link.
AVRational av_mul_q(AVRational b, AVRational c)
Multiply two rationals.
@ AV_PIX_FMT_YUV444P
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
static void lowpass_line_c(uint8_t *dstp, ptrdiff_t width, const uint8_t *srcp, ptrdiff_t mref, ptrdiff_t pref, int clip_max)
@ AV_PIX_FMT_YUV422P
planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples)
static enum AVPixelFormat full_scale_yuvj_pix_fmts[]
int64_t frame_count_in
Number of past frames sent through the link.
@ AV_PIX_FMT_YUVA444P10LE
planar YUV 4:4:4 40bpp, (1 Cr & Cb sample per 1x1 Y & A samples, little-endian)
static void lowpass_line_complex_c_16(uint8_t *dst8, ptrdiff_t width, const uint8_t *src8, ptrdiff_t mref, ptrdiff_t pref, int clip_max)
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
@ AV_PIX_FMT_YUV411P
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples)
#define TINTERLACE_FLAG_VLPF
static const AVOption interlace_options[]
#define flags(name, subs,...)
static void lowpass_line_c_16(uint8_t *dst8, ptrdiff_t width, const uint8_t *src8, ptrdiff_t mref, ptrdiff_t pref, int clip_max)
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
@ AV_PIX_FMT_YUV410P
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
@ AV_PIX_FMT_YUV422P12LE
planar YUV 4:2:2,24bpp, (1 Cr & Cb sample per 2x1 Y samples), little-endian
AVFILTER_DEFINE_CLASS(tinterlace)
@ AV_PIX_FMT_YUVA422P
planar YUV 4:2:2 24bpp, (1 Cr & Cb sample per 2x1 Y & A samples)
static const AVFilterPad tinterlace_outputs[]