FFmpeg  4.4.4
vf_amplify.c
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2018 Paul B Mahol
3  *
4  * This file is part of FFmpeg.
5  *
6  * FFmpeg is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * FFmpeg is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with FFmpeg; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19  */
20 
21 #include "libavutil/imgutils.h"
22 #include "libavutil/intreadwrite.h"
23 #include "libavutil/opt.h"
24 #include "libavutil/pixdesc.h"
25 
26 #include "avfilter.h"
27 #include "formats.h"
28 #include "internal.h"
29 #include "video.h"
30 
31 typedef struct AmplifyContext {
32  const AVClass *class;
34  int radius;
35  float factor;
36  float threshold;
37  float tolerance;
38  int planes;
39 
40  int llimit;
41  int hlimit;
42  int nb_inputs;
43  int nb_frames;
44 
45  int depth;
46  int nb_planes;
47  int linesize[4];
48  int height[4];
49 
52 
54 {
55  static const enum AVPixelFormat pixel_fmts[] = {
80  };
82  if (!formats)
83  return AVERROR(ENOMEM);
85 }
86 
88 {
89  AmplifyContext *s = ctx->priv;
90 
91  s->nb_inputs = s->radius * 2 + 1;
92 
93  s->frames = av_calloc(s->nb_inputs, sizeof(*s->frames));
94  if (!s->frames)
95  return AVERROR(ENOMEM);
96 
97  return 0;
98 }
99 
100 typedef struct ThreadData {
102 } ThreadData;
103 
104 static int amplify_frame(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
105 {
106  AmplifyContext *s = ctx->priv;
107  ThreadData *td = arg;
108  AVFrame **in = td->in;
109  AVFrame *out = td->out;
110  const int radius = s->radius;
111  const int nb_inputs = s->nb_inputs;
112  const float threshold = s->threshold;
113  const float tolerance = s->tolerance;
114  const float factor = s->factor;
115  const int llimit = s->llimit;
116  const int hlimit = s->hlimit;
117  const int depth = s->depth;
118  int i, p, x, y;
119 
120  if (s->depth <= 8) {
121  for (p = 0; p < s->nb_planes; p++) {
122  const int slice_start = (s->height[p] * jobnr) / nb_jobs;
123  const int slice_end = (s->height[p] * (jobnr+1)) / nb_jobs;
124  uint8_t *dst = out->data[p] + slice_start * out->linesize[p];
125 
126  if (!((1 << p) & s->planes)) {
127  av_image_copy_plane(dst, out->linesize[p],
128  in[radius]->data[p] + slice_start * in[radius]->linesize[p],
129  in[radius]->linesize[p],
130  s->linesize[p], slice_end - slice_start);
131  continue;
132  }
133 
134  for (y = slice_start; y < slice_end; y++) {
135  for (x = 0; x < s->linesize[p]; x++) {
136  int src = in[radius]->data[p][y * in[radius]->linesize[p] + x];
137  float diff, avg;
138  int sum = 0;
139 
140  for (i = 0; i < nb_inputs; i++) {
141  sum += in[i]->data[p][y * in[i]->linesize[p] + x];
142  }
143 
144  avg = sum / (float)nb_inputs;
145  diff = src - avg;
146  if (fabsf(diff) < threshold && fabsf(diff) > tolerance) {
147  int amp;
148  if (diff < 0) {
149  amp = -FFMIN(FFABS(diff * factor), llimit);
150  } else {
151  amp = FFMIN(FFABS(diff * factor), hlimit);
152  }
153  dst[x] = av_clip_uint8(src + amp);
154  } else {
155  dst[x] = src;
156  }
157  }
158 
159  dst += out->linesize[p];
160  }
161  }
162  } else {
163  for (p = 0; p < s->nb_planes; p++) {
164  const int slice_start = (s->height[p] * jobnr) / nb_jobs;
165  const int slice_end = (s->height[p] * (jobnr+1)) / nb_jobs;
166  uint16_t *dst = (uint16_t *)(out->data[p] + slice_start * out->linesize[p]);
167 
168  if (!((1 << p) & s->planes)) {
169  av_image_copy_plane((uint8_t *)dst, out->linesize[p],
170  in[radius]->data[p] + slice_start * in[radius]->linesize[p],
171  in[radius]->linesize[p],
172  s->linesize[p], slice_end - slice_start);
173  continue;
174  }
175 
176  for (y = slice_start; y < slice_end; y++) {
177  for (x = 0; x < s->linesize[p] / 2; x++) {
178  int src = AV_RN16(in[radius]->data[p] + y * in[radius]->linesize[p] + x * 2);
179  float diff, avg;
180  int sum = 0;
181 
182  for (i = 0; i < nb_inputs; i++) {
183  sum += AV_RN16(in[i]->data[p] + y * in[i]->linesize[p] + x * 2);
184  }
185 
186  avg = sum / (float)nb_inputs;
187  diff = src - avg;
188 
189  if (fabsf(diff) < threshold && fabsf(diff) > tolerance) {
190  int amp;
191  if (diff < 0) {
192  amp = -FFMIN(FFABS(diff * factor), llimit);
193  } else {
194  amp = FFMIN(FFABS(diff * factor), hlimit);
195  }
196  dst[x] = av_clip_uintp2_c(src + amp, depth);
197  } else {
198  dst[x] = src;
199  }
200  }
201 
202  dst += out->linesize[p] / 2;
203  }
204  }
205  }
206 
207  return 0;
208 }
209 
210 static int config_output(AVFilterLink *outlink)
211 {
212  AVFilterContext *ctx = outlink->src;
213  AmplifyContext *s = ctx->priv;
214  AVFilterLink *inlink = ctx->inputs[0];
215  int ret;
216 
217  s->desc = av_pix_fmt_desc_get(outlink->format);
218  if (!s->desc)
219  return AVERROR_BUG;
220  s->nb_planes = av_pix_fmt_count_planes(outlink->format);
221  s->depth = s->desc->comp[0].depth;
222 
223  if ((ret = av_image_fill_linesizes(s->linesize, inlink->format, inlink->w)) < 0)
224  return ret;
225 
226  s->height[1] = s->height[2] = AV_CEIL_RSHIFT(inlink->h, s->desc->log2_chroma_h);
227  s->height[0] = s->height[3] = inlink->h;
228 
229  return 0;
230 }
231 
233 {
234  AmplifyContext *s = ctx->priv;
235  int i;
236 
237  if (s->frames) {
238  for (i = 0; i < s->nb_frames; i++)
239  av_frame_free(&s->frames[i]);
240  }
241  av_freep(&s->frames);
242 }
243 
244 static int filter_frame(AVFilterLink *inlink, AVFrame *in)
245 {
246  AVFilterContext *ctx = inlink->dst;
247  AVFilterLink *outlink = ctx->outputs[0];
248  AmplifyContext *s = ctx->priv;
249  ThreadData td;
250  AVFrame *out;
251 
252  if (s->nb_frames < s->nb_inputs) {
253  s->frames[s->nb_frames] = in;
254  s->nb_frames++;
255  return 0;
256  } else {
257  av_frame_free(&s->frames[0]);
258  memmove(&s->frames[0], &s->frames[1], sizeof(*s->frames) * (s->nb_inputs - 1));
259  s->frames[s->nb_inputs - 1] = in;
260  }
261 
262  if (!ctx->is_disabled) {
263  out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
264  if (!out)
265  return AVERROR(ENOMEM);
266  av_frame_copy_props(out, s->frames[0]);
267 
268  td.out = out;
269  td.in = s->frames;
270  ctx->internal->execute(ctx, amplify_frame, &td, NULL, FFMIN(s->height[1], ff_filter_get_nb_threads(ctx)));
271  } else {
272  out = av_frame_clone(s->frames[s->radius]);
273  if (!out)
274  return AVERROR(ENOMEM);
275  out->pts = s->frames[0]->pts;
276  }
277 
278  return ff_filter_frame(outlink, out);
279 }
280 
281 #define OFFSET(x) offsetof(AmplifyContext, x)
282 #define FLAGS AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_FILTERING_PARAM
283 #define VFT AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_FILTERING_PARAM | AV_OPT_FLAG_RUNTIME_PARAM
284 
285 static const AVOption amplify_options[] = {
286  { "radius", "set radius", OFFSET(radius), AV_OPT_TYPE_INT, {.i64=2}, 1, 63, .flags = FLAGS },
287  { "factor", "set factor", OFFSET(factor), AV_OPT_TYPE_FLOAT, {.dbl=2}, 0, UINT16_MAX, .flags = VFT },
288  { "threshold", "set threshold", OFFSET(threshold), AV_OPT_TYPE_FLOAT, {.dbl=10}, 0, UINT16_MAX, .flags = VFT },
289  { "tolerance", "set tolerance", OFFSET(tolerance), AV_OPT_TYPE_FLOAT, {.dbl=0}, 0, UINT16_MAX, .flags = VFT },
290  { "low", "set low limit for amplification", OFFSET(llimit), AV_OPT_TYPE_INT, {.i64=UINT16_MAX}, 0, UINT16_MAX, .flags = VFT },
291  { "high", "set high limit for amplification", OFFSET(hlimit), AV_OPT_TYPE_INT, {.i64=UINT16_MAX}, 0, UINT16_MAX, .flags = VFT },
292  { "planes", "set what planes to filter", OFFSET(planes), AV_OPT_TYPE_FLAGS, {.i64=7}, 0, 15, VFT },
293  { NULL },
294 };
295 
296 static const AVFilterPad inputs[] = {
297  {
298  .name = "default",
299  .type = AVMEDIA_TYPE_VIDEO,
300  .filter_frame = filter_frame,
301  },
302  { NULL }
303 };
304 
305 static const AVFilterPad outputs[] = {
306  {
307  .name = "default",
308  .type = AVMEDIA_TYPE_VIDEO,
309  .config_props = config_output,
310  },
311  { NULL }
312 };
313 
315 
317  .name = "amplify",
318  .description = NULL_IF_CONFIG_SMALL("Amplify changes between successive video frames."),
319  .priv_size = sizeof(AmplifyContext),
320  .priv_class = &amplify_class,
322  .outputs = outputs,
323  .inputs = inputs,
324  .init = init,
325  .uninit = uninit,
328 };
static int process_command(AVFilterContext *ctx, const char *cmd, const char *args, char *res, int res_len, int flags)
Definition: af_acrusher.c:336
#define av_cold
Definition: attributes.h:88
uint8_t pi<< 24) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi - 0x80) *(1.0f/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi - 0x80) *(1.0/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S16, int16_t,(*(const int16_t *) pi >> 8)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S16, int16_t, *(const int16_t *) pi *(1.0f/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S16, int16_t, *(const int16_t *) pi *(1.0/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S32, int32_t,(*(const int32_t *) pi >> 24)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S32, int32_t, *(const int32_t *) pi *(1.0f/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S32, int32_t, *(const int32_t *) pi *(1.0/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_FLT, float, av_clip_uint8(lrintf(*(const float *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_FLT, float, av_clip_int16(lrintf(*(const float *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_FLT, float, av_clipl_int32(llrintf(*(const float *) pi *(1U<< 31)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_DBL, double, av_clip_uint8(lrint(*(const double *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_DBL, double, av_clip_int16(lrint(*(const double *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_DBL, double, av_clipl_int32(llrint(*(const double *) pi *(1U<< 31)))) #define SET_CONV_FUNC_GROUP(ofmt, ifmt) static void set_generic_function(AudioConvert *ac) { } void ff_audio_convert_free(AudioConvert **ac) { if(! *ac) return;ff_dither_free(&(*ac) ->dc);av_freep(ac);} AudioConvert *ff_audio_convert_alloc(AVAudioResampleContext *avr, enum AVSampleFormat out_fmt, enum AVSampleFormat in_fmt, int channels, int sample_rate, int apply_map) { AudioConvert *ac;int in_planar, out_planar;ac=av_mallocz(sizeof(*ac));if(!ac) return NULL;ac->avr=avr;ac->out_fmt=out_fmt;ac->in_fmt=in_fmt;ac->channels=channels;ac->apply_map=apply_map;if(avr->dither_method !=AV_RESAMPLE_DITHER_NONE &&av_get_packed_sample_fmt(out_fmt)==AV_SAMPLE_FMT_S16 &&av_get_bytes_per_sample(in_fmt) > 2) { ac->dc=ff_dither_alloc(avr, out_fmt, in_fmt, channels, sample_rate, apply_map);if(!ac->dc) { av_free(ac);return NULL;} return ac;} in_planar=ff_sample_fmt_is_planar(in_fmt, channels);out_planar=ff_sample_fmt_is_planar(out_fmt, channels);if(in_planar==out_planar) { ac->func_type=CONV_FUNC_TYPE_FLAT;ac->planes=in_planar ? ac->channels :1;} else if(in_planar) ac->func_type=CONV_FUNC_TYPE_INTERLEAVE;else ac->func_type=CONV_FUNC_TYPE_DEINTERLEAVE;set_generic_function(ac);if(ARCH_AARCH64) ff_audio_convert_init_aarch64(ac);if(ARCH_ARM) ff_audio_convert_init_arm(ac);if(ARCH_X86) ff_audio_convert_init_x86(ac);return ac;} int ff_audio_convert(AudioConvert *ac, AudioData *out, AudioData *in) { int use_generic=1;int len=in->nb_samples;int p;if(ac->dc) { av_log(ac->avr, AV_LOG_TRACE, "%d samples - audio_convert: %s to %s (dithered)\n", len, av_get_sample_fmt_name(ac->in_fmt), av_get_sample_fmt_name(ac->out_fmt));return ff_convert_dither(ac-> in
uint8_t
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
Definition: avfilter.c:1096
int ff_filter_process_command(AVFilterContext *ctx, const char *cmd, const char *arg, char *res, int res_len, int flags)
Generic processing of user supplied commands that are set in the same way as the filter options.
Definition: avfilter.c:882
int ff_filter_get_nb_threads(AVFilterContext *ctx)
Get number of threads for current filter instance.
Definition: avfilter.c:802
Main libavfilter public API header.
#define flags(name, subs,...)
Definition: cbs_av1.c:561
#define s(width, name)
Definition: cbs_vp9.c:257
#define avg(a, b, c, d)
#define FFMIN(a, b)
Definition: common.h:105
#define AV_CEIL_RSHIFT(a, b)
Definition: common.h:58
static av_always_inline av_const unsigned av_clip_uintp2_c(int a, int p)
Clip a signed integer to an unsigned power of two range.
Definition: common.h:302
#define av_clip_uint8
Definition: common.h:128
#define FFABS(a)
Absolute value, Note, INT_MIN / INT64_MIN result in undefined behavior as they are not representable ...
Definition: common.h:72
#define NULL
Definition: coverity.c:32
static __device__ float fabsf(float a)
Definition: cuda_runtime.h:181
int ff_set_common_formats(AVFilterContext *ctx, AVFilterFormats *formats)
A helper for query_formats() which sets all links to the same list of formats.
Definition: formats.c:587
AVFilterFormats * ff_make_format_list(const int *fmts)
Create a list of supported formats.
Definition: formats.c:286
@ AV_OPT_TYPE_FLAGS
Definition: opt.h:224
@ AV_OPT_TYPE_INT
Definition: opt.h:225
@ AV_OPT_TYPE_FLOAT
Definition: opt.h:228
#define AVFILTER_FLAG_SLICE_THREADS
The filter supports multithreading by splitting frames into multiple parts and processing them concur...
Definition: avfilter.h:117
#define AVFILTER_FLAG_SUPPORT_TIMELINE_INTERNAL
Same as AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC, except that the filter will have its filter_frame() c...
Definition: avfilter.h:134
#define AVERROR_BUG
Internal bug, also see AVERROR_BUG2.
Definition: error.h:50
#define AVERROR(e)
Definition: error.h:43
AVFrame * av_frame_clone(const AVFrame *src)
Create a new frame that references the same data as src.
Definition: frame.c:540
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
Definition: frame.c:203
int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
Copy only "metadata" fields from src to dst.
Definition: frame.c:658
void * av_calloc(size_t nmemb, size_t size)
Non-inlined equivalent of av_mallocz_array().
Definition: mem.c:245
@ AVMEDIA_TYPE_VIDEO
Definition: avutil.h:201
void av_image_copy_plane(uint8_t *dst, int dst_linesize, const uint8_t *src, int src_linesize, int bytewidth, int height)
Copy image plane from src to dst.
Definition: imgutils.c:373
int av_image_fill_linesizes(int linesizes[4], enum AVPixelFormat pix_fmt, int width)
Fill plane linesizes for an image with pixel format pix_fmt and width width.
Definition: imgutils.c:89
misc image utilities
int i
Definition: input.c:407
#define AV_RN16(p)
Definition: intreadwrite.h:360
const char * arg
Definition: jacosubdec.c:66
common internal API header
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
Definition: internal.h:117
static const struct @322 planes[]
static int slice_end(AVCodecContext *avctx, AVFrame *pict)
Handle slice ends.
Definition: mpeg12dec.c:2033
const char data[16]
Definition: mxf.c:142
AVOptions.
int av_pix_fmt_count_planes(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2613
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2573
#define AV_PIX_FMT_GBRAP12
Definition: pixfmt.h:420
#define AV_PIX_FMT_YUV420P16
Definition: pixfmt.h:410
#define AV_PIX_FMT_YUV444P12
Definition: pixfmt.h:406
#define AV_PIX_FMT_YUV444P9
Definition: pixfmt.h:398
#define AV_PIX_FMT_YUV420P10
Definition: pixfmt.h:399
#define AV_PIX_FMT_YUV440P12
Definition: pixfmt.h:405
#define AV_PIX_FMT_GRAY9
Definition: pixfmt.h:379
#define AV_PIX_FMT_GBRAP16
Definition: pixfmt.h:421
#define AV_PIX_FMT_GBRP9
Definition: pixfmt.h:414
#define AV_PIX_FMT_YUV422P9
Definition: pixfmt.h:397
#define AV_PIX_FMT_YUVA444P10
Definition: pixfmt.h:438
#define AV_PIX_FMT_YUVA420P16
Definition: pixfmt.h:441
#define AV_PIX_FMT_YUV420P12
Definition: pixfmt.h:403
#define AV_PIX_FMT_YUVA420P10
Definition: pixfmt.h:436
#define AV_PIX_FMT_YUVA422P9
Definition: pixfmt.h:434
#define AV_PIX_FMT_YUV422P12
Definition: pixfmt.h:404
#define AV_PIX_FMT_GBRP10
Definition: pixfmt.h:415
#define AV_PIX_FMT_YUV422P10
Definition: pixfmt.h:400
#define AV_PIX_FMT_GRAY12
Definition: pixfmt.h:381
#define AV_PIX_FMT_GBRP12
Definition: pixfmt.h:416
#define AV_PIX_FMT_YUV420P9
Definition: pixfmt.h:396
#define AV_PIX_FMT_YUVA420P9
Definition: pixfmt.h:433
#define AV_PIX_FMT_YUVA422P10
Definition: pixfmt.h:437
#define AV_PIX_FMT_YUV420P14
Definition: pixfmt.h:407
AVPixelFormat
Pixel format.
Definition: pixfmt.h:64
@ AV_PIX_FMT_NONE
Definition: pixfmt.h:65
@ AV_PIX_FMT_YUV420P
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
Definition: pixfmt.h:66
@ AV_PIX_FMT_YUV440P
planar YUV 4:4:0 (1 Cr & Cb sample per 1x2 Y samples)
Definition: pixfmt.h:99
@ AV_PIX_FMT_YUV422P
planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples)
Definition: pixfmt.h:70
@ AV_PIX_FMT_GRAY8
Y , 8bpp.
Definition: pixfmt.h:74
@ AV_PIX_FMT_YUVA420P
planar YUV 4:2:0, 20bpp, (1 Cr & Cb sample per 2x2 Y & A samples)
Definition: pixfmt.h:101
@ AV_PIX_FMT_YUVJ440P
planar YUV 4:4:0 full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV440P and setting color_range
Definition: pixfmt.h:100
@ AV_PIX_FMT_YUV410P
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
Definition: pixfmt.h:72
@ AV_PIX_FMT_YUV411P
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples)
Definition: pixfmt.h:73
@ AV_PIX_FMT_YUV444P
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
Definition: pixfmt.h:71
@ AV_PIX_FMT_YUVA444P
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
Definition: pixfmt.h:177
@ AV_PIX_FMT_YUVJ411P
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples) full scale (JPEG), deprecated in favor ...
Definition: pixfmt.h:258
@ AV_PIX_FMT_GBRAP
planar GBRA 4:4:4:4 32bpp
Definition: pixfmt.h:215
@ AV_PIX_FMT_YUVJ422P
planar YUV 4:2:2, 16bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV422P and setting col...
Definition: pixfmt.h:79
@ AV_PIX_FMT_YUVA422P
planar YUV 4:2:2 24bpp, (1 Cr & Cb sample per 2x1 Y & A samples)
Definition: pixfmt.h:176
@ AV_PIX_FMT_GBRP
planar GBR 4:4:4 24bpp
Definition: pixfmt.h:168
@ AV_PIX_FMT_YUVJ444P
planar YUV 4:4:4, 24bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV444P and setting col...
Definition: pixfmt.h:80
@ AV_PIX_FMT_YUVJ420P
planar YUV 4:2:0, 12bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV420P and setting col...
Definition: pixfmt.h:78
#define AV_PIX_FMT_YUVA422P12
Definition: pixfmt.h:439
#define AV_PIX_FMT_YUV422P14
Definition: pixfmt.h:408
#define AV_PIX_FMT_GRAY10
Definition: pixfmt.h:380
#define AV_PIX_FMT_GRAY14
Definition: pixfmt.h:382
#define AV_PIX_FMT_YUV422P16
Definition: pixfmt.h:411
#define AV_PIX_FMT_YUV440P10
Definition: pixfmt.h:401
#define AV_PIX_FMT_GRAY16
Definition: pixfmt.h:383
#define AV_PIX_FMT_GBRAP10
Definition: pixfmt.h:419
#define AV_PIX_FMT_YUVA444P16
Definition: pixfmt.h:443
#define AV_PIX_FMT_YUVA422P16
Definition: pixfmt.h:442
#define AV_PIX_FMT_GBRP16
Definition: pixfmt.h:418
#define AV_PIX_FMT_YUV444P14
Definition: pixfmt.h:409
#define AV_PIX_FMT_YUVA444P9
Definition: pixfmt.h:435
#define AV_PIX_FMT_GBRP14
Definition: pixfmt.h:417
#define AV_PIX_FMT_YUVA444P12
Definition: pixfmt.h:440
#define AV_PIX_FMT_YUV444P16
Definition: pixfmt.h:412
#define AV_PIX_FMT_YUV444P10
Definition: pixfmt.h:402
#define td
Definition: regdef.h:70
formats
Definition: signature.h:48
Describe the class of an AVClass context structure.
Definition: log.h:67
An instance of a filter.
Definition: avfilter.h:341
A list of supported formats for one end of a filter link.
Definition: formats.h:65
A filter pad used for either input or output.
Definition: internal.h:54
const char * name
Pad name.
Definition: internal.h:60
Filter definition.
Definition: avfilter.h:145
const char * name
Filter name.
Definition: avfilter.h:149
AVFormatInternal * internal
An opaque field for libavformat internal usage.
Definition: avformat.h:1699
This structure describes decoded (raw) audio or video data.
Definition: frame.h:318
AVOption.
Definition: opt.h:248
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
Definition: pixdesc.h:81
const AVPixFmtDescriptor * desc
Definition: vf_amplify.c:33
int height[4]
Definition: vf_amplify.c:48
float tolerance
Definition: vf_amplify.c:37
AVFrame ** frames
Definition: vf_amplify.c:50
float threshold
Definition: vf_amplify.c:36
int linesize[4]
Definition: vf_amplify.c:47
Used for passing data between threads.
Definition: dsddec.c:67
AVFrame ** in
Definition: vf_amplify.c:101
AVFrame * out
Definition: af_adeclick.c:502
#define av_freep(p)
#define src
Definition: vp8dsp.c:255
FILE * out
Definition: movenc.c:54
AVFormatContext * ctx
Definition: movenc.c:48
#define VFT
Definition: vf_amplify.c:283
AVFILTER_DEFINE_CLASS(amplify)
AVFilter ff_vf_amplify
Definition: vf_amplify.c:316
static int query_formats(AVFilterContext *ctx)
Definition: vf_amplify.c:53
#define FLAGS
Definition: vf_amplify.c:282
static const AVFilterPad inputs[]
Definition: vf_amplify.c:296
static const AVFilterPad outputs[]
Definition: vf_amplify.c:305
static int filter_frame(AVFilterLink *inlink, AVFrame *in)
Definition: vf_amplify.c:244
static const AVOption amplify_options[]
Definition: vf_amplify.c:285
static av_cold int init(AVFilterContext *ctx)
Definition: vf_amplify.c:87
static av_cold void uninit(AVFilterContext *ctx)
Definition: vf_amplify.c:232
#define OFFSET(x)
Definition: vf_amplify.c:281
static int config_output(AVFilterLink *outlink)
Definition: vf_amplify.c:210
static int amplify_frame(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_amplify.c:104
static av_always_inline int diff(const uint32_t a, const uint32_t b)
static const int factor[16]
Definition: vf_pp7.c:77
AVFrame * ff_get_video_buffer(AVFilterLink *link, int w, int h)
Request a picture buffer with a specific set of permissions.
Definition: video.c:104