FFmpeg  4.4.4
vf_atadenoise.c
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2015 Paul B Mahol
3  *
4  * This file is part of FFmpeg.
5  *
6  * FFmpeg is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * FFmpeg is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with FFmpeg; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19  */
20 
21 /**
22  * @file
23  * Adaptive Temporal Averaging Denoiser,
24  * based on paper "Video Denoising Based on Adaptive Temporal Averaging" by
25  * David Bartovčak and Miroslav Vrankić
26  */
27 
28 #include "libavutil/imgutils.h"
29 #include "libavutil/opt.h"
30 #include "libavutil/pixdesc.h"
31 #include "avfilter.h"
32 
33 #define FF_BUFQUEUE_SIZE 129
34 #include "bufferqueue.h"
35 
36 #include "atadenoise.h"
37 #include "formats.h"
38 #include "internal.h"
39 #include "video.h"
40 
41 #define SIZE FF_BUFQUEUE_SIZE
42 
43 typedef struct ATADenoiseContext {
44  const AVClass *class;
45 
46  float fthra[4], fthrb[4];
47  float sigma[4];
48  int thra[4], thrb[4];
49  int algorithm;
50 
51  int planes;
52  int nb_planes;
53  int planewidth[4];
54  int planeheight[4];
55 
56  struct FFBufQueue q;
57  void *data[4][SIZE];
58  int linesize[4][SIZE];
59  float weights[4][SIZE];
60  int size, mid, radius;
61  int available;
62 
63  int (*filter_slice)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs);
64 
67 
68 #define OFFSET(x) offsetof(ATADenoiseContext, x)
69 #define FLAGS AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_RUNTIME_PARAM
70 #define VF AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM
71 
72 static const AVOption atadenoise_options[] = {
73  { "0a", "set threshold A for 1st plane", OFFSET(fthra[0]), AV_OPT_TYPE_FLOAT, {.dbl=0.02}, 0, 0.3, FLAGS },
74  { "0b", "set threshold B for 1st plane", OFFSET(fthrb[0]), AV_OPT_TYPE_FLOAT, {.dbl=0.04}, 0, 5.0, FLAGS },
75  { "1a", "set threshold A for 2nd plane", OFFSET(fthra[1]), AV_OPT_TYPE_FLOAT, {.dbl=0.02}, 0, 0.3, FLAGS },
76  { "1b", "set threshold B for 2nd plane", OFFSET(fthrb[1]), AV_OPT_TYPE_FLOAT, {.dbl=0.04}, 0, 5.0, FLAGS },
77  { "2a", "set threshold A for 3rd plane", OFFSET(fthra[2]), AV_OPT_TYPE_FLOAT, {.dbl=0.02}, 0, 0.3, FLAGS },
78  { "2b", "set threshold B for 3rd plane", OFFSET(fthrb[2]), AV_OPT_TYPE_FLOAT, {.dbl=0.04}, 0, 5.0, FLAGS },
79  { "s", "set how many frames to use", OFFSET(size), AV_OPT_TYPE_INT, {.i64=9}, 5, SIZE, VF },
80  { "p", "set what planes to filter", OFFSET(planes), AV_OPT_TYPE_FLAGS, {.i64=7}, 0, 15, FLAGS },
81  { "a", "set variant of algorithm", OFFSET(algorithm),AV_OPT_TYPE_INT, {.i64=PARALLEL}, 0, NB_ATAA-1, FLAGS, "a" },
82  { "p", "parallel", 0, AV_OPT_TYPE_CONST, {.i64=PARALLEL}, 0, 0, FLAGS, "a" },
83  { "s", "serial", 0, AV_OPT_TYPE_CONST, {.i64=SERIAL}, 0, 0, FLAGS, "a" },
84  { "0s", "set sigma for 1st plane", OFFSET(sigma[0]), AV_OPT_TYPE_FLOAT, {.dbl=INT16_MAX}, 0, INT16_MAX, FLAGS },
85  { "1s", "set sigma for 2nd plane", OFFSET(sigma[1]), AV_OPT_TYPE_FLOAT, {.dbl=INT16_MAX}, 0, INT16_MAX, FLAGS },
86  { "2s", "set sigma for 3rd plane", OFFSET(sigma[2]), AV_OPT_TYPE_FLOAT, {.dbl=INT16_MAX}, 0, INT16_MAX, FLAGS },
87  { NULL }
88 };
89 
91 
93 {
94  static const enum AVPixelFormat pixel_fmts[] = {
122  };
124  if (!formats)
125  return AVERROR(ENOMEM);
127 }
128 
130 {
131  ATADenoiseContext *s = ctx->priv;
132 
133  if (!(s->size & 1)) {
134  av_log(ctx, AV_LOG_WARNING, "size %d is invalid. Must be an odd value, setting it to %d.\n", s->size, s->size|1);
135  s->size |= 1;
136  }
137  s->radius = s->size / 2;
138  s->mid = s->radius;
139 
140  return 0;
141 }
142 
143 typedef struct ThreadData {
144  AVFrame *in, *out;
145 } ThreadData;
146 
147 #define WFILTER_ROW(type, name) \
148 static void fweight_row##name(const uint8_t *ssrc, uint8_t *ddst, \
149  const uint8_t *ssrcf[SIZE], \
150  int w, int mid, int size, \
151  int thra, int thrb, const float *weights) \
152 { \
153  const type *src = (const type *)ssrc; \
154  const type **srcf = (const type **)ssrcf; \
155  type *dst = (type *)ddst; \
156  \
157  for (int x = 0; x < w; x++) { \
158  const int srcx = src[x]; \
159  unsigned lsumdiff = 0, rsumdiff = 0; \
160  unsigned ldiff, rdiff; \
161  float sum = srcx; \
162  float wsum = 1.f; \
163  int l = 0, r = 0; \
164  int srcjx, srcix; \
165  \
166  for (int j = mid - 1, i = mid + 1; j >= 0 && i < size; j--, i++) { \
167  srcjx = srcf[j][x]; \
168  \
169  ldiff = FFABS(srcx - srcjx); \
170  lsumdiff += ldiff; \
171  if (ldiff > thra || \
172  lsumdiff > thrb) \
173  break; \
174  l++; \
175  sum += srcjx * weights[j]; \
176  wsum += weights[j]; \
177  \
178  srcix = srcf[i][x]; \
179  \
180  rdiff = FFABS(srcx - srcix); \
181  rsumdiff += rdiff; \
182  if (rdiff > thra || \
183  rsumdiff > thrb) \
184  break; \
185  r++; \
186  sum += srcix * weights[i]; \
187  wsum += weights[i]; \
188  } \
189  \
190  dst[x] = lrintf(sum / wsum); \
191  } \
192 }
193 
195 WFILTER_ROW(uint16_t, 16)
196 
197 #define WFILTER_ROW_SERIAL(type, name) \
198 static void fweight_row##name##_serial(const uint8_t *ssrc, uint8_t *ddst, \
199  const uint8_t *ssrcf[SIZE], \
200  int w, int mid, int size, \
201  int thra, int thrb, \
202  const float *weights) \
203 { \
204  const type *src = (const type *)ssrc; \
205  const type **srcf = (const type **)ssrcf; \
206  type *dst = (type *)ddst; \
207  \
208  for (int x = 0; x < w; x++) { \
209  const int srcx = src[x]; \
210  unsigned lsumdiff = 0, rsumdiff = 0; \
211  unsigned ldiff, rdiff; \
212  float sum = srcx; \
213  float wsum = 1.f; \
214  int l = 0, r = 0; \
215  int srcjx, srcix; \
216  \
217  for (int j = mid - 1; j >= 0; j--) { \
218  srcjx = srcf[j][x]; \
219  \
220  ldiff = FFABS(srcx - srcjx); \
221  lsumdiff += ldiff; \
222  if (ldiff > thra || \
223  lsumdiff > thrb) \
224  break; \
225  l++; \
226  sum += srcjx * weights[j]; \
227  wsum += weights[j]; \
228  } \
229  \
230  for (int i = mid + 1; i < size; i++) { \
231  srcix = srcf[i][x]; \
232  \
233  rdiff = FFABS(srcx - srcix); \
234  rsumdiff += rdiff; \
235  if (rdiff > thra || \
236  rsumdiff > thrb) \
237  break; \
238  r++; \
239  sum += srcix * weights[i]; \
240  wsum += weights[i]; \
241  } \
242  \
243  dst[x] = lrintf(sum / wsum); \
244  } \
245 }
246 
248 WFILTER_ROW_SERIAL(uint16_t, 16)
249 
250 #define FILTER_ROW(type, name) \
251 static void filter_row##name(const uint8_t *ssrc, uint8_t *ddst, \
252  const uint8_t *ssrcf[SIZE], \
253  int w, int mid, int size, \
254  int thra, int thrb, const float *weights) \
255 { \
256  const type *src = (const type *)ssrc; \
257  const type **srcf = (const type **)ssrcf; \
258  type *dst = (type *)ddst; \
259  \
260  for (int x = 0; x < w; x++) { \
261  const int srcx = src[x]; \
262  unsigned lsumdiff = 0, rsumdiff = 0; \
263  unsigned ldiff, rdiff; \
264  unsigned sum = srcx; \
265  int l = 0, r = 0; \
266  int srcjx, srcix; \
267  \
268  for (int j = mid - 1, i = mid + 1; j >= 0 && i < size; j--, i++) { \
269  srcjx = srcf[j][x]; \
270  \
271  ldiff = FFABS(srcx - srcjx); \
272  lsumdiff += ldiff; \
273  if (ldiff > thra || \
274  lsumdiff > thrb) \
275  break; \
276  l++; \
277  sum += srcjx; \
278  \
279  srcix = srcf[i][x]; \
280  \
281  rdiff = FFABS(srcx - srcix); \
282  rsumdiff += rdiff; \
283  if (rdiff > thra || \
284  rsumdiff > thrb) \
285  break; \
286  r++; \
287  sum += srcix; \
288  } \
289  \
290  dst[x] = (sum + ((r + l + 1) >> 1)) / (r + l + 1); \
291  } \
292 }
293 
295 FILTER_ROW(uint16_t, 16)
296 
297 #define FILTER_ROW_SERIAL(type, name) \
298 static void filter_row##name##_serial(const uint8_t *ssrc, uint8_t *ddst, \
299  const uint8_t *ssrcf[SIZE], \
300  int w, int mid, int size, \
301  int thra, int thrb, \
302  const float *weights) \
303 { \
304  const type *src = (const type *)ssrc; \
305  const type **srcf = (const type **)ssrcf; \
306  type *dst = (type *)ddst; \
307  \
308  for (int x = 0; x < w; x++) { \
309  const int srcx = src[x]; \
310  unsigned lsumdiff = 0, rsumdiff = 0; \
311  unsigned ldiff, rdiff; \
312  unsigned sum = srcx; \
313  int l = 0, r = 0; \
314  int srcjx, srcix; \
315  \
316  for (int j = mid - 1; j >= 0; j--) { \
317  srcjx = srcf[j][x]; \
318  \
319  ldiff = FFABS(srcx - srcjx); \
320  lsumdiff += ldiff; \
321  if (ldiff > thra || \
322  lsumdiff > thrb) \
323  break; \
324  l++; \
325  sum += srcjx; \
326  } \
327  \
328  for (int i = mid + 1; i < size; i++) { \
329  srcix = srcf[i][x]; \
330  \
331  rdiff = FFABS(srcx - srcix); \
332  rsumdiff += rdiff; \
333  if (rdiff > thra || \
334  rsumdiff > thrb) \
335  break; \
336  r++; \
337  sum += srcix; \
338  } \
339  \
340  dst[x] = (sum + ((r + l + 1) >> 1)) / (r + l + 1); \
341  } \
342 }
343 
345 FILTER_ROW_SERIAL(uint16_t, 16)
346 
347 static int filter_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
348 {
349  ATADenoiseContext *s = ctx->priv;
350  ThreadData *td = arg;
351  AVFrame *in = td->in;
352  AVFrame *out = td->out;
353  const int size = s->size;
354  const int mid = s->mid;
355  int p, y, i;
356 
357  for (p = 0; p < s->nb_planes; p++) {
358  const float *weights = s->weights[p];
359  const int h = s->planeheight[p];
360  const int w = s->planewidth[p];
361  const int slice_start = (h * jobnr) / nb_jobs;
362  const int slice_end = (h * (jobnr+1)) / nb_jobs;
363  const uint8_t *src = in->data[p] + slice_start * in->linesize[p];
364  uint8_t *dst = out->data[p] + slice_start * out->linesize[p];
365  const int thra = s->thra[p];
366  const int thrb = s->thrb[p];
367  const uint8_t **data = (const uint8_t **)s->data[p];
368  const int *linesize = (const int *)s->linesize[p];
369  const uint8_t *srcf[SIZE];
370 
371  if (!((1 << p) & s->planes)) {
372  av_image_copy_plane(dst, out->linesize[p], src, in->linesize[p],
373  w, slice_end - slice_start);
374  continue;
375  }
376 
377  for (i = 0; i < size; i++)
378  srcf[i] = data[i] + slice_start * linesize[i];
379 
380  for (y = slice_start; y < slice_end; y++) {
381  s->dsp.filter_row[p](src, dst, srcf, w, mid, size, thra, thrb, weights);
382 
383  dst += out->linesize[p];
384  src += in->linesize[p];
385 
386  for (i = 0; i < size; i++)
387  srcf[i] += linesize[i];
388  }
389  }
390 
391  return 0;
392 }
393 
394 static int config_input(AVFilterLink *inlink)
395 {
397  AVFilterContext *ctx = inlink->dst;
398  ATADenoiseContext *s = ctx->priv;
399  int depth;
400 
401  s->nb_planes = desc->nb_components;
402 
403  s->planeheight[1] = s->planeheight[2] = AV_CEIL_RSHIFT(inlink->h, desc->log2_chroma_h);
404  s->planeheight[0] = s->planeheight[3] = inlink->h;
405  s->planewidth[1] = s->planewidth[2] = AV_CEIL_RSHIFT(inlink->w, desc->log2_chroma_w);
406  s->planewidth[0] = s->planewidth[3] = inlink->w;
407 
408  depth = desc->comp[0].depth;
409  s->filter_slice = filter_slice;
410 
411  for (int p = 0; p < s->nb_planes; p++) {
412  if (depth == 8 && s->sigma[p] == INT16_MAX)
413  s->dsp.filter_row[p] = s->algorithm == PARALLEL ? filter_row8 : filter_row8_serial;
414  else if (s->sigma[p] == INT16_MAX)
415  s->dsp.filter_row[p] = s->algorithm == PARALLEL ? filter_row16 : filter_row16_serial;
416  else if (depth == 8 && s->sigma[p] < INT16_MAX)
417  s->dsp.filter_row[p] = s->algorithm == PARALLEL ? fweight_row8 : fweight_row8_serial;
418  else if (s->sigma[p] < INT16_MAX)
419  s->dsp.filter_row[p] = s->algorithm == PARALLEL ? fweight_row16 : fweight_row16_serial;
420  }
421 
422  s->thra[0] = s->fthra[0] * (1 << depth) - 1;
423  s->thra[1] = s->fthra[1] * (1 << depth) - 1;
424  s->thra[2] = s->fthra[2] * (1 << depth) - 1;
425  s->thrb[0] = s->fthrb[0] * (1 << depth) - 1;
426  s->thrb[1] = s->fthrb[1] * (1 << depth) - 1;
427  s->thrb[2] = s->fthrb[2] * (1 << depth) - 1;
428 
429  for (int p = 0; p < s->nb_planes; p++) {
430  float sigma = s->radius * s->sigma[p];
431 
432  s->weights[p][s->radius] = 1.f;
433  for (int n = 1; n <= s->radius; n++) {
434  s->weights[p][s->radius + n] =
435  s->weights[p][s->radius - n] = expf(-0.5 * (n + 1) * (n + 1) / (sigma * sigma));
436  }
437  }
438 
439  if (ARCH_X86)
440  ff_atadenoise_init_x86(&s->dsp, depth, s->algorithm, s->sigma);
441 
442  return 0;
443 }
444 
445 static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
446 {
447  AVFilterContext *ctx = inlink->dst;
448  AVFilterLink *outlink = ctx->outputs[0];
449  ATADenoiseContext *s = ctx->priv;
450  AVFrame *out, *in;
451  int i;
452 
453  if (s->q.available != s->size) {
454  if (s->q.available < s->mid) {
455  for (i = 0; i < s->mid; i++) {
456  out = av_frame_clone(buf);
457  if (!out) {
458  av_frame_free(&buf);
459  return AVERROR(ENOMEM);
460  }
461  ff_bufqueue_add(ctx, &s->q, out);
462  }
463  }
464  if (s->q.available < s->size) {
465  ff_bufqueue_add(ctx, &s->q, buf);
466  s->available++;
467  }
468  return 0;
469  }
470 
471  in = ff_bufqueue_peek(&s->q, s->mid);
472 
473  if (!ctx->is_disabled) {
474  ThreadData td;
475 
476  out = ff_get_video_buffer(outlink, outlink->w, outlink->h);
477  if (!out) {
478  av_frame_free(&buf);
479  return AVERROR(ENOMEM);
480  }
481 
482  for (i = 0; i < s->size; i++) {
483  AVFrame *frame = ff_bufqueue_peek(&s->q, i);
484 
485  s->data[0][i] = frame->data[0];
486  s->data[1][i] = frame->data[1];
487  s->data[2][i] = frame->data[2];
488  s->linesize[0][i] = frame->linesize[0];
489  s->linesize[1][i] = frame->linesize[1];
490  s->linesize[2][i] = frame->linesize[2];
491  }
492 
493  td.in = in; td.out = out;
494  ctx->internal->execute(ctx, s->filter_slice, &td, NULL,
495  FFMIN3(s->planeheight[1],
496  s->planeheight[2],
499  } else {
500  out = av_frame_clone(in);
501  if (!out) {
502  av_frame_free(&buf);
503  return AVERROR(ENOMEM);
504  }
505  }
506 
507  in = ff_bufqueue_get(&s->q);
508  av_frame_free(&in);
509  ff_bufqueue_add(ctx, &s->q, buf);
510 
511  return ff_filter_frame(outlink, out);
512 }
513 
514 static int request_frame(AVFilterLink *outlink)
515 {
516  AVFilterContext *ctx = outlink->src;
517  ATADenoiseContext *s = ctx->priv;
518  int ret = 0;
519 
520  ret = ff_request_frame(ctx->inputs[0]);
521 
522  if (ret == AVERROR_EOF && !ctx->is_disabled && s->available) {
523  AVFrame *buf = av_frame_clone(ff_bufqueue_peek(&s->q, s->available));
524  if (!buf)
525  return AVERROR(ENOMEM);
526 
527  ret = filter_frame(ctx->inputs[0], buf);
528  s->available--;
529  }
530 
531  return ret;
532 }
533 
535 {
536  ATADenoiseContext *s = ctx->priv;
537 
539 }
540 
542  const char *cmd,
543  const char *arg,
544  char *res,
545  int res_len,
546  int flags)
547 {
548  int ret = ff_filter_process_command(ctx, cmd, arg, res, res_len, flags);
549 
550  if (ret < 0)
551  return ret;
552 
553  return config_input(ctx->inputs[0]);
554 }
555 
556 static const AVFilterPad inputs[] = {
557  {
558  .name = "default",
559  .type = AVMEDIA_TYPE_VIDEO,
560  .filter_frame = filter_frame,
561  .config_props = config_input,
562  },
563  { NULL }
564 };
565 
566 static const AVFilterPad outputs[] = {
567  {
568  .name = "default",
569  .type = AVMEDIA_TYPE_VIDEO,
570  .request_frame = request_frame,
571  },
572  { NULL }
573 };
574 
576  .name = "atadenoise",
577  .description = NULL_IF_CONFIG_SMALL("Apply an Adaptive Temporal Averaging Denoiser."),
578  .priv_size = sizeof(ATADenoiseContext),
579  .priv_class = &atadenoise_class,
580  .init = init,
581  .uninit = uninit,
583  .inputs = inputs,
584  .outputs = outputs,
587 };
@ NB_ATAA
Definition: atadenoise.h:30
@ PARALLEL
Definition: atadenoise.h:28
@ SERIAL
Definition: atadenoise.h:29
void ff_atadenoise_init_x86(ATADenoiseDSPContext *dsp, int depth, int algorithm, const float *sigma)
#define av_cold
Definition: attributes.h:88
uint8_t pi<< 24) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi - 0x80) *(1.0f/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi - 0x80) *(1.0/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S16, int16_t,(*(const int16_t *) pi >> 8)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S16, int16_t, *(const int16_t *) pi *(1.0f/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S16, int16_t, *(const int16_t *) pi *(1.0/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S32, int32_t,(*(const int32_t *) pi >> 24)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S32, int32_t, *(const int32_t *) pi *(1.0f/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S32, int32_t, *(const int32_t *) pi *(1.0/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_FLT, float, av_clip_uint8(lrintf(*(const float *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_FLT, float, av_clip_int16(lrintf(*(const float *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_FLT, float, av_clipl_int32(llrintf(*(const float *) pi *(1U<< 31)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_DBL, double, av_clip_uint8(lrint(*(const double *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_DBL, double, av_clip_int16(lrint(*(const double *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_DBL, double, av_clipl_int32(llrint(*(const double *) pi *(1U<< 31)))) #define SET_CONV_FUNC_GROUP(ofmt, ifmt) static void set_generic_function(AudioConvert *ac) { } void ff_audio_convert_free(AudioConvert **ac) { if(! *ac) return;ff_dither_free(&(*ac) ->dc);av_freep(ac);} AudioConvert *ff_audio_convert_alloc(AVAudioResampleContext *avr, enum AVSampleFormat out_fmt, enum AVSampleFormat in_fmt, int channels, int sample_rate, int apply_map) { AudioConvert *ac;int in_planar, out_planar;ac=av_mallocz(sizeof(*ac));if(!ac) return NULL;ac->avr=avr;ac->out_fmt=out_fmt;ac->in_fmt=in_fmt;ac->channels=channels;ac->apply_map=apply_map;if(avr->dither_method !=AV_RESAMPLE_DITHER_NONE &&av_get_packed_sample_fmt(out_fmt)==AV_SAMPLE_FMT_S16 &&av_get_bytes_per_sample(in_fmt) > 2) { ac->dc=ff_dither_alloc(avr, out_fmt, in_fmt, channels, sample_rate, apply_map);if(!ac->dc) { av_free(ac);return NULL;} return ac;} in_planar=ff_sample_fmt_is_planar(in_fmt, channels);out_planar=ff_sample_fmt_is_planar(out_fmt, channels);if(in_planar==out_planar) { ac->func_type=CONV_FUNC_TYPE_FLAT;ac->planes=in_planar ? ac->channels :1;} else if(in_planar) ac->func_type=CONV_FUNC_TYPE_INTERLEAVE;else ac->func_type=CONV_FUNC_TYPE_DEINTERLEAVE;set_generic_function(ac);if(ARCH_AARCH64) ff_audio_convert_init_aarch64(ac);if(ARCH_ARM) ff_audio_convert_init_arm(ac);if(ARCH_X86) ff_audio_convert_init_x86(ac);return ac;} int ff_audio_convert(AudioConvert *ac, AudioData *out, AudioData *in) { int use_generic=1;int len=in->nb_samples;int p;if(ac->dc) { av_log(ac->avr, AV_LOG_TRACE, "%d samples - audio_convert: %s to %s (dithered)\n", len, av_get_sample_fmt_name(ac->in_fmt), av_get_sample_fmt_name(ac->out_fmt));return ff_convert_dither(ac-> in
uint8_t
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
Definition: avfilter.c:1096
int ff_filter_process_command(AVFilterContext *ctx, const char *cmd, const char *arg, char *res, int res_len, int flags)
Generic processing of user supplied commands that are set in the same way as the filter options.
Definition: avfilter.c:882
int ff_request_frame(AVFilterLink *link)
Request an input frame from the filter at the other end of the link.
Definition: avfilter.c:408
int ff_filter_get_nb_threads(AVFilterContext *ctx)
Get number of threads for current filter instance.
Definition: avfilter.c:802
Main libavfilter public API header.
static void ff_bufqueue_add(void *log, struct FFBufQueue *queue, AVFrame *buf)
Add a buffer to the queue.
Definition: bufferqueue.h:71
static AVFrame * ff_bufqueue_peek(struct FFBufQueue *queue, unsigned index)
Get a buffer from the queue without altering it.
Definition: bufferqueue.h:87
static void ff_bufqueue_discard_all(struct FFBufQueue *queue)
Unref and remove all buffers from the queue.
Definition: bufferqueue.h:111
static AVFrame * ff_bufqueue_get(struct FFBufQueue *queue)
Get the first buffer from the queue and remove it.
Definition: bufferqueue.h:98
#define flags(name, subs,...)
Definition: cbs_av1.c:561
#define s(width, name)
Definition: cbs_vp9.c:257
#define AV_CEIL_RSHIFT(a, b)
Definition: common.h:58
#define FFMIN3(a, b, c)
Definition: common.h:106
#define ARCH_X86
Definition: config.h:39
#define NULL
Definition: coverity.c:32
static AVFrame * frame
int
int ff_set_common_formats(AVFilterContext *ctx, AVFilterFormats *formats)
A helper for query_formats() which sets all links to the same list of formats.
Definition: formats.c:587
AVFilterFormats * ff_make_format_list(const int *fmts)
Create a list of supported formats.
Definition: formats.c:286
@ AV_OPT_TYPE_CONST
Definition: opt.h:234
@ AV_OPT_TYPE_FLAGS
Definition: opt.h:224
@ AV_OPT_TYPE_INT
Definition: opt.h:225
@ AV_OPT_TYPE_FLOAT
Definition: opt.h:228
#define AVFILTER_FLAG_SLICE_THREADS
The filter supports multithreading by splitting frames into multiple parts and processing them concur...
Definition: avfilter.h:117
#define AVFILTER_FLAG_SUPPORT_TIMELINE_INTERNAL
Same as AVFILTER_FLAG_SUPPORT_TIMELINE_GENERIC, except that the filter will have its filter_frame() c...
Definition: avfilter.h:134
#define AVERROR_EOF
End of file.
Definition: error.h:55
#define AVERROR(e)
Definition: error.h:43
AVFrame * av_frame_clone(const AVFrame *src)
Create a new frame that references the same data as src.
Definition: frame.c:540
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
Definition: frame.c:203
int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
Copy only "metadata" fields from src to dst.
Definition: frame.c:658
#define AV_LOG_WARNING
Something somehow does not look correct.
Definition: log.h:200
@ AVMEDIA_TYPE_VIDEO
Definition: avutil.h:201
void av_image_copy_plane(uint8_t *dst, int dst_linesize, const uint8_t *src, int src_linesize, int bytewidth, int height)
Copy image plane from src to dst.
Definition: imgutils.c:373
static const int weights[]
Definition: hevc_pel.c:32
misc image utilities
int i
Definition: input.c:407
const char * arg
Definition: jacosubdec.c:66
common internal API header
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
Definition: internal.h:117
#define expf(x)
Definition: libm.h:283
const char * desc
Definition: libsvtav1.c:79
uint8_t w
Definition: llviddspenc.c:39
static const struct @322 planes[]
static int slice_end(AVCodecContext *avctx, AVFrame *pict)
Handle slice ends.
Definition: mpeg12dec.c:2033
const char data[16]
Definition: mxf.c:142
AVOptions.
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:2573
#define AV_PIX_FMT_GBRAP12
Definition: pixfmt.h:420
#define AV_PIX_FMT_YUV420P16
Definition: pixfmt.h:410
#define AV_PIX_FMT_YUV444P12
Definition: pixfmt.h:406
#define AV_PIX_FMT_YUV444P9
Definition: pixfmt.h:398
#define AV_PIX_FMT_YUV420P10
Definition: pixfmt.h:399
#define AV_PIX_FMT_YUV440P12
Definition: pixfmt.h:405
#define AV_PIX_FMT_GRAY9
Definition: pixfmt.h:379
#define AV_PIX_FMT_GBRAP16
Definition: pixfmt.h:421
#define AV_PIX_FMT_GBRP9
Definition: pixfmt.h:414
#define AV_PIX_FMT_YUV422P9
Definition: pixfmt.h:397
#define AV_PIX_FMT_YUVA444P10
Definition: pixfmt.h:438
#define AV_PIX_FMT_YUVA420P16
Definition: pixfmt.h:441
#define AV_PIX_FMT_YUV420P12
Definition: pixfmt.h:403
#define AV_PIX_FMT_YUVA420P10
Definition: pixfmt.h:436
#define AV_PIX_FMT_YUVA422P9
Definition: pixfmt.h:434
#define AV_PIX_FMT_YUV422P12
Definition: pixfmt.h:404
#define AV_PIX_FMT_GBRP10
Definition: pixfmt.h:415
#define AV_PIX_FMT_YUV422P10
Definition: pixfmt.h:400
#define AV_PIX_FMT_GRAY12
Definition: pixfmt.h:381
#define AV_PIX_FMT_GBRP12
Definition: pixfmt.h:416
#define AV_PIX_FMT_YUV420P9
Definition: pixfmt.h:396
#define AV_PIX_FMT_YUVA420P9
Definition: pixfmt.h:433
#define AV_PIX_FMT_YUVA422P10
Definition: pixfmt.h:437
#define AV_PIX_FMT_YUV420P14
Definition: pixfmt.h:407
AVPixelFormat
Pixel format.
Definition: pixfmt.h:64
@ AV_PIX_FMT_NONE
Definition: pixfmt.h:65
@ AV_PIX_FMT_YUV420P
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
Definition: pixfmt.h:66
@ AV_PIX_FMT_YUV440P
planar YUV 4:4:0 (1 Cr & Cb sample per 1x2 Y samples)
Definition: pixfmt.h:99
@ AV_PIX_FMT_YUV422P
planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples)
Definition: pixfmt.h:70
@ AV_PIX_FMT_GRAY8
Y , 8bpp.
Definition: pixfmt.h:74
@ AV_PIX_FMT_YUVA420P
planar YUV 4:2:0, 20bpp, (1 Cr & Cb sample per 2x2 Y & A samples)
Definition: pixfmt.h:101
@ AV_PIX_FMT_YUVJ440P
planar YUV 4:4:0 full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV440P and setting color_range
Definition: pixfmt.h:100
@ AV_PIX_FMT_YUV410P
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
Definition: pixfmt.h:72
@ AV_PIX_FMT_YUV411P
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples)
Definition: pixfmt.h:73
@ AV_PIX_FMT_YUV444P
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
Definition: pixfmt.h:71
@ AV_PIX_FMT_YUVA444P
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
Definition: pixfmt.h:177
@ AV_PIX_FMT_YUVJ411P
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples) full scale (JPEG), deprecated in favor ...
Definition: pixfmt.h:258
@ AV_PIX_FMT_GBRAP
planar GBRA 4:4:4:4 32bpp
Definition: pixfmt.h:215
@ AV_PIX_FMT_YUVJ422P
planar YUV 4:2:2, 16bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV422P and setting col...
Definition: pixfmt.h:79
@ AV_PIX_FMT_YUVA422P
planar YUV 4:2:2 24bpp, (1 Cr & Cb sample per 2x1 Y & A samples)
Definition: pixfmt.h:176
@ AV_PIX_FMT_GBRP
planar GBR 4:4:4 24bpp
Definition: pixfmt.h:168
@ AV_PIX_FMT_YUVJ444P
planar YUV 4:4:4, 24bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV444P and setting col...
Definition: pixfmt.h:80
@ AV_PIX_FMT_YUVJ420P
planar YUV 4:2:0, 12bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV420P and setting col...
Definition: pixfmt.h:78
#define AV_PIX_FMT_YUVA422P12
Definition: pixfmt.h:439
#define AV_PIX_FMT_YUV422P14
Definition: pixfmt.h:408
#define AV_PIX_FMT_GRAY10
Definition: pixfmt.h:380
#define AV_PIX_FMT_GRAY14
Definition: pixfmt.h:382
#define AV_PIX_FMT_YUV422P16
Definition: pixfmt.h:411
#define AV_PIX_FMT_YUV440P10
Definition: pixfmt.h:401
#define AV_PIX_FMT_GRAY16
Definition: pixfmt.h:383
#define AV_PIX_FMT_GBRAP10
Definition: pixfmt.h:419
#define AV_PIX_FMT_YUVA444P16
Definition: pixfmt.h:443
#define AV_PIX_FMT_YUVA422P16
Definition: pixfmt.h:442
#define AV_PIX_FMT_GBRP16
Definition: pixfmt.h:418
#define AV_PIX_FMT_YUV444P14
Definition: pixfmt.h:409
#define AV_PIX_FMT_YUVA444P9
Definition: pixfmt.h:435
#define AV_PIX_FMT_GBRP14
Definition: pixfmt.h:417
#define AV_PIX_FMT_YUVA444P12
Definition: pixfmt.h:440
#define AV_PIX_FMT_YUV444P16
Definition: pixfmt.h:412
#define AV_PIX_FMT_YUV444P10
Definition: pixfmt.h:402
#define td
Definition: regdef.h:70
formats
Definition: signature.h:48
int(* filter_slice)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
Definition: vf_atadenoise.c:63
int linesize[4][SIZE]
Definition: vf_atadenoise.c:58
float weights[4][SIZE]
Definition: vf_atadenoise.c:59
struct FFBufQueue q
Definition: vf_atadenoise.c:56
ATADenoiseDSPContext dsp
Definition: vf_atadenoise.c:65
void * data[4][SIZE]
Definition: vf_atadenoise.c:57
Describe the class of an AVClass context structure.
Definition: log.h:67
An instance of a filter.
Definition: avfilter.h:341
A list of supported formats for one end of a filter link.
Definition: formats.h:65
A filter pad used for either input or output.
Definition: internal.h:54
const char * name
Pad name.
Definition: internal.h:60
Filter definition.
Definition: avfilter.h:145
const char * name
Filter name.
Definition: avfilter.h:149
AVFormatInternal * internal
An opaque field for libavformat internal usage.
Definition: avformat.h:1699
This structure describes decoded (raw) audio or video data.
Definition: frame.h:318
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:332
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
Definition: frame.h:349
AVOption.
Definition: opt.h:248
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
Definition: pixdesc.h:81
Structure holding the queue.
Definition: bufferqueue.h:49
Used for passing data between threads.
Definition: dsddec.c:67
AVFrame * out
Definition: af_adeclick.c:502
AVFrame * in
Definition: af_adenorm.c:223
#define av_log(a,...)
#define src
Definition: vp8dsp.c:255
FILE * out
Definition: movenc.c:54
AVFormatContext * ctx
Definition: movenc.c:48
int size
static int process_command(AVFilterContext *ctx, const char *cmd, const char *arg, char *res, int res_len, int flags)
static int filter_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
#define FILTER_ROW(type, name)
#define FILTER_ROW_SERIAL(type, name)
#define WFILTER_ROW(type, name)
static int query_formats(AVFilterContext *ctx)
Definition: vf_atadenoise.c:92
static int config_input(AVFilterLink *inlink)
#define FLAGS
Definition: vf_atadenoise.c:69
static const AVFilterPad inputs[]
AVFilter ff_vf_atadenoise
#define SIZE
Definition: vf_atadenoise.c:41
static int request_frame(AVFilterLink *outlink)
static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
static const AVFilterPad outputs[]
#define VF
Definition: vf_atadenoise.c:70
static const AVOption atadenoise_options[]
Definition: vf_atadenoise.c:72
static av_cold int init(AVFilterContext *ctx)
static av_cold void uninit(AVFilterContext *ctx)
#define OFFSET(x)
Definition: vf_atadenoise.c:68
AVFILTER_DEFINE_CLASS(atadenoise)
#define WFILTER_ROW_SERIAL(type, name)
AVFrame * ff_get_video_buffer(AVFilterLink *link, int w, int h)
Request a picture buffer with a specific set of permissions.
Definition: video.c:104