FFmpeg
vf_blurdetect.c
Go to the documentation of this file.
1 /*
2  * Copyright (c) 2021 Thilo Borgmann <thilo.borgmann _at_ mail.de>
3  *
4  * This file is part of FFmpeg.
5  *
6  * FFmpeg is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * FFmpeg is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with FFmpeg; if not, write to the Free Software
18  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19  */
20 
21 /**
22  * @file
23  * No-reference blurdetect filter
24  *
25  * Implementing:
26  * Marziliano, Pina, et al. "A no-reference perceptual blur metric." Proceedings.
27  * International conference on image processing. Vol. 3. IEEE, 2002.
28  * https://infoscience.epfl.ch/record/111802/files/14%20A%20no-reference%20perceptual%20blur%20metric.pdf
29  *
30  * @author Thilo Borgmann <thilo.borgmann _at_ mail.de>
31  */
32 
33 #include "libavutil/mem.h"
34 #include "libavutil/opt.h"
35 #include "libavutil/pixdesc.h"
36 #include "libavutil/qsort.h"
37 
38 #include "filters.h"
39 #include "edge_common.h"
40 #include "video.h"
41 
42 static int comp(const float *a,const float *b)
43 {
44  return FFDIFFSIGN(*a, *b);
45 }
46 
47 typedef struct BLRContext {
48  const AVClass *class;
49 
50  int hsub, vsub;
51  int nb_planes;
52 
53  float low, high;
54  uint8_t low_u8, high_u8;
55  int radius; // radius during local maxima detection
56  int block_pct; // percentage of "sharpest" blocks in the image to use for bluriness calculation
57  int block_width; // width for block abbreviation
58  int block_height; // height for block abbreviation
59  int planes; // number of planes to filter
60 
61  double blur_total;
62  uint64_t nb_frames;
63 
64  float *blks;
65  uint8_t *filterbuf;
66  uint8_t *tmpbuf;
67  uint16_t *gradients;
68  int8_t *directions;
69 } BLRContext;
70 
71 #define OFFSET(x) offsetof(BLRContext, x)
72 #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
73 static const AVOption blurdetect_options[] = {
74  { "high", "set high threshold", OFFSET(high), AV_OPT_TYPE_FLOAT, {.dbl=30/255.}, 0, 1, FLAGS },
75  { "low", "set low threshold", OFFSET(low), AV_OPT_TYPE_FLOAT, {.dbl=15/255.}, 0, 1, FLAGS },
76  { "radius", "search radius for maxima detection", OFFSET(radius), AV_OPT_TYPE_INT, {.i64=50}, 1, 100, FLAGS },
77  { "block_pct", "block pooling threshold when calculating blurriness", OFFSET(block_pct), AV_OPT_TYPE_INT, {.i64=80}, 1, 100, FLAGS },
78  { "block_width", "block size for block-based abbreviation of blurriness", OFFSET(block_width), AV_OPT_TYPE_INT, {.i64=-1}, -1, INT_MAX, FLAGS },
79  { "block_height", "block size for block-based abbreviation of blurriness", OFFSET(block_height), AV_OPT_TYPE_INT, {.i64=-1}, -1, INT_MAX, FLAGS },
80  { "planes", "set planes to filter", OFFSET(planes), AV_OPT_TYPE_INT, {.i64=1}, 0, 15, FLAGS },
81  { NULL }
82 };
83 
84 AVFILTER_DEFINE_CLASS(blurdetect);
85 
87 {
88  BLRContext *s = ctx->priv;
89 
90  s->low_u8 = s->low * 255. + .5;
91  s->high_u8 = s->high * 255. + .5;
92 
93  return 0;
94 }
95 
97 {
98  AVFilterContext *ctx = inlink->dst;
99  BLRContext *s = ctx->priv;
100  const int bufsize = inlink->w * inlink->h;
101  const AVPixFmtDescriptor *pix_desc;
102 
103  pix_desc = av_pix_fmt_desc_get(inlink->format);
104  s->hsub = pix_desc->log2_chroma_w;
105  s->vsub = pix_desc->log2_chroma_h;
106  s->nb_planes = av_pix_fmt_count_planes(inlink->format);
107 
108  if (s->block_width < 1 || s->block_height < 1) {
109  s->block_width = inlink->w;
110  s->block_height = inlink->h;
111  }
112 
113  s->tmpbuf = av_malloc(bufsize);
114  s->filterbuf = av_malloc(bufsize);
115  s->gradients = av_calloc(bufsize, sizeof(*s->gradients));
116  s->directions = av_malloc(bufsize);
117  s->blks = av_calloc((inlink->w / s->block_width) * (inlink->h / s->block_height),
118  sizeof(*s->blks));
119 
120  if (!s->tmpbuf || !s->filterbuf || !s->gradients || !s->directions || !s->blks)
121  return AVERROR(ENOMEM);
122 
123  return 0;
124 }
125 
126 // edge width is defined as the distance between surrounding maxima of the edge pixel
127 static float edge_width(BLRContext *blr, int i, int j, int8_t dir, int w, int h,
128  int edge, const uint8_t *src, int src_linesize)
129 {
130  float width = 0;
131  int dX, dY;
132  int sign;
133  int tmp;
134  int p1;
135  int p2;
136  int k, x, y;
137  int radius = blr->radius;
138 
139  switch(dir) {
140  case DIRECTION_HORIZONTAL: dX = 1; dY = 0; break;
141  case DIRECTION_VERTICAL: dX = 0; dY = 1; break;
142  case DIRECTION_45UP: dX = 1; dY = -1; break;
143  case DIRECTION_45DOWN: dX = 1; dY = 1; break;
144  default: dX = 1; dY = 1; break;
145  }
146 
147  // determines if search in direction dX/dY is looking for a maximum or minimum
148  sign = src[j * src_linesize + i] > src[(j - dY) * src_linesize + i - dX] ? 1 : -1;
149 
150  // search in -(dX/dY) direction
151  for (k = 0; k < radius; k++) {
152  x = i - k*dX;
153  y = j - k*dY;
154  p1 = y * src_linesize + x;
155  x -= dX;
156  y -= dY;
157  p2 = y * src_linesize + x;
158  if (x < 0 || x >= w || y < 0 || y >= h)
159  return 0;
160 
161  tmp = (src[p1] - src[p2]) * sign;
162 
163  if (tmp <= 0) // local maximum found
164  break;
165  }
166  width += k;
167 
168  // search in +(dX/dY) direction
169  for (k = 0; k < radius; k++) {
170  x = i + k * dX;
171  y = j + k * dY;
172  p1 = y * src_linesize + x;
173  x += dX;
174  y += dY;
175  p2 = y * src_linesize + x;
176  if (x < 0 || x >= w || y < 0 || y >= h)
177  return 0;
178 
179  tmp = (src[p1] - src[p2]) * sign;
180 
181  if (tmp >= 0) // local maximum found
182  break;
183  }
184  width += k;
185 
186  // for 45 degree directions approximate edge width in pixel units: 0.7 ~= sqrt(2)/2
187  if (dir == DIRECTION_45UP || dir == DIRECTION_45DOWN)
188  width *= 0.7;
189 
190  return width;
191 }
192 
193 static float calculate_blur(BLRContext *s, int w, int h, int hsub, int vsub,
194  int8_t* dir, int dir_linesize,
195  uint8_t* dst, int dst_linesize,
196  uint8_t* src, int src_linesize)
197 {
198  float total_width = 0.0;
199 
200  int blkcnt = 0;
201 
202  float *blks = s->blks;
203  float block_pool_threshold = s->block_pct / 100.0;
204 
205  int block_width = AV_CEIL_RSHIFT(s->block_width, hsub);
206  int block_height = AV_CEIL_RSHIFT(s->block_height, vsub);
207  int brows = h / block_height;
208  int bcols = w / block_width;
209 
210  for (int blkj = 0; blkj < brows; blkj++) {
211  for (int blki = 0; blki < bcols; blki++) {
212  double block_total_width = 0.0;
213  int block_count = 0;
214  for (int inj = 0; inj < block_height; inj++) {
215  for (int ini = 0; ini < block_width; ini++) {
216  int i = blki * block_width + ini;
217  int j = blkj * block_height + inj;
218 
219  if (dst[j * dst_linesize + i] > 0) {
220  float width = edge_width(s, i, j, dir[j*dir_linesize+i],
221  w, h, dst[j*dst_linesize+i],
222  src, src_linesize);
223  if (width > 0.001) { // throw away zeros
224  block_count++;
225  block_total_width += width;
226  }
227  }
228  }
229  }
230  // if not enough edge pixels in a block, consider it smooth
231  if (block_total_width >= 2 && block_count) {
232  blks[blkcnt] = block_total_width / block_count;
233  blkcnt++;
234  }
235  }
236  }
237 
238  // simple block pooling by sorting and keeping the sharper blocks
239  AV_QSORT(blks, blkcnt, float, comp);
240  blkcnt = ceil(blkcnt * block_pool_threshold);
241  for (int i = 0; i < blkcnt; i++) {
242  total_width += blks[i];
243  }
244 
245  return total_width / blkcnt;
246 }
247 
248 static void set_meta(AVDictionary **metadata, const char *key, float d)
249 {
250  char value[128];
251  snprintf(value, sizeof(value), "%f", d);
253 }
254 
256 {
258  AVFilterContext *ctx = inlink->dst;
259  BLRContext *s = ctx->priv;
260  AVFilterLink *outlink = ctx->outputs[0];
261 
262  const int inw = inlink->w;
263  const int inh = inlink->h;
264 
265  uint8_t *tmpbuf = s->tmpbuf;
266  uint8_t *filterbuf = s->filterbuf;
267  uint16_t *gradients = s->gradients;
268  int8_t *directions = s->directions;
269 
270  float blur = 0.0f;
271  int nplanes = 0;
273  metadata = &in->metadata;
274 
275  for (int plane = 0; plane < s->nb_planes; plane++) {
276  int hsub = plane == 1 || plane == 2 ? s->hsub : 0;
277  int vsub = plane == 1 || plane == 2 ? s->vsub : 0;
278  int w = AV_CEIL_RSHIFT(inw, hsub);
279  int h = AV_CEIL_RSHIFT(inh, vsub);
280 
281  if (!((1 << plane) & s->planes))
282  continue;
283 
284  nplanes++;
285 
286  // gaussian filter to reduce noise
287  ff_gaussian_blur_8(w, h,
288  filterbuf, w,
289  in->data[plane], in->linesize[plane], 1);
290 
291  // compute the 16-bits gradients and directions for the next step
292  ff_sobel_8(w, h, gradients, w, directions, w, filterbuf, w, 1);
293 
294  // non_maximum_suppression() will actually keep & clip what's necessary and
295  // ignore the rest, so we need a clean output buffer
296  memset(tmpbuf, 0, inw * inh);
297  ff_non_maximum_suppression(w, h, tmpbuf, w, directions, w, gradients, w);
298 
299 
300  // keep high values, or low values surrounded by high values
301  ff_double_threshold(s->low_u8, s->high_u8, w, h,
302  tmpbuf, w, tmpbuf, w);
303 
304  blur += calculate_blur(s, w, h, hsub, vsub, directions, w,
305  tmpbuf, w, filterbuf, w);
306  }
307 
308  if (nplanes)
309  blur /= nplanes;
310 
311  s->blur_total += blur;
312 
313  // write stats
314  av_log(ctx, AV_LOG_VERBOSE, "blur: %.7f\n", blur);
315 
316  set_meta(metadata, "lavfi.blur", blur);
317 
318  s->nb_frames = inl->frame_count_in;
319 
320  return ff_filter_frame(outlink, in);
321 }
322 
324 {
325  BLRContext *s = ctx->priv;
326 
327  if (s->nb_frames > 0) {
328  av_log(ctx, AV_LOG_INFO, "blur mean: %.7f\n",
329  s->blur_total / s->nb_frames);
330  }
331 
332  av_freep(&s->tmpbuf);
333  av_freep(&s->filterbuf);
334  av_freep(&s->gradients);
335  av_freep(&s->directions);
336  av_freep(&s->blks);
337 }
338 
339 static const enum AVPixelFormat pix_fmts[] = {
349 };
350 
351 static const AVFilterPad blurdetect_inputs[] = {
352  {
353  .name = "default",
354  .type = AVMEDIA_TYPE_VIDEO,
355  .config_props = blurdetect_config_input,
356  .filter_frame = blurdetect_filter_frame,
357  },
358 };
359 
361  .p.name = "blurdetect",
362  .p.description = NULL_IF_CONFIG_SMALL("Blurdetect filter."),
363  .p.priv_class = &blurdetect_class,
364  .p.flags = AVFILTER_FLAG_METADATA_ONLY,
365  .priv_size = sizeof(BLRContext),
371 };
AVPixelFormat
AVPixelFormat
Pixel format.
Definition: pixfmt.h:71
AVERROR
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
opt.h
BLRContext::hsub
int hsub
Definition: vf_blurdetect.c:50
FILTER_PIXFMTS_ARRAY
#define FILTER_PIXFMTS_ARRAY(array)
Definition: filters.h:244
BLRContext::gradients
uint16_t * gradients
Definition: vf_blurdetect.c:67
ff_filter_frame
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
Definition: avfilter.c:1067
av_pix_fmt_desc_get
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:3456
BLRContext::high_u8
uint8_t high_u8
Definition: vf_blurdetect.c:54
inlink
The exact code depends on how similar the blocks are and how related they are to the and needs to apply these operations to the correct inlink or outlink if there are several Macros are available to factor that when no extra processing is inlink
Definition: filter_design.txt:212
FILTER_INPUTS
#define FILTER_INPUTS(array)
Definition: filters.h:264
AVFrame
This structure describes decoded (raw) audio or video data.
Definition: frame.h:427
pixdesc.h
BLRContext::vsub
int vsub
Definition: vf_blurdetect.c:50
edge_common.h
calculate_blur
static float calculate_blur(BLRContext *s, int w, int h, int hsub, int vsub, int8_t *dir, int dir_linesize, uint8_t *dst, int dst_linesize, uint8_t *src, int src_linesize)
Definition: vf_blurdetect.c:193
AVOption
AVOption.
Definition: opt.h:429
b
#define b
Definition: input.c:42
blurdetect_init
static av_cold int blurdetect_init(AVFilterContext *ctx)
Definition: vf_blurdetect.c:86
high
int high
Definition: dovi_rpuenc.c:39
AV_LOG_VERBOSE
#define AV_LOG_VERBOSE
Detailed information.
Definition: log.h:226
AV_PIX_FMT_YUV440P
@ AV_PIX_FMT_YUV440P
planar YUV 4:4:0 (1 Cr & Cb sample per 1x2 Y samples)
Definition: pixfmt.h:106
AVDictionary
Definition: dict.c:32
AVFilter::name
const char * name
Filter name.
Definition: avfilter.h:220
BLRContext::block_height
int block_height
Definition: vf_blurdetect.c:58
DIRECTION_VERTICAL
@ DIRECTION_VERTICAL
Definition: edge_common.h:36
video.h
pix_fmts
static enum AVPixelFormat pix_fmts[]
Definition: vf_blurdetect.c:339
AVFrame::data
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
Definition: frame.h:448
hsub
static void hsub(htype *dst, const htype *src, int bins)
Definition: vf_median.c:74
av_pix_fmt_count_planes
int av_pix_fmt_count_planes(enum AVPixelFormat pix_fmt)
Definition: pixdesc.c:3496
BLRContext::low
float low
Definition: vf_blurdetect.c:53
BLRContext::planes
int planes
Definition: vf_blurdetect.c:59
AV_PIX_FMT_GBRAP
@ AV_PIX_FMT_GBRAP
planar GBRA 4:4:4:4 32bpp
Definition: pixfmt.h:212
FLAGS
#define FLAGS
Definition: vf_blurdetect.c:72
set_meta
static void set_meta(AVDictionary **metadata, const char *key, float d)
Definition: vf_blurdetect.c:248
AVFilterPad
A filter pad used for either input or output.
Definition: filters.h:40
FFDIFFSIGN
#define FFDIFFSIGN(x, y)
Comparator.
Definition: macros.h:45
AV_PIX_FMT_YUVJ411P
@ AV_PIX_FMT_YUVJ411P
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples) full scale (JPEG), deprecated in favor ...
Definition: pixfmt.h:283
ceil
static __device__ float ceil(float a)
Definition: cuda_runtime.h:176
av_cold
#define av_cold
Definition: attributes.h:106
BLRContext::tmpbuf
uint8_t * tmpbuf
Definition: vf_blurdetect.c:66
BLRContext
Definition: vf_blurdetect.c:47
DIRECTION_45DOWN
@ DIRECTION_45DOWN
Definition: edge_common.h:34
ff_video_default_filterpad
const AVFilterPad ff_video_default_filterpad[1]
An AVFilterPad array whose only entry has name "default" and is of type AVMEDIA_TYPE_VIDEO.
Definition: video.c:37
FFFilter
Definition: filters.h:267
AV_PIX_FMT_YUVJ422P
@ AV_PIX_FMT_YUVJ422P
planar YUV 4:2:2, 16bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV422P and setting col...
Definition: pixfmt.h:86
s
#define s(width, name)
Definition: cbs_vp9.c:198
AV_PIX_FMT_YUVA420P
@ AV_PIX_FMT_YUVA420P
planar YUV 4:2:0, 20bpp, (1 Cr & Cb sample per 2x2 Y & A samples)
Definition: pixfmt.h:108
AV_CEIL_RSHIFT
#define AV_CEIL_RSHIFT(a, b)
Definition: common.h:60
BLRContext::block_width
int block_width
Definition: vf_blurdetect.c:57
filters.h
ctx
static AVFormatContext * ctx
Definition: movenc.c:49
AVPixFmtDescriptor::log2_chroma_w
uint8_t log2_chroma_w
Amount to shift the luma width right to find the chroma width.
Definition: pixdesc.h:80
AV_PIX_FMT_YUV420P
@ AV_PIX_FMT_YUV420P
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
Definition: pixfmt.h:73
key
const char * key
Definition: hwcontext_opencl.c:189
FILTER_OUTPUTS
#define FILTER_OUTPUTS(array)
Definition: filters.h:265
comp
static int comp(const float *a, const float *b)
Definition: vf_blurdetect.c:42
tmp
static uint8_t tmp[40]
Definition: aes_ctr.c:52
AV_PIX_FMT_YUVJ444P
@ AV_PIX_FMT_YUVJ444P
planar YUV 4:4:4, 24bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV444P and setting col...
Definition: pixfmt.h:87
BLRContext::radius
int radius
Definition: vf_blurdetect.c:55
OFFSET
#define OFFSET(x)
Definition: vf_blurdetect.c:71
AVClass
Describe the class of an AVClass context structure.
Definition: log.h:76
metadata
Stream codec metadata
Definition: ogg-flac-chained-meta.txt:2
NULL
#define NULL
Definition: coverity.c:32
AV_PIX_FMT_YUVJ420P
@ AV_PIX_FMT_YUVJ420P
planar YUV 4:2:0, 12bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV420P and setting col...
Definition: pixfmt.h:85
DIRECTION_HORIZONTAL
@ DIRECTION_HORIZONTAL
Definition: edge_common.h:35
planes
static const struct @562 planes[]
ff_non_maximum_suppression
void ff_non_maximum_suppression(int w, int h, uint8_t *dst, int dst_linesize, const int8_t *dir, int dir_linesize, const uint16_t *src, int src_linesize)
Filters rounded gradients to drop all non-maxima pixels in the magnitude image Expects gradients gene...
Definition: edge_common.c:60
blurdetect_uninit
static av_cold void blurdetect_uninit(AVFilterContext *ctx)
Definition: vf_blurdetect.c:323
AV_PIX_FMT_GRAY8
@ AV_PIX_FMT_GRAY8
Y , 8bpp.
Definition: pixfmt.h:81
ff_filter_link
static FilterLink * ff_filter_link(AVFilterLink *link)
Definition: filters.h:199
qsort.h
blur
static void blur(uint8_t *dst, int dst_step, const uint8_t *src, int src_step, int len, int radius, int pixsize)
Definition: vf_boxblur.c:162
init
int(* init)(AVBSFContext *ctx)
Definition: dts2pts.c:550
NULL_IF_CONFIG_SMALL
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
Definition: internal.h:94
ff_double_threshold
void ff_double_threshold(int low, int high, int w, int h, uint8_t *dst, int dst_linesize, const uint8_t *src, int src_linesize)
Filters all pixels in src to keep all pixels > high, and keep all pixels > low where all surrounding ...
Definition: edge_common.c:89
dst
uint8_t ptrdiff_t const uint8_t ptrdiff_t int intptr_t intptr_t int int16_t * dst
Definition: dsp.h:87
i
#define i(width, name, range_min, range_max)
Definition: cbs_h264.c:63
BLRContext::blks
float * blks
Definition: vf_blurdetect.c:64
a
The reader does not expect b to be semantically here and if the code is changed by maybe adding a a division or other the signedness will almost certainly be mistaken To avoid this confusion a new type was SUINT is the C unsigned type but it holds a signed int to use the same example SUINT a
Definition: undefined.txt:41
AV_PIX_FMT_YUVA444P
@ AV_PIX_FMT_YUVA444P
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
Definition: pixfmt.h:174
BLRContext::low_u8
uint8_t low_u8
Definition: vf_blurdetect.c:54
BLRContext::nb_planes
int nb_planes
Definition: vf_blurdetect.c:51
AV_LOG_INFO
#define AV_LOG_INFO
Standard information.
Definition: log.h:221
AV_OPT_TYPE_FLOAT
@ AV_OPT_TYPE_FLOAT
Underlying C type is float.
Definition: opt.h:271
BLRContext::blur_total
double blur_total
Definition: vf_blurdetect.c:61
uninit
static void uninit(AVBSFContext *ctx)
Definition: pcm_rechunk.c:68
AV_QSORT
#define AV_QSORT(p, num, type, cmp)
Quicksort This sort is fast, and fully inplace but not stable and it is possible to construct input t...
Definition: qsort.h:33
BLRContext::nb_frames
uint64_t nb_frames
Definition: vf_blurdetect.c:62
value
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf default value
Definition: writing_filters.txt:86
AV_PIX_FMT_YUVJ440P
@ AV_PIX_FMT_YUVJ440P
planar YUV 4:4:0 full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV440P and setting color_range
Definition: pixfmt.h:107
ff_vf_blurdetect
const FFFilter ff_vf_blurdetect
Definition: vf_blurdetect.c:360
AVFilterPad::name
const char * name
Pad name.
Definition: filters.h:46
blurdetect_options
static const AVOption blurdetect_options[]
Definition: vf_blurdetect.c:73
av_calloc
void * av_calloc(size_t nmemb, size_t size)
Definition: mem.c:264
edge_width
static float edge_width(BLRContext *blr, int i, int j, int8_t dir, int w, int h, int edge, const uint8_t *src, int src_linesize)
Definition: vf_blurdetect.c:127
BLRContext::block_pct
int block_pct
Definition: vf_blurdetect.c:56
av_malloc
void * av_malloc(size_t size)
Allocate a memory block with alignment suitable for all memory accesses (including vectors if availab...
Definition: mem.c:98
blurdetect_inputs
static const AVFilterPad blurdetect_inputs[]
Definition: vf_blurdetect.c:351
blurdetect_filter_frame
static int blurdetect_filter_frame(AVFilterLink *inlink, AVFrame *in)
Definition: vf_blurdetect.c:255
AVFILTER_DEFINE_CLASS
AVFILTER_DEFINE_CLASS(blurdetect)
BLRContext::high
float high
Definition: vf_blurdetect.c:53
AV_PIX_FMT_NONE
@ AV_PIX_FMT_NONE
Definition: pixfmt.h:72
AV_OPT_TYPE_INT
@ AV_OPT_TYPE_INT
Underlying C type is int.
Definition: opt.h:259
AVFrame::metadata
AVDictionary * metadata
metadata.
Definition: frame.h:705
BLRContext::filterbuf
uint8_t * filterbuf
Definition: vf_blurdetect.c:65
AVFILTER_FLAG_METADATA_ONLY
#define AVFILTER_FLAG_METADATA_ONLY
The filter is a "metadata" filter - it does not modify the frame data in any way.
Definition: avfilter.h:183
BLRContext::directions
int8_t * directions
Definition: vf_blurdetect.c:68
AV_PIX_FMT_YUV444P
@ AV_PIX_FMT_YUV444P
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
Definition: pixfmt.h:78
AVFilterContext
An instance of a filter.
Definition: avfilter.h:274
AV_PIX_FMT_GBRP
@ AV_PIX_FMT_GBRP
planar GBR 4:4:4 24bpp
Definition: pixfmt.h:165
AVMEDIA_TYPE_VIDEO
@ AVMEDIA_TYPE_VIDEO
Definition: avutil.h:200
FFFilter::p
AVFilter p
The public AVFilter.
Definition: filters.h:271
AV_PIX_FMT_YUV422P
@ AV_PIX_FMT_YUV422P
planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples)
Definition: pixfmt.h:77
mem.h
AVPixFmtDescriptor
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
Definition: pixdesc.h:69
w
uint8_t w
Definition: llvidencdsp.c:39
DIRECTION_45UP
@ DIRECTION_45UP
Definition: edge_common.h:33
av_freep
#define av_freep(p)
Definition: tableprint_vlc.h:35
av_dict_set
int av_dict_set(AVDictionary **pm, const char *key, const char *value, int flags)
Set the given entry in *pm, overwriting an existing entry.
Definition: dict.c:86
AV_PIX_FMT_YUV411P
@ AV_PIX_FMT_YUV411P
planar YUV 4:1:1, 12bpp, (1 Cr & Cb sample per 4x1 Y samples)
Definition: pixfmt.h:80
blurdetect_config_input
static int blurdetect_config_input(AVFilterLink *inlink)
Definition: vf_blurdetect.c:96
AVFrame::linesize
int linesize[AV_NUM_DATA_POINTERS]
For video, a positive or negative value, which is typically indicating the size in bytes of each pict...
Definition: frame.h:472
AV_PIX_FMT_YUV410P
@ AV_PIX_FMT_YUV410P
planar YUV 4:1:0, 9bpp, (1 Cr & Cb sample per 4x4 Y samples)
Definition: pixfmt.h:79
av_log
#define av_log(a,...)
Definition: tableprint_vlc.h:27
h
h
Definition: vp9dsp_template.c:2070
width
#define width
Definition: dsp.h:89
snprintf
#define snprintf
Definition: snprintf.h:34
AVPixFmtDescriptor::log2_chroma_h
uint8_t log2_chroma_h
Amount to shift the luma height right to find the chroma height.
Definition: pixdesc.h:89
src
#define src
Definition: vp8dsp.c:248
AV_PIX_FMT_YUVA422P
@ AV_PIX_FMT_YUVA422P
planar YUV 4:2:2 24bpp, (1 Cr & Cb sample per 2x1 Y & A samples)
Definition: pixfmt.h:173