FFmpeg
dnn_backend_common.h
Go to the documentation of this file.
1 /*
2  * This file is part of FFmpeg.
3  *
4  * FFmpeg is free software; you can redistribute it and/or
5  * modify it under the terms of the GNU Lesser General Public
6  * License as published by the Free Software Foundation; either
7  * version 2.1 of the License, or (at your option) any later version.
8  *
9  * FFmpeg is distributed in the hope that it will be useful,
10  * but WITHOUT ANY WARRANTY; without even the implied warranty of
11  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
12  * Lesser General Public License for more details.
13  *
14  * You should have received a copy of the GNU Lesser General Public
15  * License along with FFmpeg; if not, write to the Free Software
16  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
17  */
18 
19 /**
20  * @file
21  * DNN common functions different backends.
22  */
23 
24 #ifndef AVFILTER_DNN_DNN_BACKEND_COMMON_H
25 #define AVFILTER_DNN_DNN_BACKEND_COMMON_H
26 
27 #include "queue.h"
28 #include "../dnn_interface.h"
29 #include "libavutil/thread.h"
30 
31 #define DNN_DEFINE_CLASS_EXT(name, desc, options) \
32  { \
33  .class_name = desc, \
34  .item_name = av_default_item_name, \
35  .option = options, \
36  .version = LIBAVUTIL_VERSION_INT, \
37  .category = AV_CLASS_CATEGORY_FILTER, \
38  }
39 #define DNN_DEFINE_CLASS(fname) \
40  DNN_DEFINE_CLASS_EXT(fname, #fname, fname##_options)
41 
42 // one task for one function call from dnn interface
43 typedef struct TaskItem {
44  void *model; // model for the backend
47  const char *input_name;
48  const char **output_names;
49  uint8_t async;
50  uint8_t do_ioproc;
51  uint32_t nb_output;
52  uint32_t inference_todo;
53  uint32_t inference_done;
54 } TaskItem;
55 
56 // one task might have multiple inferences
57 typedef struct LastLevelTaskItem {
59  uint32_t bbox_index;
61 
62 /**
63  * Common Async Execution Mechanism for the DNN Backends.
64  */
65 typedef struct DNNAsyncExecModule {
66  /**
67  * Synchronous inference function for the backend
68  * with corresponding request item as the argument.
69  */
70  int (*start_inference)(void *request);
71 
72  /**
73  * Completion Callback for the backend.
74  * Expected argument type of callback must match that
75  * of the inference function.
76  */
77  void (*callback)(void *args);
78 
79  /**
80  * Argument for the execution functions.
81  * i.e. Request item for the backend.
82  */
83  void *args;
84 #if HAVE_PTHREAD_CANCEL
85  pthread_t thread_id;
86  pthread_attr_t thread_attr;
87 #endif
89 
90 int ff_check_exec_params(void *ctx, DNNBackendType backend, DNNFunctionType func_type, DNNExecBaseParams *exec_params);
91 
92 /**
93  * Fill the Task for Backend Execution. It should be called after
94  * checking execution parameters using ff_check_exec_params.
95  *
96  * @param task pointer to the allocated task
97  * @param exec_param pointer to execution parameters
98  * @param backend_model void pointer to the backend model
99  * @param async flag for async execution. Must be 0 or 1
100  * @param do_ioproc flag for IO processing. Must be 0 or 1
101  *
102  * @returns 0 if successful or error code otherwise.
103  */
104 int ff_dnn_fill_task(TaskItem *task, DNNExecBaseParams *exec_params, void *backend_model, int async, int do_ioproc);
105 
106 /**
107  * Join the Async Execution thread and set module pointers to NULL.
108  *
109  * @param async_module pointer to DNNAsyncExecModule module
110  *
111  * @returns 0 if successful or error code otherwise.
112  */
114 
115 /**
116  * Start asynchronous inference routine for the TensorFlow
117  * model on a detached thread. It calls the completion callback
118  * after the inference completes. Completion callback and inference
119  * function must be set before calling this function.
120  *
121  * If POSIX threads aren't supported, the execution rolls back
122  * to synchronous mode, calling completion callback after inference.
123  *
124  * @param ctx pointer to the backend context
125  * @param async_module pointer to DNNAsyncExecModule module
126  *
127  * @returns 0 on the start of async inference or error code otherwise.
128  */
129 int ff_dnn_start_inference_async(void *ctx, DNNAsyncExecModule *async_module);
130 
131 /**
132  * Extract input and output frame from the Task Queue after
133  * asynchronous inference.
134  *
135  * @param task_queue pointer to the task queue of the backend
136  * @param in double pointer to the input frame
137  * @param out double pointer to the output frame
138  *
139  * @retval DAST_EMPTY_QUEUE if task queue is empty
140  * @retval DAST_NOT_READY if inference not completed yet.
141  * @retval DAST_SUCCESS if result successfully extracted
142  */
144 
145 /**
146  * Allocate input and output frames and fill the Task
147  * with execution parameters.
148  *
149  * @param task pointer to the allocated task
150  * @param exec_params pointer to execution parameters
151  * @param backend_model void pointer to the backend model
152  * @param input_height height of input frame
153  * @param input_width width of input frame
154  * @param ctx pointer to the backend context
155  *
156  * @returns 0 if successful or error code otherwise.
157  */
158 int ff_dnn_fill_gettingoutput_task(TaskItem *task, DNNExecBaseParams *exec_params, void *backend_model, int input_height, int input_width, void *ctx);
159 
160 #endif
out
FILE * out
Definition: movenc.c:55
thread.h
DNNAsyncExecModule
Common Async Execution Mechanism for the DNN Backends.
Definition: dnn_backend_common.h:65
DNNFunctionType
DNNFunctionType
Definition: dnn_interface.h:52
ff_dnn_start_inference_async
int ff_dnn_start_inference_async(void *ctx, DNNAsyncExecModule *async_module)
Start asynchronous inference routine for the TensorFlow model on a detached thread.
Definition: dnn_backend_common.c:105
LastLevelTaskItem
Definition: dnn_backend_common.h:57
LastLevelTaskItem::bbox_index
uint32_t bbox_index
Definition: dnn_backend_common.h:59
AVFrame
This structure describes decoded (raw) audio or video data.
Definition: frame.h:374
TaskItem
Definition: dnn_backend_common.h:43
DNNAsyncExecModule::callback
void(* callback)(void *args)
Completion Callback for the backend.
Definition: dnn_backend_common.h:77
TaskItem::model
void * model
Definition: dnn_backend_common.h:44
Queue
Linear double-ended data structure.
Definition: queue.c:33
ff_dnn_fill_gettingoutput_task
int ff_dnn_fill_gettingoutput_task(TaskItem *task, DNNExecBaseParams *exec_params, void *backend_model, int input_height, int input_width, void *ctx)
Allocate input and output frames and fill the Task with execution parameters.
Definition: dnn_backend_common.c:156
LastLevelTaskItem::task
TaskItem * task
Definition: dnn_backend_common.h:58
pthread_attr_t
void pthread_attr_t
Definition: os2threads.h:51
ctx
AVFormatContext * ctx
Definition: movenc.c:49
TaskItem::inference_todo
uint32_t inference_todo
Definition: dnn_backend_common.h:52
ff_dnn_fill_task
int ff_dnn_fill_task(TaskItem *task, DNNExecBaseParams *exec_params, void *backend_model, int async, int do_ioproc)
Fill the Task for Backend Execution.
Definition: dnn_backend_common.c:50
TaskItem::in_frame
AVFrame * in_frame
Definition: dnn_backend_common.h:45
ff_dnn_async_module_cleanup
int ff_dnn_async_module_cleanup(DNNAsyncExecModule *async_module)
Join the Async Execution thread and set module pointers to NULL.
Definition: dnn_backend_common.c:86
TaskItem::async
uint8_t async
Definition: dnn_backend_common.h:49
TaskItem::inference_done
uint32_t inference_done
Definition: dnn_backend_common.h:53
DNNBackendType
DNNBackendType
Definition: dnn_interface.h:35
queue.h
pthread_t
Definition: os2threads.h:44
ff_check_exec_params
int ff_check_exec_params(void *ctx, DNNBackendType backend, DNNFunctionType func_type, DNNExecBaseParams *exec_params)
Definition: dnn_backend_common.c:30
ff_dnn_get_result_common
DNNAsyncStatusType ff_dnn_get_result_common(Queue *task_queue, AVFrame **in, AVFrame **out)
Extract input and output frame from the Task Queue after asynchronous inference.
Definition: dnn_backend_common.c:136
DNNAsyncExecModule::start_inference
int(* start_inference)(void *request)
Synchronous inference function for the backend with corresponding request item as the argument.
Definition: dnn_backend_common.h:70
DNNAsyncExecModule::args
void * args
Argument for the execution functions.
Definition: dnn_backend_common.h:83
TaskItem::output_names
const char ** output_names
Definition: dnn_backend_common.h:48
TaskItem::out_frame
AVFrame * out_frame
Definition: dnn_backend_common.h:46
TaskItem::input_name
const char * input_name
Definition: dnn_backend_common.h:47
DNNExecBaseParams
Definition: dnn_interface.h:76
TaskItem::do_ioproc
uint8_t do_ioproc
Definition: dnn_backend_common.h:50
int
int
Definition: ffmpeg_filter.c:424
DNNAsyncStatusType
DNNAsyncStatusType
Definition: dnn_interface.h:45
TaskItem::nb_output
uint32_t nb_output
Definition: dnn_backend_common.h:51