forked from zephyrproject-rtos/zephyr
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathgna.h
212 lines (188 loc) · 5.35 KB
/
gna.h
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
/*
* Copyright (c) 2018 Intel Corporation.
*
* Author: Sathish Kuttan <sathish.k.kuttan@intel.com>
*
* SPDX-License-Identifier: Apache-2.0
*/
/**
* @file
* @brief Public API header file for Intel GNA driver
*
* This file contains the driver APIs for Intel's
* Gaussian Mixture Model and Neural Network Accelerator (GNA)
*/
#ifndef __INCLUDE_GNA__
#define __INCLUDE_GNA__
#ifdef __cplusplus
extern "C" {
#endif
/**
* GNA driver configuration structure.
* Currently empty.
*/
struct gna_config {
};
/**
* GNA Neural Network model header
* Describes the key parameters of the neural network model
*/
struct gna_model_header {
u32_t labase_offset;
u32_t model_size;
u32_t gna_mode;
u32_t layer_count;
u32_t bytes_per_input;
u32_t bytes_per_output;
u32_t num_input_nodes;
u32_t num_output_nodes;
u32_t input_ptr_offset;
u32_t output_ptr_offset;
u32_t rw_region_size;
u32_t input_scaling_factor;
u32_t output_scaling_factor;
};
/**
* GNA Neural Network model information to be provided by application
* during model registration
*/
struct gna_model_info {
struct gna_model_header *header;
void *rw_region;
void *ro_region;
};
/**
* Request to perform inference on the given neural network model
*/
struct gna_inference_req {
void *model_handle;
void *input;
void *output;
void *intermediate;
};
/**
* Statistics of the inference operation returned after completion
*/
struct gna_inference_stats {
u32_t total_cycles;
u32_t stall_cycles;
u32_t cycles_per_sec;
};
/**
* Result of an inference operation
*/
enum gna_result {
GNA_RESULT_INFERENCE_COMPLETE,
GNA_RESULT_SATURATION_OCCURRED,
GNA_RESULT_OUTPUT_BUFFER_FULL_ERROR,
GNA_RESULT_PARAM_OUT_OF_RANGE_ERROR,
GNA_RESULT_GENERIC_ERROR,
};
/**
* Structure containing a response to the inference request
*/
struct gna_inference_resp {
enum gna_result result;
void *output;
size_t output_len;
struct gna_inference_stats stats;
};
/**
* @cond INTERNAL_HIDDEN
*
* Internal documentation. Skip in public documentation
*/
typedef int (*gna_callback)(struct gna_inference_resp *result);
typedef int (*gna_api_config)(struct device *dev, struct gna_config *cfg);
typedef int (*gna_api_register)(struct device *dev,
struct gna_model_info *model, void **model_handle);
typedef int (*gna_api_deregister)(struct device *dev, void *model_handle);
typedef int (*gna_api_infer)(struct device *dev, struct gna_inference_req *req,
gna_callback callback);
struct gna_driver_api {
gna_api_config configure;
gna_api_register register_model;
gna_api_deregister deregister_model;
gna_api_infer infer;
};
/**
* @endcond
*/
/**
* @brief Configure the GNA device.
*
* Configure the GNA device. The GNA device must be configured before
* registering a model or performing inference
*
* @param dev Pointer to the device structure for the driver instance.
* @param cfg Device configuration information
*
* @retval 0 If the configuration is successful
* @retval A negative error code in case of a failure.
*/
static inline int gna_configure(struct device *dev, struct gna_config *cfg)
{
const struct gna_driver_api *api = dev->driver_api;
return api->configure(dev, cfg);
}
/**
* @brief Register a neural network model
*
* Register a neural network model with the GNA device
* A model needs to be registered before it can be used to perform inference
*
* @param dev Pointer to the device structure for the driver instance.
* @param model Information about the neural network model
* @param model_handle Handle to the registered model if registration succeeds
*
* @retval 0 If registration of the model is successful.
* @retval A negative error code in case of a failure.
*/
static inline int gna_register_model(struct device *dev,
struct gna_model_info *model, void **model_handle)
{
const struct gna_driver_api *api = dev->driver_api;
return api->register_model(dev, model, model_handle);
}
/**
* @brief De-register a previously registered neural network model
*
* De-register a previously registered neural network model from the GNA device
* De-registration may be done to free up memory for registering another model
* Once de-registered, the model can no longer be used to perform inference
*
* @param dev Pointer to the device structure for the driver instance.
* @param model Model handle output by gna_register_model API
*
* @retval 0 If de-registration of the model is successful.
* @retval A negative error code in case of a failure.
*/
static inline int gna_deregister_model(struct device *dev, void *model)
{
const struct gna_driver_api *api = dev->driver_api;
return api->deregister_model(dev, model);
}
/**
* @brief Perform inference on a model with input vectors
*
* Make an inference request on a previously registered model with an of
* input data vector
* A callback is provided for notification of inference completion
*
* @param dev Pointer to the device structure for the driver instance.
* @param req Information required to perform inference on a neural network
* @param callback A callback function to notify inference completion
*
* @retval 0 If the request is accepted
* @retval A negative error code in case of a failure.
*/
static inline int gna_infer(struct device *dev, struct gna_inference_req *req,
gna_callback callback)
{
const struct gna_driver_api *api = dev->driver_api;
return api->infer(dev, req, callback);
}
#ifdef __cplusplus
}
#endif
#endif /* __INCLUDE_GNA__ */