gna.h 5.7 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233
  1. /*
  2. * Copyright (c) 2018 Intel Corporation.
  3. *
  4. * Author: Sathish Kuttan <sathish.k.kuttan@intel.com>
  5. *
  6. * SPDX-License-Identifier: Apache-2.0
  7. */
  8. /**
  9. * @file
  10. * @brief Public API header file for Intel GNA driver
  11. */
  12. #ifndef __INCLUDE_GNA__
  13. #define __INCLUDE_GNA__
  14. /**
  15. * @defgroup gna_interface GNA Interface
  16. * @ingroup io_interfaces
  17. * @{
  18. *
  19. * This file contains the driver APIs for Intel's
  20. * Gaussian Mixture Model and Neural Network Accelerator (GNA)
  21. */
  22. #ifdef __cplusplus
  23. extern "C" {
  24. #endif
  25. /**
  26. * GNA driver configuration structure.
  27. * Currently empty.
  28. */
  29. struct gna_config {
  30. };
  31. /**
  32. * GNA Neural Network model header
  33. * Describes the key parameters of the neural network model
  34. */
  35. struct gna_model_header {
  36. uint32_t labase_offset;
  37. uint32_t model_size;
  38. uint32_t gna_mode;
  39. uint32_t layer_count;
  40. uint32_t bytes_per_input;
  41. uint32_t bytes_per_output;
  42. uint32_t num_input_nodes;
  43. uint32_t num_output_nodes;
  44. uint32_t input_ptr_offset;
  45. uint32_t output_ptr_offset;
  46. uint32_t rw_region_size;
  47. uint32_t input_scaling_factor;
  48. uint32_t output_scaling_factor;
  49. };
  50. /**
  51. * GNA Neural Network model information to be provided by application
  52. * during model registration
  53. */
  54. struct gna_model_info {
  55. struct gna_model_header *header;
  56. void *rw_region;
  57. void *ro_region;
  58. };
  59. /**
  60. * Request to perform inference on the given neural network model
  61. */
  62. struct gna_inference_req {
  63. void *model_handle;
  64. void *input;
  65. void *output;
  66. void *intermediate;
  67. };
  68. /**
  69. * Statistics of the inference operation returned after completion
  70. */
  71. struct gna_inference_stats {
  72. uint32_t total_cycles;
  73. uint32_t stall_cycles;
  74. uint32_t cycles_per_sec;
  75. };
  76. /**
  77. * Result of an inference operation
  78. */
  79. enum gna_result {
  80. GNA_RESULT_INFERENCE_COMPLETE,
  81. GNA_RESULT_SATURATION_OCCURRED,
  82. GNA_RESULT_OUTPUT_BUFFER_FULL_ERROR,
  83. GNA_RESULT_PARAM_OUT_OF_RANGE_ERROR,
  84. GNA_RESULT_GENERIC_ERROR,
  85. };
  86. /**
  87. * Structure containing a response to the inference request
  88. */
  89. struct gna_inference_resp {
  90. enum gna_result result;
  91. void *output;
  92. size_t output_len;
  93. struct gna_inference_stats stats;
  94. };
  95. /**
  96. * @cond INTERNAL_HIDDEN
  97. *
  98. * Internal documentation. Skip in public documentation
  99. */
  100. typedef int (*gna_callback)(struct gna_inference_resp *result);
  101. typedef int (*gna_api_config)(const struct device *dev,
  102. struct gna_config *cfg);
  103. typedef int (*gna_api_register)(const struct device *dev,
  104. struct gna_model_info *model,
  105. void **model_handle);
  106. typedef int (*gna_api_deregister)(const struct device *dev,
  107. void *model_handle);
  108. typedef int (*gna_api_infer)(const struct device *dev,
  109. struct gna_inference_req *req,
  110. gna_callback callback);
  111. struct gna_driver_api {
  112. gna_api_config configure;
  113. gna_api_register register_model;
  114. gna_api_deregister deregister_model;
  115. gna_api_infer infer;
  116. };
  117. /**
  118. * @endcond
  119. */
  120. /**
  121. * @brief Configure the GNA device.
  122. *
  123. * Configure the GNA device. The GNA device must be configured before
  124. * registering a model or performing inference
  125. *
  126. * @param dev Pointer to the device structure for the driver instance.
  127. * @param cfg Device configuration information
  128. *
  129. * @retval 0 If the configuration is successful
  130. * @retval A negative error code in case of a failure.
  131. */
  132. static inline int gna_configure(const struct device *dev,
  133. struct gna_config *cfg)
  134. {
  135. const struct gna_driver_api *api =
  136. (const struct gna_driver_api *)dev->api;
  137. return api->configure(dev, cfg);
  138. }
  139. /**
  140. * @brief Register a neural network model
  141. *
  142. * Register a neural network model with the GNA device
  143. * A model needs to be registered before it can be used to perform inference
  144. *
  145. * @param dev Pointer to the device structure for the driver instance.
  146. * @param model Information about the neural network model
  147. * @param model_handle Handle to the registered model if registration succeeds
  148. *
  149. * @retval 0 If registration of the model is successful.
  150. * @retval A negative error code in case of a failure.
  151. */
  152. static inline int gna_register_model(const struct device *dev,
  153. struct gna_model_info *model,
  154. void **model_handle)
  155. {
  156. const struct gna_driver_api *api =
  157. (const struct gna_driver_api *)dev->api;
  158. return api->register_model(dev, model, model_handle);
  159. }
  160. /**
  161. * @brief De-register a previously registered neural network model
  162. *
  163. * De-register a previously registered neural network model from the GNA device
  164. * De-registration may be done to free up memory for registering another model
  165. * Once de-registered, the model can no longer be used to perform inference
  166. *
  167. * @param dev Pointer to the device structure for the driver instance.
  168. * @param model Model handle output by gna_register_model API
  169. *
  170. * @retval 0 If de-registration of the model is successful.
  171. * @retval A negative error code in case of a failure.
  172. */
  173. static inline int gna_deregister_model(const struct device *dev, void *model)
  174. {
  175. const struct gna_driver_api *api =
  176. (const struct gna_driver_api *)dev->api;
  177. return api->deregister_model(dev, model);
  178. }
  179. /**
  180. * @brief Perform inference on a model with input vectors
  181. *
  182. * Make an inference request on a previously registered model with an of
  183. * input data vector
  184. * A callback is provided for notification of inference completion
  185. *
  186. * @param dev Pointer to the device structure for the driver instance.
  187. * @param req Information required to perform inference on a neural network
  188. * @param callback A callback function to notify inference completion
  189. *
  190. * @retval 0 If the request is accepted
  191. * @retval A negative error code in case of a failure.
  192. */
  193. static inline int gna_infer(const struct device *dev,
  194. struct gna_inference_req *req,
  195. gna_callback callback)
  196. {
  197. const struct gna_driver_api *api =
  198. (const struct gna_driver_api *)dev->api;
  199. return api->infer(dev, req, callback);
  200. }
  201. #ifdef __cplusplus
  202. }
  203. #endif
  204. /**
  205. * @}
  206. */
  207. #endif /* __INCLUDE_GNA__ */