123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233 |
- /*
- * Copyright (c) 2018 Intel Corporation.
- *
- * Author: Sathish Kuttan <sathish.k.kuttan@intel.com>
- *
- * SPDX-License-Identifier: Apache-2.0
- */
- /**
- * @file
- * @brief Public API header file for Intel GNA driver
- */
- #ifndef __INCLUDE_GNA__
- #define __INCLUDE_GNA__
- /**
- * @defgroup gna_interface GNA Interface
- * @ingroup io_interfaces
- * @{
- *
- * This file contains the driver APIs for Intel's
- * Gaussian Mixture Model and Neural Network Accelerator (GNA)
- */
- #ifdef __cplusplus
- extern "C" {
- #endif
- /**
- * GNA driver configuration structure.
- * Currently empty.
- */
- struct gna_config {
- };
- /**
- * GNA Neural Network model header
- * Describes the key parameters of the neural network model
- */
- struct gna_model_header {
- uint32_t labase_offset;
- uint32_t model_size;
- uint32_t gna_mode;
- uint32_t layer_count;
- uint32_t bytes_per_input;
- uint32_t bytes_per_output;
- uint32_t num_input_nodes;
- uint32_t num_output_nodes;
- uint32_t input_ptr_offset;
- uint32_t output_ptr_offset;
- uint32_t rw_region_size;
- uint32_t input_scaling_factor;
- uint32_t output_scaling_factor;
- };
- /**
- * GNA Neural Network model information to be provided by application
- * during model registration
- */
- struct gna_model_info {
- struct gna_model_header *header;
- void *rw_region;
- void *ro_region;
- };
- /**
- * Request to perform inference on the given neural network model
- */
- struct gna_inference_req {
- void *model_handle;
- void *input;
- void *output;
- void *intermediate;
- };
- /**
- * Statistics of the inference operation returned after completion
- */
- struct gna_inference_stats {
- uint32_t total_cycles;
- uint32_t stall_cycles;
- uint32_t cycles_per_sec;
- };
- /**
- * Result of an inference operation
- */
- enum gna_result {
- GNA_RESULT_INFERENCE_COMPLETE,
- GNA_RESULT_SATURATION_OCCURRED,
- GNA_RESULT_OUTPUT_BUFFER_FULL_ERROR,
- GNA_RESULT_PARAM_OUT_OF_RANGE_ERROR,
- GNA_RESULT_GENERIC_ERROR,
- };
- /**
- * Structure containing a response to the inference request
- */
- struct gna_inference_resp {
- enum gna_result result;
- void *output;
- size_t output_len;
- struct gna_inference_stats stats;
- };
- /**
- * @cond INTERNAL_HIDDEN
- *
- * Internal documentation. Skip in public documentation
- */
- typedef int (*gna_callback)(struct gna_inference_resp *result);
- typedef int (*gna_api_config)(const struct device *dev,
- struct gna_config *cfg);
- typedef int (*gna_api_register)(const struct device *dev,
- struct gna_model_info *model,
- void **model_handle);
- typedef int (*gna_api_deregister)(const struct device *dev,
- void *model_handle);
- typedef int (*gna_api_infer)(const struct device *dev,
- struct gna_inference_req *req,
- gna_callback callback);
- struct gna_driver_api {
- gna_api_config configure;
- gna_api_register register_model;
- gna_api_deregister deregister_model;
- gna_api_infer infer;
- };
- /**
- * @endcond
- */
- /**
- * @brief Configure the GNA device.
- *
- * Configure the GNA device. The GNA device must be configured before
- * registering a model or performing inference
- *
- * @param dev Pointer to the device structure for the driver instance.
- * @param cfg Device configuration information
- *
- * @retval 0 If the configuration is successful
- * @retval A negative error code in case of a failure.
- */
- static inline int gna_configure(const struct device *dev,
- struct gna_config *cfg)
- {
- const struct gna_driver_api *api =
- (const struct gna_driver_api *)dev->api;
- return api->configure(dev, cfg);
- }
- /**
- * @brief Register a neural network model
- *
- * Register a neural network model with the GNA device
- * A model needs to be registered before it can be used to perform inference
- *
- * @param dev Pointer to the device structure for the driver instance.
- * @param model Information about the neural network model
- * @param model_handle Handle to the registered model if registration succeeds
- *
- * @retval 0 If registration of the model is successful.
- * @retval A negative error code in case of a failure.
- */
- static inline int gna_register_model(const struct device *dev,
- struct gna_model_info *model,
- void **model_handle)
- {
- const struct gna_driver_api *api =
- (const struct gna_driver_api *)dev->api;
- return api->register_model(dev, model, model_handle);
- }
- /**
- * @brief De-register a previously registered neural network model
- *
- * De-register a previously registered neural network model from the GNA device
- * De-registration may be done to free up memory for registering another model
- * Once de-registered, the model can no longer be used to perform inference
- *
- * @param dev Pointer to the device structure for the driver instance.
- * @param model Model handle output by gna_register_model API
- *
- * @retval 0 If de-registration of the model is successful.
- * @retval A negative error code in case of a failure.
- */
- static inline int gna_deregister_model(const struct device *dev, void *model)
- {
- const struct gna_driver_api *api =
- (const struct gna_driver_api *)dev->api;
- return api->deregister_model(dev, model);
- }
- /**
- * @brief Perform inference on a model with input vectors
- *
- * Make an inference request on a previously registered model with an of
- * input data vector
- * A callback is provided for notification of inference completion
- *
- * @param dev Pointer to the device structure for the driver instance.
- * @param req Information required to perform inference on a neural network
- * @param callback A callback function to notify inference completion
- *
- * @retval 0 If the request is accepted
- * @retval A negative error code in case of a failure.
- */
- static inline int gna_infer(const struct device *dev,
- struct gna_inference_req *req,
- gna_callback callback)
- {
- const struct gna_driver_api *api =
- (const struct gna_driver_api *)dev->api;
- return api->infer(dev, req, callback);
- }
- #ifdef __cplusplus
- }
- #endif
- /**
- * @}
- */
- #endif /* __INCLUDE_GNA__ */
|