The STMicroelectronics X-LINUX-AI software environment provides tools to perform inference on embedded systems using neural network models. Examples of applications that typically use neural network inference include object/pattern recognition, gesture control, voice processing, and sound monitoring.
X-LINUX-AI includes support for two standard inference engines:
Include X-LINUX-AI packages in Digi Embedded Yocto
Add the meta-st-stm32mpu-ai layer to your conf/bblayers.conf
configuration file if it isn’t there already:
/usr/local/dey-4.0/sources/meta-digi/meta-digi-arm \
/usr/local/dey-4.0/sources/meta-digi/meta-digi-dey \
+ /usr/local/dey-4.0/sources/meta-st-stm32mpu-ai \
"
Edit your conf/local.conf
file to include one of the X-LINUX-AI package groups in your Digi Embedded Yocto image:
IMAGE_INSTALL:append = " packagegroup-x-linux-ai"
There are a few package groups to choose from:
-
packagegroup-x-linux-ai-tflite includes TensorFlow Lite packages and examples, for use with the ConnectCore MP13’s CPU
-
packagegroup-x-linux-ai-tflite-edgetpu includes TensorFlow Lite packages and examples, for use with a Coral USB Accelerator
-
packagegroup-x-linux-ai-onnxruntime includes ONNX Runtime packages and examples
-
packagegroup-x-linux-ai includes all of the above
Including either of these packageg roups increases the size of the rootfs image significantly. To minimize the increase in image size, select a subset of their packages depending on your needs. You might need to remove additional packages for the rootfs image to fit on your device. |
All X-LINUX-AI examples are GUI-based, so they cannot be run on the ConnectCore MP13. |
More information
See ST’s X-LINUX-AI OpenSTLinux Expansion Package article for more information on X-LINUX-AI.