Edge AI Gesture Detector
Demonstrates how to perform on-device inference on a Blecon Smart Beacon device
Overview
The Edge AI Gesture Detector example demonstrates how ML workflows can leverage Blecon connectivity. Inference can be performed on a Blecon-enabled device and results logged and transmitted upon connection. Models can then be updated using an OTA firmware update. This application integrates an Edge Impulse-generated model.
Source Code
The source code for this app is provided in the sensor-ml
directory of the blecon-oem-device-firmware
repository: https://github.com/blecon/blecon-oem-device-firmware/tree/main/apps/sensor-ml
Edge Impulse Model
The application integrates the gesture detection ML model from the corresponding Edge Impulse example. Inference results will also be displayed using the device's LEDs if the device's button has been pressed within a configurable period:
Red: Up/Down motion
Green: Wave
Results are reported within log
messages in the ml-sensor
namespace.
Supported Device
This code supports the MOKO SMART L02S-BCN Smart Beacon. The Zephyr board name is mkbnl02sn/nrf54l15/cpuapp
.
Configuration Options
Configurations for each board type is found in the configuration
sub-directory.
Reporting Frequency
REPORTING_PERIOD_SEC
- Configures the maximum frequency with which the device submits logged data to the cloud. If connectivity is not available when the device requests a connection, the device will continue to request
LED mapping
LED mappings are configured in the app.overlay
file for the selected board under the chosen
node. The blecon-led-gesture-1
, blecon-led-gesture-2
, and blecon-led-status
can be assigned to LED device nodes in the devicetree.
Blecon LED Status Library
To display Blecon connection status on LEDs, the BLECON_LED_LIB
option must be enabled. For devices with limited memory, this library can be disabled to reduce the application's memory footprint.
Last updated