Bring blazing-fast, ultra-compact neural inference to Ambiq’s family of ultra-low-power SoCs.
📖 Overview
HeliaAOT is an ahead-of-time compiler that transforms trained models (e.g. LiteRT) into highly optimized, standalone C inference modules tailored specifically for Ambiq’s ultra low-power System-on-Chips (SoCs). By shifting neural network inference code generation to build time, HeliaAOT delivers compact, efficient, and readable C output with zero runtime overhead. No more guessing arena sizes or packaging dead code!
🚀 Key Advantages
- 10× smaller code size Dramatically reduce your flash footprint versus vanilla TFLM.
- Expanded kernel library Get best-in-class performance with more operators and fused layers.
- Ambiq-tuned efficiency Optimized code generation for Cortex-M4 & M55 (Helium).
- Highly customizable Fine-tune at the operator, tensor, or graph level.
- Zero arena-size guessing Automatic memory planning for tensors—no more over-allocating.
- Readable, structured C Your generated code mirrors your network topology for easy debugging.
- Seamless integration Supports integration as a CMake library, Zephyr module, or neuralSPOT module.
- Advanced optimizations Layer fusion, tensor reordering, and memory-type placement out of the box.
📚 Quick Links
-
Install HeliaAOT and getting up and running in minutes. Install HeliaAOT
-
How-To examples demonstrating specific commands and features. How-To...
-
Reference for available commands, flags, and configuration options. Reference
-
Usage Examples showcasing real-world applications and best practices. Guides
-
API Documentation for the core HeliaAOT classes and functions. API Documentation
-
Performance Benchmarks comparing HeliaAOT to other frameworks. Benchmarks
🔧 Install
Using pipx is the recommended way to install helia-aot as it allows installing and running python apps in isolated environments. Upon installation, helia-aot will be available in your PATH.
Grab the app for your OS from the latest release page. No Python, no dependencies.
Supported OS: macOS – Linux – Windows
Checkout Installation Guide for alternative installation methods including installing via pip or building from source.
🎬 First Run
$ helia-aot convert --model.path ./ad01-int8.tflite
INFO Started AOT conversion
INFO Step: Loading model ad01_int8.tflite…
INFO Step completed in 0.01s ✔
INFO Step: Applying graph-level transforms…
INFO Step completed in 0.00s ✔
INFO Step: Resolving handlers…
INFO Step completed in 0.00s ✔
INFO Step: Memory planning via greedy…
INFO Step completed in 0.00s ✔
INFO Step: Emitting code as neuralspot module…
INFO Step completed in 0.10s ✔
INFO Step: Exporting module…
INFO Step completed in 0.01s ✔
INFO AOT conversion completed
This generates a complete C inference module with following files:
module_name
├── LICENSE ← license for this module
├── README.md ← module documentation
├── includes-api/ ← API header files
├── src/ ← C operator source files
└── module.mk ← For neuralSPOT integration
📜 License
HeliaAOT generates C modules that include a license header restricting use to Ambiq hardware. See the generated module’s LICENSE for terms and the repository’s LICENSE.md for tooling.
Ready to dive in? Head over to the Getting Started guide and generate your first module in minutes.

