The Volm Edge Compiler

The Volm Edge Compiler converts standard ML models (ONNX, TensorFlow Lite, PyTorch exported to ONNX) into lightweight C code or pre-compiled binaries that can run directly on MCUs (Microcontrollers) or IoT devices.

Why OEM Needs This:

  • Most IoT/MCU devices run on 256 KB – 1 MB SRAM. Traditional ML runtimes (TensorFlow Lite Micro, PyTorch Mobile) are too heavy, adding ~100 KB+ runtime overhead.

  • Volm’s compiler reduces overhead to <10 KB, allowing inference at near bare-metal speed.

  • Supports quantization (INT8, INT4, even binary neural nets), making models fit into ultra-constrained devices.

Example Flow:

  1. An OEM engineer trains a model in PyTorch (e.g., anomaly detection for motor vibration).

  2. They export the model to the ONNX format.

  3. They run the Volm Compiler, which outputs an optimized C library + lightweight Volm Node hooks.

  4. The OEM integrates the compiled code into their device firmware image.

Last updated