Let’s make your AI
tiny and fast.
Try out our inference engine for Arm Cortex-M. It is the fastest and smallest in the world. MobileNetV2 runs with 40% lower latency and uses 49% less RAM. Model accuracy does not change. No binarization, no additional quantization, no pruning.