Adaptive MCUs Facilitate Multimodal Context-Aware Computing
Intended to accelerate the development of low-power, cognitive IoT devices, Synaptics extended its Astra AI-Native platform with the SR-Series high-performance adaptive microcontrollers for scalable context-aware edge AI. The deliver 100-GOPS performance with a high level of power efficiency that leverages ultra-low-power, always-on technology to deliver intelligence at every power level.
Features include an Arm Cortex-M55 core and the Arm Ethos-U55 neural processing unit. The SR-Series is supported by the Astra Machina Micro development kit and open-source SDK.
>>Check out our coverage of embedded world 2025
The MCUs are optimized for multimodal consumer, enterprise, and industrial IoT workloads with accelerators and adaptive vision, audio, and voice algorithms They offer a range of peripherals that include multiple camera interfaces to help minimize system cost while enabling integration into a wide range of devices. The Synaptics Astra AI-Native compute platform for the IoT combines scalable, low-power compute silicon for the device edge with open-source software, tools, and Veros wireless connectivity.
The SR-Series consists of three MCUs—the SR110, SR105, and SR102—each able to serve a range of multimodal application requirements, with a Cortex-M55 core and Arm Helium technology running up to 400 MHz. Furthermore, the SR110 has a Cortex-M4 core and Arm Ethos-U55 NPU (it's sampling now); the SR105 has an Ethos-U55 NPU; and the SR102 is a single Cortex-M55 device.
Also in the mix are up to 4 MB of system memory, including ULP AON memory; streaming vision and audio processing; MIPI-CSI camera input and passthrough; low-power image signal processing; and secure OTP, TRNG, AES-256, RSA-4096, and SHA-512.
Related links: