Skip to content

Cramming Bulky AI Tech into Compact Embedded Devices

CEO Sam Fok from Femtosense discusses AI/ML at the edge and the firm's dual-sparsity processing unit in an interview.

Cramming Bulky AI Technology into Compact Embedded Devices
Cramming Bulky AI Technology into Compact Embedded Devices

Cramming Bulky AI Tech into Compact Embedded Devices

The SPU-001 AI accelerator, developed by Femtosense, is making waves in the world of edge Artificial Intelligence (AI) and Machine Learning (ML) applications. This compact device, despite its small size, offers capabilities typically found in larger AI/ML systems.

At its core, the SPU-001 combines a Neural Processing Unit (NPU) with a Cortex-M0+ MCU for control and management tasks. The NPU, built on a 22 nm process, boasts 1 MB of SRAM, which is effectively expanded to 10 MB using sparsity techniques. This design not only reduces space requirements but also power consumption, a key focus for edge devices.

The SPU-001 leverages dual sparsity optimization, storing and computing only on the weights that matter. This approach results in roughly a 10x improvement in speed, efficiency, and memory consumption. Moreover, it dramatically reduces energy consumption by activating only a fraction of neurons compared to conventional methods.

Femtosense's offering comes with an SDK compatible with major machine learning frameworks like PyTorch, TensorFlow, and JAX. This enables developers to deploy models with minimal barrier, supporting both dense and sparse models. The SDK also provides tools to optimize models via sparsity regularization and quantization-aware training.

The SPU-001 is tailored for edge devices where power efficiency, low latency, and compact memory footprint are critical. Typical applications involve AI/ML inference in constrained settings, such as IoT devices, wearables, industrial sensors, and more.

In summary, the SPU-001 stands out due to its combination of specialized hardware with advanced sparsity-driven optimizations and comprehensive software tooling. This makes it well suited for deploying modern, sparse neural networks in real-world, resource-constrained environments.

For more insights into the SPU-001's future functionality, tune into the 15:04 - Future of Functionality segment of the Inside Electronics Podcast.

Discussing the future functionality of the SPU-001 AI accelerator, listen to the 15:04 - Future of Functionality segment of the Inside Electronics Podcast to learn about its potential integration with podcasts that focus on advancements in technology, particularly artificial-intelligence. The SPU-001's development by Femtosense is already revolutionizing edge Artificial Intelligence (AI) and Machine Learning (ML) applications, and insights into its upcoming capabilities could aid in creating more efficient AI/ML systems in the future.

Read also:

    Latest