Provides a complete system
The Coral System-on-Module (SoM) is a fully-integrated system that helps you build embedded systems that demand fast machine learning (ML) inferencing. It contains NXP's iMX8M system-on-chip (SoC), eMMC memory, LPDDR4 RAM, Wi-Fi, and Bluetooth, but its unique power comes from Google's Edge TPU coprocessor.
Performs high-speed ML inferencing
The on-board Edge TPU coprocessor is capable of performing 4 trillion operations (tera-operations) per second (TOPS), using 0.5 watts for each TOPS (2 TOPS per watt). For example, it can execute state-of-the-art mobile vision models such as MobileNet v2 at 400 FPS, in a power efficient manner. This on-device processing reduces latency, increases data privacy, and removes the need for a high-bandwidth connection used to perform ML inferencing in the cloud.
Integrates with your custom hardware
The SoM connects to your own baseboard hardware with three 100-pin connectors.
Supports TensorFlow Lite
No need to build models from the ground up. TensorFlow Lite models can be compiled to run on the Edge TPU.
Supports AutoML Vision Edge
Easily build and deploy fast, high-accuracy custom image classification models to your device with AutoML Vision Edge.
Also available with a baseboard as part of the Coral Dev Board.
Key benefits of the SoM: