Code examples and project tutorials to build intelligent devices with Coral

Coral examples

Simple code examples showing how to run pre-trained models on your Coral device. More pre-trained models are on our Models page.


Image recognition with video

Multiple examples showing how to stream images from a camera and run classification or detection models with the TensorFlow Lite API. Each example uses a different camera library, such as GStreamer, OpenCV, PyGame, and PiCamera.


Person segmentation with video

This example takes in a camera feed and performs body-part segmentation using the BodyPix model (with both MobileNet v1 and ResNet50 backbones). In addition to identifying different body parts, it can anonymize people from images.


Object tracking with video

This example takes a camera feed and tracks each uniquely identified object, assigning each object with a persistent ID. The example detection script allows you to specify the tracker program you want to use (the Sort tracker is included).


Pose estimation with video

Multiple examples showing how to use the PoseNet model to detect human poses from images and video, such as locating the position of someone’s elbow, shoulder or foot.

Semantic segmentation

This example performs semantic segmentation on an image. It takes an image as input and creates a new version of that image showing which pixels correspond to each recognized object.

Keyphrase detector

A few examples using a keyphrase detection model that can detect over 140 short phrases such as "start game" and "next song." Includes a snake game and a YouTube player that respond to voice commands.

Basic object detection

An example that performs object detection with a photo and draws a square around each object. Also works with face detection models.

Basic image classification

An example that performs image classification with a photo.

Partner examples

More examples that use ML tools from our partners.

Image classification with PerceptiLabs

This tutorial shows you how to take a trained model built in PerceptiLabs and export it for compatibility with Coral.

Sound categorization with balenaCloud

This tutorial teaches you how to deploy a Coral Dev Board with a pre-trained sound categorization model, and use a balenaCloud backend to manually review the classifications.

Fleet management with balenaCloud

This example uses balenaCloud to deploy an object detection model to a Dev Board and view live inferencing from a web page.

Project tutorials

Detailed instructions to help you bring local AI into the real world.


An open source AI experiment that introduces the basics of machine learning by helping you build a teachable object using a Raspberry Pi Zero and the Coral USB Accelerator.

Teachable Sorter

A physical machine that you can teach to rapidly recognize and sort objects using your own custom machine learning models.

Smart Bird Feeder

A smart bird feeder that uses an image classification model to identify birds, record animal visits, and deter squirrels from stealing bird seed.

Embedded Teachable Machine

A machine that can quickly learn to recognize new objects by re-training a vision classification model directly on your device.


An implementation of AlphaGo Zero called Minigo, which uses machine learning to play the strategy board game "go" at expert levels.