Free Online Training Events
Free Technical Resources
3 days (In-Person) 8 hours per day or
4 sessions (Live Online) 6 hours per session
Essential Edge AI for Embedded Developers is designed for engineers who need a practical understanding of how to fine-tune and deploy AI/ML models to constrained edge devices.
This course follows a systematic MLOps workflow, starting with a review of the foundation models, data planning & feature pre-processing, application of transfer learning, followed by deployment of these models and in-field monitoring for further tuning.
Using basic principles and procedures, along with important rules and helpful tricks, the course enables attendees to appreciate a systems perspective of embedding a deep learning model inference into an application. This includes interfacing the model to inputs and outputs of the system to make it useful.
The course considers the application of neural networks, including explainable classical ML algorithms, to a wide variety of constrained edge devices; from microcontrollers, single-board computers through to neural network accelerators.
The practical side of the course demonstrates model fine-tuning, data preprocessing, model-format conversion, model inference, object detection, model acceleration, and model deployment as a Docker container. Exercises related to model creation and preprocessing use Python. Model inference exercises use ONNX and TFLite (now LiteRT) runtimes and support both Python and C++ APIs. These exercises comprise approximately 50% of class time.
Embedded Developers, System Architects and Product Managers, who plan to deploy an AI/ML model for an application in one or more of the following constrained edge device types: 64-bit Linux based Single Board Computers (SBCs) or 32-bit microcontrollers with or without Neural Network Accelerators.
Please note that the course does not delve into details of different types of neural network architecture. These details are covered in the Practical Deep Learning course. Inference of large Generative AI models are not part of the course, as they are more likely to run on servers or workstations.
Attendees should be familiar with the basic idea of an embedded AI application. Specifically, you should have:
Please contact Doulos directly to discuss and assess your specific experience against the pre-requisites.
Doulos training materials are renowned for being the most comprehensive and user-friendly available. Their style, content and coverage are unique in the training world and have made them sought after resources. The materials include:
Supervised learning • Inference engines at the edge • Edge AI components • Edge AI applications
Practicals: Try out model training environments based on CPU and GPU
Model training • Loss functions • Sequential and Functional API use and examples • Frequently used Keras APIs
CRISP-DM Methodology & MLOps for Edge AI • Use case of audio feature pre-processing • Split audio file for training • Conversion of audio segments to Spectrogram • Training data preparation
Practicals: Use SoX (Sound eXchange) software to trim, filter, and plot Spectrograms • View Spectrograms before training • Organize spectrograms into folders for training • Work with a subset of Environment Sound Classification (ESC) dataset
Review of different kinds of neural networks • Convolutional Neural Network (CNN) using 2D and 1D convolution • Transfer Learning • Chaining NN models
Practicals: Code (using Keras) simple Convolutional neural networks using small datasets • Write code to use pretrained models (such as MobileNet) for transfer learning
Constrained Inference platforms • Inference server • Linux based SBC • 32 bit Microcontroller • NN Accelerator • GPU • FPGA
Practicals: Take picture, play and record audio using SBC • Read sensor data using serial port from SBC and Microcontroller • Structure accelerometer data for supervised model training
Model Formats • Open Neural Network Exchange Format (ONNX) • TensorFlow Lite (TFLite) • Viewing Model Graph • Model Format Conversion • Model Quantization
Practicals: Train and convert Scikit-Learn (ML) model to ONNX • Train and convert Keras Models to ONNX and TFLite
Inferencing steps • Input and Output Tensor Shape • ONNX runtime, TFLite Interpreter Python methods • TFLite C++ Classes
Practicals: Convert Scikit-learn model to ONNX format. Perform Inference of NN model using ONNX/TFLite formats in Python environment, Inference of TFLite and ONNX models using C++ classes
TinyML Implementation frameworks • TFLite Micro • Converting Keras Model • Setting up and using TFLite Micro Interpreter
Practicals: Quantize Fully Connected TFLite Model • Convert Model to C array for storing in MPU • Read sensor value and execute model using TinyML (TFLite Micro) on Cortex-M4 device
Neural Network Accelerator platforms • Model compiler and workload partitioning • Executing Model on Accelerator • Working with multiple accelerators
Practicals: Quantize and compile CNN Model for NN Accelerator • Run compiled model on accelerator and compare execution time with SBC
MobileNet architecture •Object Detection using MobileNet based SSD • Object Detector Output • Decoding and boxing detected output • Object Detection using You Only Look Once (YOLO) model
Practicals: Use object detection models to locate objects of interest. Employ post-processing on model output for Non-Maximum Suppression (NMS)
Share Model Inference using MQTT and REST API • Monitor Model Drift
Practicals: Communicate model inference output using MQTT and Flask webserver
Enumerate Model Inference Dependencies • Create Dockerfile for Model Inference Container • Push Model Inference Container to Edge Device
Practicals: Create ONNX/TFLite inference Dockerfiles • Use Dockerfile to create Docker container and test it on Edge device
Edge AI application planning template • Discussion on creating Edge AI solution for different use cases
Classical computer vision algorithms • OpenCV pre-processing and post-processing functions • AI Model inferencing using OpenCV
C++ application compilation • Use of CMake to build application • Use of precompiled shared object (.so) files of model runtime and OpenCV
Terminology – univariate, multivariate, regression, classification • Pre-processing using Pandas • Time Series as supervised learning problem • Keras Time Series Generator API • Time Series classification using CNN • Time Series Database
Practicals: Set up time series dataset as supervised learning task • Perform CNN based fingerprint analysis of sections of time series data
For on-site, team-based training, please contact Doulos about tailoring this course to suit your particular target hardware and software environments.
25 Aug 2025 | ONLINE Americas | Enquire |
08 Dec 2025 | ONLINE EurAsia | Enquire |
08 Dec 2025 | Munich, DE | Enquire |
15 Dec 2025 | ONLINE Americas | Enquire |
Complete an enquiry form and a Doulos representative will get back to you.
Enquiry FormPrice on request