Can we Learn on the Edge? Al systems need to continually adapt to new data collected from the sensors Not only inference, but also run back-propagation on edge devices
3
#Activation is the Memory Bottleneck, not #Trainable Parameters
4
Related Work: Parameter-Efficient Transfer Learning
5
Address Optimization Difficulty of Quantized Graphs
6
QAS: Quantization-Aware Scaling
7
Sparse Layer/Tensor Update
8
Find Layers to Update by Contribution Analysis
9
Tiny Training Engine (TTE)
10
Tiny Training Engine Workflow
11
Deep Gradient Compression: Reduce Bandwidth
12
Post-training testing (high accuracy)
13
Real-life testing
Description:
Explore on-device learning techniques for IoT devices with limited memory in this tinyML Forum talk by MIT EECS Assistant Professor Song Han. Discover an algorithm-system co-design framework enabling neural network fine-tuning with only 256KB of memory. Learn about quantization-aware scaling to stabilize quantized training and sparse update techniques to reduce memory footprint. Understand the implementation of Tiny Training Engine, a lightweight system that prunes backward computation graphs and offloads runtime auto-differentiation. Gain insights into practical solutions for on-device transfer learning in visual recognition tasks on microcontrollers, using significantly less memory than existing frameworks while matching cloud training accuracy. Delve into topics such as edge learning challenges, memory bottlenecks, parameter-efficient transfer learning, and deep gradient compression. See how this approach enables continuous adaptation and lifelong learning for tiny IoT devices, demonstrating their capability to perform both inference and on-device learning.
Read more
On-Device Learning Under 256KB Memory - Challenges and Solutions for IoT Devices