New Problem: Imbalanced Memory Distribution of CNNS
7
Solving the Imbalance with Patch-based Inference
8
MCUNet-v2 Takeaways
9
Once-for-All Network
10
Problem in Training for Tiny Models
11
NetAug for TinyML
12
Problem: Training Memory is much larger
13
TinyTL: Up to 6.5x Memory Saving without Accuracy Loss
14
Differentiable Augmentation
15
TinyML for LIDAR & Point Cloud
16
Full Stack LIDAR & Point Cloud Processing
17
Takeaways: Coming Back to MCUNets
18
Fundamental Problems in TinyML
19
OmniML "Compress" the Model Before Training
20
OmniML: Enable TinyML for All Vision Tasks
21
Founding Team
Description:
Explore full-stack optimization techniques for diverse edge AI platforms in this tinyML Summit 2022 conference talk. Delve into the challenges of deploying TinyML on resource-constrained devices and discover innovative solutions to improve neural network efficiency. Learn about model compression, neural architecture rebalancing, and new design primitives that address the mismatch between AI models and hardware. Gain insights into deploying real-world AI applications on tiny microcontroller units (MCUs) despite limited memory and compute power. Examine topics such as patch-based inference, MCUNet-v2, Once-for-All Network, NetAug for TinyML, TinyTL, and full-stack LIDAR and point cloud processing. Understand the fundamental problems in TinyML and explore how OmniML's approach of compressing models before training enables TinyML for various vision tasks.
TinyML for All: Full-stack Optimization for Diverse Edge AI Platforms