Главная
Study mode:
on
1
Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed MoE
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Learn about enhancing supernet architecture training through a 35-minute AutoML seminar that introduces the Mixture-of-Supernets formulation. Explore how Mixture-of-Experts (MoE) concepts generate flexible weights for subnetworks, leading to improved Neural Architecture Search (NAS) efficiency and high-quality architectures. Discover the practical applications in constructing efficient BERT and Machine Translation models while meeting user-defined constraints. Join speaker Ganesh Jawahar as he presents this ACL 2024 research that demonstrates significant improvements in retraining time and overall NAS effectiveness.

Mixture-of-Supernets - Improving Weight-Sharing Supernet Training with Architecture-Routed MoE

AutoML Seminars
Add to list
0:00 / 0:00