Главная
Study mode:
on
1
Intro
2
Pulling Data
3
Preprocessing
4
Data Input Pipeline
5
Defining Model
6
Model Training
7
Saving and Loading Models
8
Making Predictions
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Learn how to build a multi-class language classification model using BERT and TensorFlow in this comprehensive 43-minute tutorial. Explore the power of transformers in natural language processing as you work through each step of the process, from data preprocessing to model training and prediction. Follow along with clearly defined chapters for each section, including data input pipeline creation, model definition, and saving/loading techniques. Gain insights into the significance of transformers in deep learning and their dominance in NLP benchmarks. Utilize the HuggingFace transformers library to create an efficient and high-performing solution for multi-class text classification tasks.

Multi-Class Language Classification With BERT in TensorFlow

James Briggs
Add to list
0:00 / 0:00