- Introduction to Dataset building for fine-tuning.
2
- The Reddit dataset options Torrent, Archive.org, BigQuery
3
- Exporting BigQuery Reddit and some other data
4
- Decompressing all of the gzip archives
5
- Re-combining the archives for target subreddits
6
- How to structure the data
7
- Building training samples and saving to database
8
- Creating customized training json files
9
- QLoRA training and results
Description:
Learn how to build a QLoRA fine-tuning dataset for language models in this comprehensive video tutorial. Explore various Reddit dataset options, including torrent files, Archive.org, and BigQuery. Follow step-by-step instructions on exporting BigQuery Reddit data, decompressing gzip archives, and recombining archives for target subreddits. Discover the proper data structure, build training samples, and save them to a database. Create customized training JSON files and dive into QLoRA training and results. Gain valuable insights into dataset building for fine-tuning language models through practical demonstrations and explanations.
Building an LLM Fine-Tuning Dataset - From Reddit Comments to QLoRA Training