Course Information
Course Overview
How to train Open Source LLMs with LoRA QLoRA, DPO and ORPO.
Unlock the full potential of Large Language Models (LLMs) with this comprehensive course designed for developers and data scientists eager to master advanced training and optimization techniques.
I'll cover everything from A to Z, helping developers understand how LLMs works and data scientists learn simple and advance training techniques.
Starting with the fundamentals of language models and the transformative power of the Transformer architecture, you'll set up your development environment and train your first model from scratch.
Dive deep into cutting-edge fine-tuning methods like LoRA, QLoRA, and DoRA to enhance model performance efficiently. Learn how to improve LLM robustness against noisy data using techniques like Flash Attention and NEFTune, and gain practical experience through hands-on coding sessions.
The course also explores aligning LLMs to human preferences using advanced methods such as Direct Preference Optimization (DPO), KTO, and ORPO. You'll implement these techniques to ensure your models not only perform well but also align with user expectations and ethical standards.
Finally, accelerate your LLM training with multi-GPU setups, model parallelism, Fully Sharded Data Parallel (FSDP) training, and the Unsloth framework to boost speed and reduce VRAM usage. By the end of this course, you'll have a good understanding and practical experience to train, fine-tune, and optimize robust open-source LLMs.
For any problem or request please use this email to communicate with me: gal@apriori.ai
Happy learning!
Course Content
- 7 section(s)
- 28 lecture(s)
- Section 1 What is a Language Model and how training pipeline looks like
- Section 2 Course Materials
- Section 3 Setup your environment and train you first Language Model
- Section 4 Fine tuning LLMs with supervised fine-tune (LoRA, QLoRA, DoRA)
- Section 5 Improve LLM performance and make training Robust to noisy data
- Section 6 Align LLMs to human preference using DPO, KTO and ORPO
- Section 7 Accelerate LLM Training
What You’ll Learn
- What is language model and how the training pipeline looks like
- Fine tuning LLMs with supervised fine-tune (LoRA, QLoRA, DoRA)
- Align LLMs to human preference using DPO, KTO and ORPO
- Accelerate LLM training with multiple GPUs training and Unsloth library
Skills covered in this course
Reviews
-
MMridul Jain
too many gaps, course does not cover practical aspects, the last sections especially with KTO, ORPO are rushed and incomplete. The course does not do justice to the subject at all.
-
LLem Dahlia
so simple and clear explanation of the transformer. never have seen anything like this lora and qlora codes is very practical. exactly what you need in practice. Thanks!
-
KKulwant Bhatia
Covers new frontiers in training LLMs
-
DDan Buffington
This is a good course because it covers a lot of ground in under three hours. It's nice to get a tour of what's going on and I feel like I can ask useful questions after learning from this course.