Udemy

Fine Tuning LLM with Hugging Face Transformers for NLP

Enroll Now
  • 6,264 Students
  • Updated 11/2025
  • Certificate Available
4.4
(672 Ratings)
CTgoodjobs selects quality courses to enhance professionals' competitiveness. By purchasing courses through links on our site, we may receive an affiliate commission.

Course Information

Registration period
Year-round Recruitment
Course Level
Study Mode
Duration
16 Hour(s) 30 Minute(s)
Language
English
Taught by
KGP Talkie | Laxmi Kant
Certificate
  • Available
  • *The delivery and distribution of the certificate are subject to the policies and arrangements of the course provider.
Rating
4.4
(672 Ratings)

Course Overview

Fine Tuning LLM with Hugging Face Transformers for NLP

Master Transformer models like Phi2, LLAMA; BERT variants, and distillation for advanced NLP applications on custom data

Do not take this course if you are an ML beginner. This course is designed for those who are interested in pure coding and want to fine-tune LLMs instead of focusing on prompt engineering. Otherwise, you may find it difficult to understand.


Welcome to "Mastering Transformer Models and LLM Fine Tuning", a comprehensive and practical course designed for all levels, from beginners to advanced practitioners in Natural Language Processing (NLP). This course delves deep into the world of Transformer models, fine-tuning techniques, and knowledge distillation, with a special focus on popular BERT variants like Phi2, LLAMA, T5, BERT, DistilBERT, MobileBERT, and TinyBERT.

Course Overview:

Section 1: Introduction

  • Get an overview of the course and understand the learning outcomes.

  • Introduction to the resources and code files you will need throughout the course.

Section 2: Understanding Transformers with Hugging Face

  • Learn the fundamentals of Hugging Face Transformers.

  • Explore Hugging Face pipelines, checkpoints, models, and datasets.

  • Gain insights into Hugging Face Spaces and Auto-Classes for seamless model management.

Section 3: Core Concepts of Transformers and LLMs

  • Delve into the architectures and key concepts behind Transformers.

  • Understand the applications of Transformers in various NLP tasks.

  • Introduction to transfer learning with Transformers.

Section 4: BERT Architecture Deep Dive

  • Detailed exploration of BERT's architecture and its importance in context understanding.

  • Learn about Masked Language Modeling (MLM) and Next Sentence Prediction (NSP) in BERT.

  • Understand BERT fine-tuning and evaluation techniques.

Section 5: Practical Fine-Tuning with BERT

  • Hands-on sessions to fine-tune BERT for sentiment classification on Twitter data.

  • Step-by-step guide on data loading, tokenization, and model training.

  • Practical application of fine-tuning techniques to build a BERT classifier.

Section 6: Knowledge Distillation Techniques for BERT

  • Introduction to knowledge distillation and its significance in model optimization.

  • Detailed study of DistilBERT, including loss functions and paper walkthroughs.

  • Explore MobileBERT and TinyBERT, with a focus on their unique distillation techniques and practical implementations.

Section 7: Applying Distilled BERT Models for Real-World Tasks like Fake News Detection

  • Use DistilBERT, MobileBERT, and TinyBERT for fake news detection.

  • Practical examples and hands-on exercises to build and evaluate models.

  • Benchmarking performance of distilled models against BERT-Base.

Section 8: Named Entity Recognition (NER) with DistilBERT

  • Techniques for fine-tuning DistilBERT for NER in restaurant search applications.

  • Detailed guide on data preparation, tokenization, and model training.

  • Hands-on sessions to build, evaluate, and deploy NER models.

Section 9: Custom Summarization with T5 Transformer

  • Practical guide to fine-tuning the T5 model for summarization tasks.

  • Detailed walkthrough of dataset analysis, tokenization, and model fine-tuning.

  • Implement summarization predictions on custom data.

Section 10: Vision Transformer for Image Classification

  • Introduction to Vision Transformers (ViT) and their applications.

  • Step-by-step guide to using ViT for classifying Indian foods.

  • Practical exercises on image preprocessing, model training, and evaluation.

Section 11: Fine-Tuning Large Language Models on Custom Datasets

  • Theoretical insights and practical steps for fine-tuning large language models (LLMs).

  • Explore various fine-tuning techniques, including PEFT, LORA, and QLORA.

  • Hands-on coding sessions to implement custom dataset fine-tuning for LLMs.

Section 12: Specialized Topics in Transformer Fine-Tuning

  • Learn about advanced topics such as 8-bit quantization and adapter-based fine-tuning.

  • Review and implement state-of-the-art techniques for optimizing Transformer models.

  • Practical sessions to generate product descriptions using fine-tuned models.

Section 13: Building Chat and Instruction Models with LLAMA

  • Learn about advanced topics such as 4-bit quantization and adapter-based fine-tuning.

  • Techniques for fine-tuning the LLAMA base model for chat and instruction-based tasks.

  • Practical examples and hands-on guidance to build, train, and deploy chat models.

  • Explore the significance of chat format datasets and model configuration for PEFT fine-tuning.

Enroll now in "Mastering Transformer Models and LLM Fine Tuning on Custom Dataset" and gain the skills to harness the power of state-of-the-art NLP models. Whether you're just starting or looking to enhance your expertise, this course offers valuable knowledge and practical experience to elevate your proficiency in the field of natural language processing.

Unlock the full potential of Transformer models with our comprehensive course. Master fine-tuning techniques for BERT variants, explore knowledge distillation with DistilBERT, MobileBERT, and TinyBERT, and apply advanced models like RoBERTa, ALBERT, XLNet, and Vision Transformers for real-world NLP applications. Dive into practical examples using Hugging Face tools, T5 for summarization, and learn to build custom chat models with LLAMA.

Keywords: Transformer models, fine-tuning BERT, DistilBERT, MobileBERT, TinyBERT, RoBERTa, ALBERT, XLNet, ELECTRA, ConvBERT, DeBERTa, Vision Transformer, T5, BART, Pegasus, GPT-3, DeiT, Swin Transformer, Hugging Face, NLP applications, knowledge distillation, custom chat models, LLAMA.

Course Content

  • 13 section(s)
  • 141 lecture(s)
  • Section 1 Introduction
  • Section 2 Hello Transformers
  • Section 3 Transformers Architectures and Basic LLM Concepts
  • Section 4 BERT Architecture Theory
  • Section 5 Fine-Tuning BERT for Multi-Class Sentiment Classification for Twitter Tweets
  • Section 6 Knowledge Distillation for BERT - DistilBERT, MobileBERT and TinyBERT [Theory]
  • Section 7 Fake News Detection using DistilBERT, MobileBERT and TinyBERT
  • Section 8 Restaurant Search NER Recognition By Fine Tuning DistilBERT
  • Section 9 Fine Tuning T5 for Custom Summarization Task
  • Section 10 Fine Tuning Vision Transformer (ViT) for Indian Foods Classification
  • Section 11 Fine Tuning LLM on Custom Dataset [Theory]
  • Section 12 Fine Tuning LLM (Phi2/ or Any LLM) on Custom Data [Coding]
  • Section 13 Fine Tuning LLAMA Base Model as Chat/Instruction Model on Custom Data

What You’ll Learn

  • Understand transformers and their role in NLP.
  • Gain hands-on experience with Hugging Face Transformers.
  • Learn about relevant datasets and evaluation metrics.
  • Fine-tune transformers for text classification, question answering, natural language inference, text summarization, and machine translation.
  • Understand the principles of transformer fine-tuning.
  • Apply transformer fine-tuning to real-world NLP problems.
  • Learn about different types of transformers, such as BERT, GPT-2, and T5.
  • Hands-on experience with the Hugging Face Transformers library


Reviews

  • N
    Nikita
    5.0

    This is the only course which finally teaches how to fine-tune your own LLM.

  • P
    Paresh Patel
    5.0

    The course is very excellent especially to start learning llms.

  • S
    Sujata Katkade
    5.0

    The course conducted very well with practical explanations

  • N
    Nguyen Hoang Huy
    5.0

    What I need

Start FollowingSee all

We use cookies to enhance your experience on our website. Please read and confirm your agreement to our Privacy Policy and Terms and Conditions before continue to browse our website.

Read and Agreed