Udemy

Mastering Generative AI and LLM Deployment.

Enroll Now
  • 535 Students
  • Updated 12/2025
4.6
(32 Ratings)
CTgoodjobs selects quality courses to enhance professionals' competitiveness. By purchasing courses through links on our site, we may receive an affiliate commission.

Course Information

Registration period
Year-round Recruitment
Course Level
Study Mode
Duration
15 Hour(s) 13 Minute(s)
Language
English
Taught by
PhD Researcher AI & Robotics Scientist Fikrat Gasimov
Rating
4.6
(32 Ratings)

Course Overview

Mastering Generative AI and LLM Deployment.

Proficiency on OpenAI,LangChain,MidJourney,LLama3,;Javascript Applications for 20X Fast Inference Prototypes.Get Hired

This course is diving into Generative AI State-Of-Art Scientific Challenges. It helps to uncover ongoing problems and develop or customize your Own Large Models Applications. Course mainly is suitable for any candidates(students, engineers,experts) that have great motivation to Large Language Models with Todays-Ongoing Challenges as well as their  deeployment with Python Based and Javascript Web Applications, as well as with C/C++ Programming Languages. Candidates will have deep knowledge on  TensorFlow , Pytorch,  Keras models, HuggingFace with Docker Service.

In addition, one will be able to optimize and quantize TensorRT frameworks for deployment in variety of sectors. Moreover, They will learn deployment of LLM quantized model to Web Pages developed with React, Javascript and FLASK
Here you will also learn how  to integrate Reinforcement Learning(PPO) to Large Language Model, in order to fine them with Human Feedback based. 
Candidates will learn how to code and debug in C/C++ Programming languages at least in intermediate level.

LLM Models used:

  • The Falcon,

  • LLAMA2,

  • BLOOM,

  • MPT,

  • Vicuna,

  • FLAN-T5,

  • GPT2/GPT3, GPT NEOX

  • BERT 101, Distil BERT

  • FINE-Tuning Small Models under supervision of BIG Models

Image Generation :

  1. LLAMA models

  2. Gemini

  3. Dall-E OpenAI

  4. Hugging face Models




  1. Learning and Installation of Docker from scratch

  2. Knowledge of Javscript, HTML ,CSS, Bootstrap

  3. React Hook, DOM and Javacscript Web Development

  4. Deep Dive on Deep Learning Transformer based Natural Language Processing

  5. Python FLASK  Rest API along with MySql

  6. Preparation of DockerFiles, Docker Compose as well as Docker Compose Debug file

  7. Configuration and Installation of Plugin packages in Visual Studio Code

  8. Learning, Installation and Confguration of frameworks such as Tensorflow, Pytorch, Kears with docker images from scratch

  9. Preprocessing and Preparation of Deep learning datasets for training and testing

  10. OpenCV  DNN with C++ Inference

  11. Training, Testing and Validation of Deep Learning frameworks

  12. Conversion of prebuilt models to Onnx  and Onnx Inference on images with C++ Programming

  13. Conversion of onnx model to TensorRT engine with C++ RunTime and Compile Time API

  14. TensorRT engine Inference on images and videos

  15. Comparison of achieved metrices and result between TensorRT and Onnx Inference

  16. Prepare Yourself for C++ Object Oriented Programming Inference!

  17. Ready to solve any programming challenge with C/C++

  18. Read to tackle Deployment issues on Edge Devices as well as Cloud Areas

  19. Large Language Models Fine Tunning

  20. Large Language Models Hands-On-Practice: BLOOM, GPT3-GPT3.5, FLAN-T5 family

  21. Large Language Models Training, Evaluation and User-Defined Prompt IN-Context Learning/On-Line Learning

  22. Human FeedBack Alignment on LLM with Reinforcement Learning (PPO) with Large Language Model : BERT and FLAN-T5

  23. How to Avoid Catastropich Forgetting Program on Large Multi-Task LLM Models.

  24. How to prepare LLM for Multi-Task Problems such as Code Generation, Summarization, Content Analizer, Image Generation.

  25. Quantization of Large Language Models with various existing state-of-art techniques


  • Importante Note:
          In this course, there is not nothing to copy & paste, you will put your hands in every line of project to be successfully LLM and Web Application Developer!

You DO NOT need any Special Hardware component. You will be delivering project either on CLOUD or on Your Local Computer.



Course Content

  • 10 section(s)
  • 143 lecture(s)
  • Section 1 All course summary
  • Section 2 Some Demos
  • Section 3 Set up Docker Images,Containers, and Visual Code
  • Section 4 Prepare YoloV7 Fast Precision Server Side
  • Section 5 Flask Server Implementation for High Security Web App
  • Section 6 Flask Server with YoloV7 Deep Learning Integration
  • Section 7 Flask Server Web APP Design
  • Section 8 React Web App Inference with Emotion Detection NLP
  • Section 9 Pytorch Dataloader & Hugging Face Framework(Large Language Models)
  • Section 10 BERT NLP Transformer : Model Freezing

What You’ll Learn

  • What is Docker and How to use Docker
  • Advance Docker Usage
  • What are OpenCL and OpenGL and when to use ?
  • (LAB) Tensorflow and Pytorch Installation, Configuration with Docker
  • (LAB)DockerFile, Docker Compile and Docker Compose Debug file configuration
  • (LAB)Different YOLO version, comparisons, and when to use which version of YOLO according to your problem
  • (LAB)Jupyter Notebook Editor as well as Visual Studio Coding Skills
  • (LAB)Learn and Prepare yourself for full stack and c++ coding exercies
  • (LAB)TENSORRT PRECISION FLOAT 32/16 MODEL QUANTIZIATION
  • Key Differences:Explicit vs. Implicit Batch Size
  • (LAB)TENSORRT PRECISION INT8 MODEL QUANTIZIATION
  • (LAB) Visual Studio Code Setup and Docker Debugger with VS and GDB Debugger
  • (LAB) what is ONNX framework C Plus and how to use apply onnx to your custom C ++ problems
  • (LAB) What is TensorRT Framework and how to use apply to your custom problems
  • (LAB) Custom Detection, Classification, Segmentation problems and inference on images and videos
  • (LAB) Advance C ++ Object Oriented Programming
  • (LAB) Deep Learning Problem Solving Skills on Edge Devices, and Cloud Computings with C++ Programming Language
  • (LAB) How to generate High Performance Inference Models on Embedded Device, in order to get high precision, FPS detection as well as less gpu memory consumption
  • (LAB) Visual Studio Code with Docker
  • (LAB) GDB Debugger with SonarLite and SonarCube Debuggers
  • (LAB) yolov4 onnx inference with opencv c++ dnn libraries
  • (LAB) yolov5 onnx inference with opencv c++ dnn libraries
  • (LAB) yolov5 onnx inference with Dynamic C++ TensorRT Libraries
  • (LAB) C++(11/14/17) compiler programming exercies
  • Key Differences: OpenCV AND CUDA/ OPENCV AND TENSORRT
  • (LAB) Deep Dive on React Development with Axios Front End Rest API
  • (LAB) Deep Dive on Flask Rest API with REACT with MySql
  • Understand model optimization techniques: Pruning, Distillation, and Quantization
  • (LAB) Deep Dive on Text Summarization Inference on Web App
  • (LAB) Prompt Penetration Testing
  • (LAB) Deep Dive on BERT (LLM) Fine tunning and Emotion Analysis on Web App
  • (LAB) Deep Dive On Distributed GPU Programming with Natural Language Processing (Large Language Models))
  • (LAB) Deep Dive on BERT (LLM) Fine tunning and Emotion Analysis on Web App
  • (LAB) Prompt Engineering from basics to advance
  • (LAB) Deep Dive on Generative AI use cases, project lifecycle, and model pre-training
  • (LAB) OPENAI GPT models with Specific prompt Engineering techniques
  • (LAB) Fine-tuning and evaluating large language models
  • (LAB) Reinforcement learning and LLM-powered applications, ALIGN Fine tunning with User Feedback
  • (LAB) Quantization of Large Language Models with Modern Nvidia GPU's
  • (LAB) C++ OOP TensorRT Quantization and Fast Inference
  • (LAB) Deep Dive on Hugging FACE Library
  • (LAB)Translation ● Text summarization ● Question answering
  • (LAB)Sequence-to-sequence models, ONLY Encoder Based Models, Only Decoder Based Models
  • (LAB)Define the terms Generative AI, large language models, prompt, and describe the transformer architecture that powers LLMs
  • (LAB)Discuss computational challenges during model pre-training and determine how to efficiently reduce memory footprint
  • (LAB)Describe how fine-tuning with instructions using prompt datasets can improve performance on one or more tasks
  • (LAB)Explain how PEFT decreases computational cost and overcomes catastrophic forgetting
  • (LAB)Describe how RLHF uses human feedback to improve the performance and alignment of large language models
  • (LAB)Discuss the challenges that LLMs face with knowledge cut-offs, and explain how information retrieval and augmentation techniques can overcome these challen
  • Recognize and understand the various strategies and techniques used in fine-tuning language models for specialized applications.
  • Master the skills necessary to preprocess datasets effectively, ensuring they are in the ideal format for AI training.
  • Investigate the vast potential of fine-tuned AI models in practical, real-world scenarios across multiple industries.
  • Acquire knowledge on how to estimate and manage the costs associated with AI model training, making the process efficient and economic
  • Distributing Computing for (DDP) Distributed Data Parallelization and Fully Shared Data Parallel across multi GPU/CPU with Pytorch together with Retrieval Augme
  • The RoBERTa model was proposed in RoBERTa: A Robustly Optimized BERT Pretraining Approach
  • Master downcasting from FP32 to BF16 and FP32 to INT8
  • Learn the difference between symmetric and asymmetric quantization
  • Implement quantization techniques in Python with real examples
  • Apply quantization to make models more efficient and deployment-ready
  • Learn the basics of data types like FP32, FP16, BFloat16, and INT8
  • Gain practical skills to optimize models for edge devices and resource-constrained environments
  • Advance Image Generation and Editing


Reviews

  • S
    Shaik Abdul Haffiez
    1.0

    No this is not the course that i was expected, the explanation was too poor, This is not for a Beginer

  • Y
    YUSHU LIU
    5.0

    Thank you very much. You are a great teacher.

  • G
    Gulia
    5.0

    Go forward! Nice course!

  • G
    Girts Karnit
    5.0

    State of art Generative AI models are well explained

Start FollowingSee all

We use cookies to enhance your experience on our website. Please read and confirm your agreement to our Privacy Policy and Terms and Conditions before continue to browse our website.

Read and Agreed