Course Information
- Available
- *The delivery and distribution of the certificate are subject to the policies and arrangements of the course provider.
Course Overview
Roadmap to become AI QA Engineer to test LLMs and AI Application using DeepEval, RAGAs and HF Evaluate with Local LLMs
Testing AI & LLM App with DeepEval, RAGAs & more using Ollama and Local Large Language Models (LLMs)
Master the essential skills for testing and evaluating AI applications, particularly Large Language Models (LLMs). This hands-on course equips QA, AI QA, Developers, data scientists, and AI practitioners with cutting-edge techniques to assess AI performance, identify biases, and ensure robust application development.
Topics Covered:
Section 1: Foundations of AI Application Testing (Introduction to LLM testing, AI application types, evaluation metrics, LLM evaluation libraries).
Section 2: Local LLM Deployment with Ollama (Local LLM deployment, AI models, running LLMs locally, Ollama implementation, GUI/CLI, setting up Ollama as API).
Section 3: Environment Setup (Jupyter Notebook for tests, setting up Confident AI).
Section 4: DeepEval Basics (Traditional LLM testing, first DeepEval code for AnswerRelevance, Context Precision, evaluating in Confident AI, testing with local LLM, understanding LLMTestCases and Goldens).
Section 5: Advanced LLM Evaluation (LangChain for LLMs, evaluating Answer Relevancy, Context Precision, bias detection, custom criteria with GEval, advanced bias testing).
Section 6: RAG Testing with DeepEval (Introduction to RAG, understanding RAG apps, demo, creating GEval for RAG, testing for conciseness & completeness).
Section 7: Advanced RAG Testing with DeepEval (Creating multiple test data, Goldens in Confident AI, actual output and retrieval context, LLMTestCases from dataset, running evaluation for RAG).
Section 8: Testing AI Agents and Tool Callings (Understanding AI Agents, working with agents, testing agents with and without actual systems, testing with multiple datasets).
Section 9: Evaluating LLMs using RAGAS (Introduction to RAGAS, Context Recall, Noise Sensitivity, MultiTurnSample, general purpose metrics for summaries and harmfulness).
Section 10: Testing RAG applications with RAGAS (Introduction and setup, creating retrievers and vector stores, MultiTurnSample dataset for RAG, evaluating RAG with RAGAS).
Course Content
- 15 section(s)
- 95 lecture(s)
- Section 1 Introduction
- Section 2 Running LLM locally using Ollama
- Section 3 Complete course Source code
- Section 4 Environment step required for Testing/Evaluating LLM Apps and LLMs
- Section 5 Understanding the Basics of DeepEval (Building Blocks)
- Section 6 Evaluating Real LLMs (Locally) as the Source and creating Dataset with LLMs
- Section 7 Testing RAG (Retrieval-Augmented Generation) application using DeepEval
- Section 8 Testing RAG application with DeepEval (Advanced)
- Section 9 Testing AI Agents and Tool Callings with Local LLMs and DeepEval
- Section 10 Evaluating/Testing LLMs using RAGAs
- Section 11 Testing RAG applications with RAGAs
- Section 12 Functional Testing of Large Language Models (LLMs) with HuggingFace and Python
- Section 13 Evaluate LLMs using HuggingFace Evaluate
- Section 14 Component Testing of RAG LLMs Application with DeepEval
- Section 15 Component Testing of AI Agent and Tool Calling with DeepEval
What You’ll Learn
- Understand the purpose of Testing LLM and LLM based Application
- Understand DeepEval and RAGAs in detail from complete ground up
- Understand different metrics and evaluations to evaluate LLMs and LLM based app using DeepEval and RAGAs
- Understand the advanced concepts of DeepEval and RAGAs
- Testing RAG based application using DeepEval and RAGAs
- Testing AI Agents using DeepEval to understand how tool callings can be tested
Skills covered in this course
Reviews
-
TTzechung Kao
It is so good that I understood the concept.
-
LLeonardo Miro
Karthik does a fine job introducing and showing DeepEval and RAGAs at a programmatic level. The last chapters are really challenging depending on your programming skills. A correct roadmap for testers interested in AI.
-
FFelix Schlegel
The course is ok to get a quick overview over several libraries. However, it is quite superficial, the example are not much more than what you find in the documentation. I am missing a real world scenario where you would create a large synthetic test dataset using tools like Ragas, then compare the performance of different RAG applications, not just passing one or two manual test cases.
-
GGourav Kumar Laha
Good explanation