Back to Courses
Advanced 6 weeks 3 sessions/week

LLM Fine-Tuning & Custom Model Training

Train and customise your own Large Language Models. Master LoRA, QLoRA, PEFT techniques, RLHF, and DPO alignment -- then deploy your fine-tuned models at scale with vLLM and TGI.

£3,997 per person

What You'll Learn

01

Hugging Face Ecosystem

Master the Hugging Face platform: Transformers library, model hub, tokenizers, datasets, and the Trainer API. Your complete toolkit for working with LLMs.

02

LoRA & QLoRA

Parameter-efficient fine-tuning with Low-Rank Adaptation. Understand the math behind LoRA, implement QLoRA for memory-efficient training, and configure rank, alpha, and target modules.

03

PEFT Techniques

Beyond LoRA: prefix tuning, prompt tuning, IA3, and adapter methods. Compare approaches, understand trade-offs, and choose the right PEFT technique for your use case.

04

Dataset Curation

Build high-quality training datasets. Data collection strategies, cleaning pipelines, deduplication, quality filtering, format conversion (Alpaca, ShareGPT, Chat), and synthetic data generation.

05

Training Infrastructure

Set up multi-GPU training environments. DeepSpeed, FSDP, gradient checkpointing, mixed precision, and distributed training across multiple nodes for large model fine-tuning.

06

Evaluation & Benchmarking

Evaluate fine-tuned models rigorously. Perplexity, BLEU, ROUGE, human evaluation frameworks, LLM-as-judge, and building custom benchmarks for your domain.

07

RLHF & DPO

Align models with human preferences. Reward modelling, PPO for RLHF, Direct Preference Optimisation (DPO), and constitutional AI approaches for safer model outputs.

08

Deploying Fine-Tuned Models

Serve your models in production with vLLM and Text Generation Inference (TGI). Quantisation for deployment (GPTQ, AWQ, GGUF), batching, streaming, and cost-efficient inference at scale.

Who Is This For

ML engineers and researchers who want to customise LLMs for specific domains and tasks

AI engineers building products that require specialised language models beyond off-the-shelf APIs

Technical leads evaluating whether to fine-tune vs. use API-based models for their organisation

Prerequisites

Deep learning, Python, and Transformer knowledge required. You should be comfortable with PyTorch, understand Transformer architecture, and have experience training neural networks. Our Deep Learning Masterclass provides ideal preparation.

Course Format

Live Online Sessions

Interactive sessions with real-time Q&A and screen sharing

Recorded Replays

All sessions recorded and available for 12 months after the course

Hands-on Projects

Real-world projects that build your portfolio as you learn

1-on-1 Mentoring

Personal mentoring sessions to address your specific questions

Certificate of Completion

Industry-recognised certificate upon successful completion

Schedule & Pricing

£3,997
6 weeks · 3 sessions per week · 18 sessions total
  • Live interactive sessions
  • 12-month replay access
  • 1-on-1 mentoring
  • Certificate included
Enrol Now

Your Instructors

PP

PeusoPeupon Expert Team

Our instructors are seasoned practitioners with years of experience building production AI systems. They hold certifications across major cloud platforms and have trained thousands of professionals worldwide.

Frequently Asked Questions

We provide GPU access through cloud platforms for all labs (typically A100 or equivalent). For personal experimentation, QLoRA allows fine-tuning on consumer GPUs (16GB VRAM). Budget approximately 100-200 pounds for cloud compute over 6 weeks.

We work primarily with Llama, Mistral, and Phi models of various sizes. You will learn to evaluate and select base models, so the skills transfer to any new model that is released.

Absolutely. Fine-tuning provides consistent behaviour, lower latency, reduced costs (smaller models), domain specialisation, and the ability to run models on your own infrastructure. It is complementary to RAG and prompting.

Yes, provided you use commercially-licensed base models (Llama 3, Mistral, etc.). We cover licensing considerations and help you choose models appropriate for your commercial use case.