Lab 15: LLM Fine-Tuning Infrastructure
Overview
Architecture
┌──────────────────────────────────────────────────────────────┐
│ LLM Fine-Tuning Infrastructure │
├──────────────────────────────────────────────────────────────┤
│ DATA PIPELINE │ TRAINING INFRASTRUCTURE │
│ Raw data → JSONL │ Base Model (HuggingFace) │
│ Instruction format │ PEFT/LoRA adapter init │
│ Train/val split │ Mixed precision (BF16) │
│ Quality filtering │ Gradient checkpointing │
│ │ DeepSpeed ZeRO / FSDP │
├──────────────────────────┴──────────────────────────────────┤
│ EVALUATION │ DEPLOYMENT │
│ BLEU, ROUGE, BERTScore │ Merge LoRA → Base model │
│ Human evaluation │ Quantize to INT4/INT8 │
│ Benchmark suites │ Serve with vLLM │
└──────────────────────────────────────────────────────────────┘Step 1: Fine-Tuning Approaches
Method
Trainable Params
GPU Memory (7B)
Quality
Use Case
Step 2: LoRA Architecture
Step 3: QLoRA
Step 4: Data Preparation
Step 5: Alignment Techniques (RLHF vs DPO)
Step 6: Compute Requirements
Step 7: Evaluation (BLEU, ROUGE, BERTScore)
Metric
Measures
Range
Use Case
Step 8: Capstone — LoRA Parameter Calculator
Summary
Concept
Key Points
Last updated
