Lab 3: The Machine Learning Taxonomy

Objective

Map the landscape of machine learning approaches. By the end you will understand:

  • The three main ML paradigms and when to use each

  • Key algorithms in each category

  • Real-world applications of each approach

  • How self-supervised learning powers modern LLMs


The Three Main Paradigms

                        MACHINE LEARNING

          ┌───────────────────┼───────────────────┐
          │                   │                   │
    Supervised          Unsupervised         Reinforcement
    Learning              Learning             Learning
    (labelled data)    (no labels needed)  (learn from rewards)
          │                   │                   │
    ┌─────┴─────┐       ┌─────┴─────┐       ┌─────┴─────┐
Classification Regression Clustering Generative  Policy
  (cat/dog)  (house price) (segments) (GANs, VAE) (games, robots)

1. Supervised Learning

The most common paradigm. Every training example has an input and a known correct output (label). The model learns the mapping.

Two subtypes:

Classification — "Which category?"

Output is a discrete class label.

Examples:

  • Email spam / not spam

  • Medical image: tumour / benign

  • Credit application: approve / deny

  • Sentiment: positive / neutral / negative

Regression — "How much?"

Output is a continuous number.

Examples:

  • House price prediction

  • Stock price forecasting

  • Demand forecasting

  • Temperature prediction

Common Supervised Algorithms

Algorithm
Best For
Interpretable?

Linear/Logistic Regression

Baseline, simple relationships

✅ Yes

Decision Trees

Categorical features, non-linear

✅ Yes

Random Forest

Tabular data, robust

Partially

Gradient Boosting (XGBoost)

Competitions, tabular data

Partially

Support Vector Machine

High-dimensional, small data

Partially

Neural Networks

Images, text, complex patterns

❌ No


2. Unsupervised Learning

No labels. The model must find structure in the data on its own.

Clustering — "Which group?"

Find natural groupings in data.

Examples:

  • Customer segmentation

  • Document clustering (group similar articles)

  • Anomaly detection (the point that doesn't fit any cluster)

  • Gene expression analysis

Dimensionality Reduction — "What's essential?"

Compress high-dimensional data while preserving structure.

Examples:

  • Visualising high-dimensional data (t-SNE, UMAP)

  • Feature compression before training

  • Noise removal from signals

  • Recommendation: compress user preferences to latent factors

Generative Models — "Create new data"

Learn the underlying distribution of the training data to generate new samples.

GANs (Generative Adversarial Networks):

VAEs (Variational Autoencoders):


3. Reinforcement Learning

The model (agent) learns by taking actions in an environment and receiving rewards or penalties. No labelled data — just trial and error.

Famous RL successes:

  • AlphaGo / AlphaZero — board games

  • OpenAI Five — Dota 2 at world championship level

  • AlphaStar — StarCraft II

  • ChatGPT — RLHF (humans rate responses; model learns to maximise ratings)

  • Robotics — Boston Dynamics locomotion

  • Data centre cooling — DeepMind reduced Google's cooling costs 40%

RL is hard because:

  • Sparse rewards — agent may do 10,000 actions before getting any feedback

  • Reward hacking — agent finds unintended ways to maximise reward

  • Sample inefficiency — requires millions of episodes to learn


4. Self-Supervised Learning — The Modern Paradigm

The paradigm that powers LLMs. Labels come from the data itself — no human annotation needed.

Masked Language Modelling (BERT):

Next Token Prediction (GPT):

By training on trillions of tokens from the internet using these self-supervised tasks, the model is forced to develop deep understanding of language, facts, reasoning patterns, and world knowledge — without a single human-labelled example.

This is the key insight behind modern LLMs: language modelling at scale = general intelligence.


Choosing the Right Approach


Summary

Paradigm
Signal
Typical Use
Example Models

Supervised

Labelled data

Classification, regression

XGBoost, ResNet, BERT

Unsupervised

None

Clustering, generation

K-Means, GAN, VAE

Reinforcement

Reward signal

Games, robotics, alignment

PPO, AlphaZero

Self-supervised

Labels from data

LLMs, vision-language

GPT, CLIP, DALL-E


Further Reading

Last updated