Skip to content
Concept-Lab
← Back to Machine Learning

Advanced Learning Algorithms

The advanced ML track. It stays transcript-driven as new advanced lessons are added into Concept Lab.

0 / 62 completed0%
62 remainingShowing 62 of 62 nodes

Concepts Covered

Advanced Learning Algorithms

0/62 done
1

Introduction to Advanced Learning Algorithms

Course 2 overview: neural networks, decision trees, and practical ML system advice.

Interactive
2

Neurons and the Brain

The biological motivation behind neural networks and why deep learning took off when it did.

Theory
3

Neural Networks: Demand Prediction

Building intuition for neural networks using a T-shirt top-seller prediction example.

Interactive
4

Recognising Images with Neural Networks

How neural networks build up visual understanding layer by layer — edges, parts, then faces.

Theory
5

Neural Network Layers

Layer notation, superscripts, and how a single hidden layer computes its activations.

Interactive
6

More Complex Neural Networks

Multi-layer networks, counting conventions, and the general activation formula for any layer.

Theory
7

Inference & Forward Propagation

The algorithm for making predictions: computing activations left to right through all layers.

Interactive
8

Inference in Code (TensorFlow)

Implementing forward propagation in TensorFlow with Dense layers — the coffee roasting example.

Theory
9

Data Representation in TensorFlow

NumPy 1D vectors vs 2D matrices, TensorFlow tensors, and why the double bracket matters.

Theory
10

Building a Neural Network in TensorFlow

The Sequential API: string layers together, compile, fit, and predict in one workflow.

Interactive
11

Forward Prop: Single Layer from Scratch

Implementing forward propagation in raw Python/NumPy — understanding what TensorFlow does under the hood.

Interactive
12

General Forward Propagation

The dense() function: a reusable loop-based implementation of any layer using a weight matrix.

Interactive
13

Is There a Path to AGI?

ANI vs AGI, the one learning algorithm hypothesis, and keeping hype calibrated.

Theory
14

Vectorised Neural Network Implementation

Why matrix multiplication makes neural networks fast and how NumPy's matmul replaces for-loops.

Theory
15

Matrix Multiplication

Dot products, transposes, and vector-matrix products — the mathematical building blocks of neural networks.

Theory
16

Matrix Multiplication Rules

The general formula for AᵀW — computing every element systematically from rows and columns.

Theory
17

Matrix Multiplication in Code

NumPy matmul, the @ operator, and the complete vectorised forward pass in TensorFlow.

Theory
18

Training a Neural Network: Overview

The three-step training loop in TensorFlow: specify architecture, compile with loss, fit to data.

Lab
19

Training Details: Loss, Cost, and Backprop

Binary cross-entropy loss, cost function over all examples, and how TensorFlow uses backprop internally.

Lab
20

Alternatives to the Sigmoid Activation

ReLU, linear activation, and why hidden layers should almost never use sigmoid.

Lab
21

Choosing Activation Functions

How to pick the right activation function for output and hidden layers based on what you're predicting.

Lab
22

Why Do We Need Activation Functions?

Without nonlinear activations, a deep neural network collapses into a simple linear model.

Interactive
23

Multiclass Classification

When y can take more than two values — classifying digits 0–9, diseases, or defect types.

Interactive
24

Softmax Regression

The generalization of logistic regression to n classes — computing mutually exclusive class probabilities.

Interactive
25

Neural Network with Softmax Output

Plugging softmax into the output layer to build a multiclass neural network.

Interactive
26

Numerically Stable Softmax

Using from_logits=True to avoid floating-point roundoff errors in softmax and logistic loss.

Interactive
27

Multi-Label Classification

When one input can have multiple independent labels simultaneously — cars AND pedestrians in one image.

Interactive
28

Adam Optimizer

Adaptive Moment Estimation — the de facto standard optimizer that auto-adjusts per-parameter learning rates.

Interactive
29

Convolutional Layers

Beyond dense layers — how convolutional layers let neurons see only local regions for speed and robustness.

Interactive
30

What is a Derivative?

Intuitive foundations of calculus derivatives — the engine behind gradient descent and backpropagation.

Interactive
31

Computation Graph and Backprop

How neural network frameworks compute gradients efficiently using forward and backward passes through a graph.

Lab
32

Backprop in a Larger Network

Tracing backpropagation through a two-layer network — seeing how gradients flow back to every parameter.

Interactive
33

Deciding What to Try Next

The systematic approach to ML debugging — why intuition fails and diagnostics save months of wasted effort.

Interactive
34

Evaluating a Model

Train/test splits and why J_train alone deceives you — measuring generalization systematically.

Interactive
35

Model Selection and Cross-Validation

The three-way split — why you need a cross-validation set to choose models without contaminating the test set.

Interactive
36

Diagnosing Bias and Variance

The two fundamental failure modes of ML models — high bias underfits, high variance overfits.

Interactive
37

Regularization and Bias-Variance

How λ shifts the bias-variance tradeoff — and how cross-validation finds the sweet spot automatically.

Interactive
38

Establishing a Baseline Level of Performance

Why raw error numbers are misleading without a baseline — human-level performance as the anchor for bias-variance judgment.

Interactive
39

Learning Curves

How training and cross-validation error change as data grows, and what that tells you about whether collecting more data is worth it.

Interactive
40

Deciding What to Try Next, Revisited

How bias and variance map directly to the next engineering move, so you stop guessing and start debugging systematically.

Interactive
41

Bias, Variance, and Neural Networks

Why deep learning changed the old bias-variance tradeoff story and gave engineers a new recipe for improving models.

Interactive
42

Iterative Loop of ML Development

The real workflow of ML engineering: choose architecture, train, diagnose, refine, and repeat until performance is good enough.

Interactive
43

Error Analysis

Manual review of model mistakes to discover which error classes matter most and where engineering effort will pay off.

Interactive
44

Adding Data

Targeted data collection, augmentation, and synthetic data generation as strategic tools for improving model quality.

Interactive
45

Transfer Learning

Use a model pre-trained on a large related dataset, then fine-tune it on your smaller task to get strong results with limited data.

Interactive
46

Full Cycle of a Machine Learning Project

Training a model is only one stage; real ML systems also require scoping, deployment, monitoring, retraining, and MLOps discipline.

Lab
47

Fairness, Bias, and Ethics

Why ML engineers must think about harm, subgroup performance, and mitigation plans before and after deployment.

Interactive
48

Error Metrics for Skewed Datasets

Why accuracy becomes misleading on rare-event problems, and how the confusion matrix gives a more truthful view of model usefulness.

Lab
49

Trading Off Precision and Recall

How threshold choices change which rare events you catch, which false alarms you accept, and why F1 is a useful but incomplete summary.

Interactive
50

Decision Tree Model

A decision tree predicts by asking a sequence of feature-based questions, routing an example down branches until it reaches a leaf decision.

Interactive
51

Decision Tree Learning Process

How a tree is built recursively: choose the best split, partition the data, repeat on each branch, and stop when further splitting is no longer worth it.

Lab
52

Measuring Purity: Entropy

Entropy is the impurity measure that tells a decision tree how mixed a node is, with 0 meaning pure and 1 meaning maximally mixed in the binary case.

Lab
53

Choosing a Split with Information Gain

Information gain measures how much a candidate split reduces weighted entropy, allowing the tree to choose the most purity-improving feature.

Interactive
54

Decision Tree: Putting It Together

The full tree-building algorithm combines repeated split selection, recursive branch construction, and stopping rules into one practical training loop.

Lab
55

One-Hot Encoding of Categorical Features

How to convert a feature with multiple discrete categories into several binary indicators so trees and other models can use it cleanly.

Lab
56

Continuous-Valued Features

How trees handle numeric features by testing candidate thresholds and selecting the split with the highest information gain.

Lab
57

Regression Trees

Generalizing decision trees from class prediction to numeric prediction by minimizing weighted variance and predicting leaf averages.

Lab
58

Using Multiple Decision Trees

Why single trees are sensitive to small data changes and how voting across many trees improves robustness.

Lab
59

Sampling with Replacement

Bootstrap sampling creates new training sets by repeatedly drawing from the original set with replacement.

Lab
60

Random Forest Algorithm

Bagging plus random feature subsets per split yields more diverse trees and stronger aggregate performance.

Lab
61

XGBoost

Boosted trees focus sequentially on hard examples and are often top-performing on structured/tabular tasks.

Interactive
62

When to Use Decision Trees

Choosing between tree ensembles and neural networks based on data modality, iteration speed, interpretability, and transfer learning needs.

Lab