Ch 3 — Machine Learning Paradigms — Under the Hood
Loss functions, gradient descent, K-Means internals, Q-learning math, and self-supervised pretraining
Under the Hood
-
Click play or press Space to begin. Click any node for deep-dive details...
ASupervised Learning InternalsLoss functions · gradient descent · evaluation
1functions
Loss FunctionsMSE, cross-entropy
measuring error
2trending_down
Gradient DescentOptimizing weights
step by step
assessment
EvaluationAccuracy, F1
train/test split
3arrow_downward From labeled to unlabeled: unsupervised internals
BUnsupervised Learning InternalsK-Means math · PCA eigenvectors · anomaly detection
hub
K-MeansLloyd’s algorithm
iterative assignment
4compress
PCAEigendecomposition
variance maximization
warning
AnomalyIsolation forest
outlier detection
5arrow_downward From patterns to actions: reinforcement learning math
CReinforcement Learning InternalsMDPs · Bellman equation · Q-learning
schema
MDPStates, actions
transitions, rewards
6calculate
Q-LearningBellman equation
value iteration
balance
ε-GreedyExplore vs
exploit balance
7arrow_downward The bias-variance tradeoff and model selection
DModel Selection & ValidationBias-variance · cross-validation · regularization
tune
Bias-VarianceUnderfit vs
overfit tradeoff
8grid_view
Cross-ValidationK-fold reliable
evaluation
shield
RegularizationL1, L2, dropout
prevent overfit
9arrow_downward Modern paradigms: self-supervised & transfer learning
EModern ParadigmsSelf-supervised pretraining · transfer learning · RLHF
auto_awesome
Self-SupervisedNext-token, masked
contrastive learning
10merge
Full PipelinePretrain + finetune
+ RLHF