← back to reading

๐Ÿค– Machine Learning

Based on Deisenroth, Faisal & Ong, Mathematics for Machine Learning, licensed CC BY 4.0.

Also draws from Zhang et al., Dive into Deep Learning, licensed CC BY-SA 4.0.

Scheme REPLs run in the browser. Python equivalents in collapsible blocks.

data model prediction loss Data in, predictions out. Loss tells the model how wrong it is.
Chapter
1. What is ML? Function approximation from data — supervised, unsupervised, reinforcement ๐Ÿค–
2. Linear Regression Find weights that minimize squared error — normal equation or gradient descent ๐Ÿค–
3. Optimization Follow the negative gradient downhill — learning rate, convexity, convergence ๐Ÿค–
4. Logistic Regression Squash linear output through sigmoid for probabilities — cross-entropy loss ๐Ÿค–
5. Kernel Methods Replace dot products with kernel functions to learn nonlinear boundaries ๐Ÿค–
6. Neural Networks Stack linear layers with nonlinearities — universal approximation by composition ๐Ÿค–
7. Backpropagation The chain rule applied layer by layer — how gradients flow backward through a network ๐Ÿค–
8. Regularization Penalize complexity to prevent overfitting — L1, L2, dropout, early stopping ๐Ÿค–
9. Convolutional Networks Exploit spatial structure — shared filters, pooling, translation equivariance ๐Ÿค–
10. Recurrent Networks Process sequences by feeding hidden state forward — vanishing gradients and LSTMs ๐Ÿค–
11. Dimensionality Reduction PCA finds the axes of maximum variance — compress data without losing structure ๐Ÿค–
12. Clustering Group data by similarity — k-means, EM, and the bias-variance tradeoff in unsupervised learning ๐Ÿค–

๐Ÿ“บ Video lectures: Stanford CS229: Machine Learning ยท 3Blue1Brown: Neural Networks

Neighbors