Lecture 02

A Brief History of Deep Learning

Reminders


Recap

Formally, a computer program is said to learn from experience $ℇ$ with respect to some task $𝒯$ and performance measure $𝒫$ if its performance at $𝒯$ as measured by $𝒫$ improves with $ℇ$.


Structured data


Machine Learning Jargon


History of Machine Learning


Artificial Neurons and Perceptrons


Backpropagation

Neural networks as computation graphs

Computation graph showing input, intermediate computations, and outputs
\[\frac{\partial f_n}{\partial x} = \sum_{i_1 \in \pi(n)} \frac{\partial f_n}{\partial f_{i_1}} \frac{\partial f_{i_1}}{\partial x} = \sum_{i_1 \in \pi(n)} \frac{\partial f_n}{\partial f_{i_1}} \sum_{i_2 \in \pi(i_1)} \frac{\partial f_{i_1}}{\partial f_{i_2}} \frac{\partial f_{i_2}}{\partial x} = \dots\]

About the term “Deep Learning”

“Representation learning is a set of methods that allows a machine to be fed with raw data and to automatically discover the representations needed for detection or classification. Deep learning methods are representation-learning methods with multiple levels of representation […]” — LeCun, Bengio, & Hinton (2015)


Activation Functions


Hardware


Large-Scale Unsupervised Learning

From GPT-1 (2018) to GPT-4:


Open Directions