# # CS294-158-SP20 Deep Unsupervised Learning Spring 2020

Another great course taught by Pieter Abbeel, some of these ideas are quite useful in research. https://www.youtube.com/watch?v=V9Roouqfu-M&list=PLwRJQ4m4UJjPiJP3691u-qWwPGVKzSlNP&t=776s&ab_channel=PieterAbbeel

Main thing for understanding this is for F1TENTH.

Lecture 1

Lecture 2

Lecture 3

### Likelihood-based models

Problems we’d like to solve:

- Generating data: synthesizing images, videos, speech, text
- Compressing data: constructing efficient codes
- Anomaly detection

Likelihood-based models: estimate $p_{data}$ from samples $x_{1},…,x_{n}$

Learns a distribution $p$ that allows:

- Computing $p(x)$ for arbitrary $x$
- Sampling $x∼p(x)$

There is a parallel with STAT206, where we learn MLE.

I’ve always wondered, when we’ve been learning to do MLE, all we are really doing is taking simple estimates, such as the sampled mean to estimate $μ$ for example for a Normal Distribution.

But if we don’t know what distribution the data follows, can’t we just sample a bunch of data, and generate a Probability Density Function out of it?

The problem is that in higher dimensions this doesn’t work, and even in 1-dimension, this is very prone to overfitting.

Instead, we do function approximation. Instead of storing each probability, we store a parameterized function $p_{θ}(x)$.

This is where we talk about autoregressive models.

Novel NN architectures carry to other disciplines.