Click on a date to show the videos for that class.
(Thanks to Professor Mike Genesereth from Stanford)
(Only the first 16 minutes are relevant)
These videos are supplemental. It's good to have a feel for how this is done, but I won't expect you to derive the backpropagation algorithm on a test.
This video is from Pieter Abbeel's AI course at Berkely. The schedule is:
| 0-9:40 | Introduction to Markov Models. WATCH THIS. |
| 9:40 -25:00 | Stationary Distributions of Markov Models. You don't need to watch this, but it is interesting. He explains Google's PageRank algorithm in the context of Markov Models. |
| 25:00-27:30 | Gibbs Sampling. Skip this. |
| 27:30-47:00 | Introduction to Hidden Markov Models. WATCH THIS. |
| 47:00-END | Inference in Hidden Markov Models. We'll do this on Monday. |
This video is from Pieter Abbeel's AI course at Berkely. The schedule is:
| 0-9:40 | Introduction to Markov Models. ALREADY WATCHED. |
| 9:40 -25:00 | Stationary Distributions of Markov Models. You don't need to watch this, but it is interesting. He explains Google's PageRank algorithm in the context of Markov Models. |
| 25:00-27:30 | Gibbs Sampling. Skip this. |
| 27:30-47:00 | Introduction to Hidden Markov Models. ALREADY WATCHED. |
| 47:00-END | Inference in Hidden Markov Models. WATCH THIS. |
This video is the sequel to the previous video:
| 0-9:00 | HMM Review. SKIP THIS. |
| 9:00 -26:50 | Introduction to particle filters. WATCH THIS. |
| 26:50-35:00 | Particle Filtering Quiz. SKIP THIS. |
| 35:00-40:00 | Example of using a particle filter for robot localization. WATCH THIS IF YOU WANT. |
| 40:00-50:00 | Simultaneous Localization and Mapping (SLAM). SKIP THIS. |
| 50:00-56:00 | Dynamic Bayes Nets. WATCH THIS. |
| 56:00-END | More application examples. Watch if you want. |
Another video from Pieter Abbeel's CS188 course at Berkely. If you have time, you should watch the whole thing. You should definitely watch:
| 0-13:30 | Introduction to Markov Decision Processes. |
| 22:40 -29:00 | Assigning values to reward sequences. |
| 41:21-101:30 | Value iteration. |
Another video from Pieter Abbeel's CS188 course at Berkely.
This is the final in our series of 188 video. Here is a rough schedule:
| 0-13:20 | Review of MDP's and Q-learning. |
| 13:20-20:20 | Exploration vs. Exploitation. WATCH THIS. It relates to PA4. |
| 20:20-30:50 | Exploration function and regret. This introduces a more efficient way to handle exploration. It is interesting, but not required. |
| 31:00-50:50 | Approximate Q-learning. WATCH THIS. This describes the idea behind the final question for the PA. |
| 50:50 - 59:00 | Justification/Explanation of approximate Q-learning in terms of linear regression. This is interesting, and ties nicely back to material from earlier in the course, but it isn't required material. |
| 59:00 - END | Policy search for reinforcement learning. A different approach to RL. This relates to Friday's discussion of heuristics for function optimization. Again, this is is interesting but not required. |
