====== Lectures ====== Lectures are given in person. Videos will be provided, but not necessarily directly from the lectures (the form is at the discretion of the individual lecturers). ===== Literature ===== In the lecture descriptions below, we refer to this **supplementary course material**: **Relevant** * [[https://web.stanford.edu/class/psych209/Readings/SuttonBartoIPRLBook2ndEd.pdf|RL]]: // R. S. Sutton, A. G. Barto: Reinforcement learning: An introduction. MIT press, 2018. // * [[https://web.stanford.edu/~jurafsky/slp3/|NLP]]: // D. Jurafsky & J. H. Martin: Speech and Language Processing - 3rd edition draft // * [[https://direct.mit.edu/books/book/2604/An-Introduction-to-Computational-Learning-Theory|COLT]]: // M. J. Kearns, U. Vazirani: An Introduction to Computational Learning Theory, MIT Press 1994// RL & NLP are available online. You are strongly discouraged from using this course's materials from previous years as you would run into confusions. The RL part of the course is heavily based on the RL course of prof Emma Brunskill. The relevant lectures from prof Brunskill's course are: [[https://youtu.be/FgzM3zpZ55o|Lecture 1]],[[https://youtu.be/E3f2Camj0Is|Lecture 2]], [[https://youtu.be/dRIhrn8cc9w|Lecture 3]], [[https://youtu.be/j080VBVGkfQ|Lecture 4]], [[https://youtu.be/buptHUzDKcE|Lecture 5]], [[https://youtu.be/gOV8-bC1_KU|Lecture 6]], [[https://youtu.be/RN8qpSs8ozY|Lecture 11]]. There are nice materials by Volodymyr Kuleshov and Stefano Ermon on probabilistic graphical models (for the Bayesian networks part of the course): [[https://ermongroup.github.io/cs228-notes/]]. The relevant chapters are: [[https://ermongroup.github.io/cs228-notes/representation/directed/]], [[https://ermongroup.github.io/cs228-notes/inference/ve/]], [[https://ermongroup.github.io/cs228-notes/inference/sampling/]]. The NLP part of the course is heavily based on the NLP course(s) from Dan Jurafsky (Stanford), following his book: Speech and Language Processing (see NLP above) - particularly its 3rd edition draft (2nd ed. is insufficient!). The relevant chapters for us are 3, 6, 7 and 9. There are also some nice related [[https://web.stanford.edu/~jurafsky/slp3/|materials]] and [[https://www.youtube.com/channel/UC_48v322owNVtORXuMeRmpA|videos]] For the COLT part: besides the monograph by Kearns et al linked above, the [[https://en.wikipedia.org/wiki/Computational_learning_theory#|Wikipedia page]] has pointers to two COLT survey papers (Angluin, Haussler) which are relevant to the PAC part. There are also external courses with lecture material available; for example, [[http://www.cs.cmu.edu/~ninamf/ML11/index.html|8803 Machine Learning Theory]] at Georgia Tech covers all COLT topics of SMU (there are subtle differences in the algorithms and proofs). Video footage of the lectures available [[https://www.youtube.com/playlist?list=PLQL6z4JeTTQlgt77fhOe2Jovjjz4THF_G|here]]. ---- ==== Lecture 1 - Reinforcement Learning 1 ==== **Slides:** {{:courses:smu:lecture_1.pdf}} **Videos:** [[https://www.dropbox.com/s/qrpgficwct1d56k/SMU%20Lecture%201%20-%20Markov%20Chains.mp4?dl=0|Markov Processes]], [[ https://www.dropbox.com/s/ytgbet2ek10chff/SMU%20Lecture%201%20-%20Markov%20Reward%20Processes.mp4?dl=0|Markov Reward Processes]], [[https://www.dropbox.com/s/swbkj2y89v55i86/Lecture%201%20-%20Markov%20Decision%20Processes.mp4?dl=0|Markov Decision Processes]]. **Coming soon:** Proofs ---- ==== Lecture 2 - Reinforcement Learning 2 ==== **Slides:** {{:courses:smu:lecture_2.pdf}} **Videos:** [[https://www.dropbox.com/s/je9hs6qhb8d9f4y/SMU%20Lecture%202%20-%20Introduction.mp4?dl=0|Introduction]], [[https://www.dropbox.com/s/s1s18oalnw34yv5/SMU%20Lecture%202%20-%20Statistical%20Properties%20of%20Estimators.mp4?dl=0|Statistical Properties of Estimators]], [[https://www.dropbox.com/s/mao9irop7y3b18o/SMU%20Lecture%202%20-%20Monte%20Carlo%20Methods.mp4?dl=0|Monte Carlo Value Evaluation]]. **Coming soon:** Temporal Difference Learning ---- ==== Lecture 3 - Reinforcement Learning 3 ==== **Slides:** {{:courses:smu:lecture_3.pdf}} **Videos:** [[https://www.dropbox.com/s/br9f5527rfbm3km/SMU%20Lecture%203.mp4?dl=0|Model-Free Control]] ---- ==== Lecture 4 - Reinforcement Learning 4 ==== **Slides:** {{:courses:smu:lecture_4.pdf}} ---- ==== Lecture 5 - Reinforcement Learning 5 ==== **Slides:** {{:courses:smu:lecture_5.pdf}} ---- ==== Lecture 6 - Bayesian Networks 1 ==== **Slides:** {{ :courses:smu:lecture_6.pdf |slides}} **Video:** [[https://www.dropbox.com/s/10mr8aadkbaiki8/SMU%20-%20Lecture%206.mp4?dl=0|Video]] ---- ==== Lecture 7 - Bayesian Networks 2 ==== **Slides:** {{:courses:smu:lecture_7.pdf |slides}} **Video:** [[https://www.dropbox.com/s/jcobatxq1eydluo/SMU%20Lecture%207%20-%20Exact%20Algorithms.mp4?dl=0|Exact Methods]], [[https://www.dropbox.com/s/rtngelczv76jgan/SMU%20Lecture%207%20-%20Approximate%20Methods.mp4?dl=0|Approximate Methods]] ---- ==== Lecture 8 - Natural Language Processing 1 ==== **Probabilistic models:** {{ :courses:smu:nlp1_1.pdf|slides}} **video:** [[ https://drive.google.com/file/d/11LJRDSXoChf43EnID0QvXQzk-sMm34Ao/view|google-drive]] [[https://www.dropbox.com/s/cgssi101sx5i8yw/NLP1.mp4?dl=0|dropbox]] ---- ==== Lecture 9 - Natural Language Processing 2 ==== **Vector models:** {{ :courses:smu:nlp_-_vector_models.pdf|slides}} **video:** [[ https://drive.google.com/file/d/12fHPPI2A6_xG9TyIVb9VPFamalwqkpFn/view|google-drive]] [[https://www.dropbox.com/s/ch23dw7ynt38dy9/NLP%20-%20vector%20models.mp4?dl=0|dropbox]] ---- ==== Lecture 10 - Natural Language Processing 3 ==== **Neural models:** {{ :courses:smu:nlp_export.pdf | slides}} **video:** [[ https://drive.google.com/file/d/12thZsnQCMu53gpx8-x8FsqOlfKiqtbXf/view?usp=sharing|google-drive]] [[https://www.dropbox.com/s/2nnbotv9vsfne01/NLP%20-%20neural%20models2.mp4?dl=0|dropbox]] ---- ==== Lecture 11 - Computational Learning Theory 1 ==== {{ :courses:smu:colt-1.pdf |COLT - lecture 1}} --- ==== Lecture 12 - Computational Learning Theory 2 ==== {{ :courses:smu:colt-2.pdf |COLT - lecture 2}} --- ==== Lecture 13 - Computational Learning Theory 3 ==== {{ :courses:smu:colt-3.pdf |COLT - lecture 3}} ---