Reinforcement Learning and Bandit Algorithms Joint Reading Group

Jan-July 2024

In this reading group we will primarily read the lecture notes Foundations of Reinforcement Learning and Interactive Decision Making by Dylan J. Foster and Alexander Rakhlin as the learning material. We refer to the book Bandit Algorithms by Tor Lattimore and Csaba Szepesvari for more detailed background knowledge.

This joint reading group is hybrid - we meet biweekly on Friday via Zooms at:

Timetable (provisional)

Please check this website regularly for the most up-to-date arrangement.

Date Topic Presenter Discussant
1 12 Jan Introduction and basic technical tools Chao Wen-Xin
2 26 Jan Multi-Armed Bandits (part 1) Myeonghun Chao
3 09 Feb Multi-Armed Bandits (part 2) / Contextual Bandits Myeonghun & Wen-Xin Rick
4 23 Feb Linear Stochastic Bandits / Bandits with Sparsity (part 1) Wen-Xin & Siyu Rick
5 08 Mar Bandits with Sparsity (part 2)/ Online Decision Making with High-Dimensional Covariates Siyu Wen-Xin
6 22 Mar Dynamic Batch Learning in High-Dimensional Sparse Linear Contextual Bandits Peiyao Wen-Xin
7 05 Apr Structured Bandits (part 1) Baiyu Chao
8 19 Apr Structured Bandits (part 2) Chao Baiyu
9 03 May Reinforcement Learning: Basics (Chapter 5) Chaowen Chaowen
10 7 June General Decision Marking Rick Chaowen
11 21 June Online Policy Learning and Inference by Matrix Completion Peiyao Chao
12 12 July Reinforcement Learning: Function Approximation and Large State Spaces Huaning Wen-Xin

Materials

Group Members

Southampton

Chicago/San Diego

Michigan

Colorado

Other Supplementary References

Roles of Presenter and Discussant

If it is your first time attend a reading group, you might find the reading group tips by Lester Mackey and Percy Liang helpful.

Every time we will have one people (presenter) present the main contents and another people (discussant) raise questions and lead the discussion.

Before each session, although not compulsory I would recommend following amount of time spent on reading:


Webpage maintained by Chao Zheng. Last updated on 06/01/2024