I’m a Postdoctoral Research Associate at The Alan Turing Institute and an Associate Member of the Department of Computer Science at the University of Oxford, where I’m mentored by Marta Kwiatkowska and Lukasz Szpruch. From 2021 to 2024, I completed my PhD at the University of Oslo supervised by Christos Dimitrakakis, and previously studied Mathematics (BSc, MSc).
My research interests are in reinforcement learning and related areas, including reward learning, preference learning, and the intersection of ML with game theory and mechanism design.
News
- 04/2025: The 2nd edition of last year’s workshop on Models of Human Feedback for AI Alignment will take place at ICML 2025! This year a special focus will be the role of AI Feedback for AI Alignment. Call for Papers will be out soon.
- 03/2025: We released three new preprints on Strategyproof RLHF, Causal Imitation Learning and Multi-Agent Cooperative RL.
- 11/2024: I’m visiting Haifeng Xu’s group at the University of Chicago and will give a talk on Strategic Interactive Decision-Making on December 5th at the CS Department.
- 07/2024: We’re organizing the ICML 2024 Workshop on Models of Human Feedback for AI Alignment. 09/2024: Recordings are now available here.
Selected Publications
Strategyproof Reinforcement Learning from Human Feedback [pdf]
Thomas Kleine Buening, Jiarui Gan, Debmalya Mandal, Marta Kwiatkowska
preprintStrategic Linear Contextual Bandits [pdf]
Thomas Kleine Buening, Aadirupa Saha, Christos Dimitrakakis, Haifeng Xu
NeurIPS 2024Environment Design for Inverse Reinforcement Learning [pdf]
Thomas Kleine Buening$^\star$, Victor Villin$^\star$, Christos Dimitrakakis
ICML 2024, Oral PresentationBandits Meet Mechanism Design to Combat Clickbait in Online Recommendation [pdf]
Thomas Kleine Buening, Aadirupa Saha, Christos Dimitrakakis, Haifeng Xu
ICLR 2024, Spotlight Presentation
You can reach me at tbuening@turing.ac.uk or thomas.kleinebuening@cs.ox.ac.uk.