Théo Vincent

Théo Vincent joined the Intelligent Autonomous Systems Group as a Ph.D. student in December 2022. He is currently working on off-policy Reinforcement Learning methods. He is part of the group SAIROL, a DFKI unit, and works under the supervision of Boris Belousov and Jan Peters.

During his studies, Théo focused on Deep Learning, (3D) Computer Vision and Reinforcement Learning. He graduated from MVA, a double master's degree program in collaboration with ENS ParisSaclay and Ponts ParisTech. He did his master's thesis at IAS on value-based methods in offline settings supersived by Carlo D’Eramo.

Before his Ph.D., Théo worked in a Parisian lab called Saint-Venant lab, supervised by Rémi Carmigniani to track professional swimmers on video clips using Deep Learning. He then joined Signality, a Swedish start-up led by Mikael Rousson, to investigate the problem of finding a homography linking a football pitch to a camera. He also worked in the group of Chirag Patel at Harvard Medical School to understand how our organs are aging.

Supervision

  • HiWi, Yogesh Tripathi; SlimRL: a simple minimal library for off-policy Reinforcement Learning.
  • IP, Yogesh Tripathi; LEM: Learned Ensemble Mixture.
  • Master Thesis, Fabian Wahren (with Boris Belousov); Adapt your network: Investigating neural network’s architecture in Q-learning methods.

Teaching

🎮 Reinforcement Learning, Summer - 2024
📊 Statistical Machine Learning, Summer - 2023 & 2024

Reviewing

IROS, RLC, ICLR
AutoRL@ICML (Outstanding reviewer)

Talks

11/2024 – at Uni Mannheim , Mannheim 🇩🇪
hosted by Leif Döring
09/2024 – at Naver Labs , Grenoble 🇫🇷
hosted by Jean-Michel Renders
05/2024 – at INSAIT, Sofia 🇧🇬
hosted by Danda Paudel
04/2023 – at Lite RL, Würzburg 🇩🇪
hosted by Carlo D'Eramo

Publications

    •   Bib
      Vincent, T.; Faust, T.; Tripathi, Y.; Peters, J.; D'Eramo, C. (2025). Eau De Q-Network: Adaptive Distillation of Neural Networks in Deep Reinforcement Learning, Conference on Reinforcement Learning and Decision Making (RLDM).
    •     Bib
      Vincent, T.; Palenicek, D.; Belousov, B.; Peters, J.; D'Eramo, C. (2025). Iterated Q-Network: Beyond One-Step Bellman Updates in Deep Reinforcement Learning, Transactions on Machine Learning Research (TMLR).
    •     Bib
      Vincent, T.; Wahren, F.; Peters, J.; Belousov, B.; D'Eramo, C. (2025). Adaptive Q-Network: On-the-fly Target Selection for Deep Reinforcement Learning, International Conference on Learning Representations (ICLR).
    •     Bib
      Vincent, T.; Metelli, A.; Belousov, B.; Peters, J.; Restelli, M.; D'Eramo, C. (2024). Parameterized Projected Bellman Operator, Proceedings of the National Conference on Artificial Intelligence (AAAI).