Publications

Guided Flow Policy: Learning from High-Value Actions in Offline Reinforcement Learning

Franki Nguimatsia Tiofack*, Théotime Le Hellard*, Fabian Schramm*, Nicolas Perrin-Gilbert, Justin Carpentier
* Equal contribution

Published in Fourteenth International Conference on Learning Representations (ICLR), 2026

We introduce Guided Flow Policy (GFP), which couples a multi-step flow-matching policy with a distilled one-step actor to focus on learning from high-value actions in offline reinforcement learning.

First-order Sobolev Reinforcement Learning

Fabian Schramm, Nicolas Perrin-Gilbert, Justin Carpentier

Published in Differentiable Systems and Scientific Machine Learning, EurIPS 2025, 2025

We propose a refinement of temporal-difference learning that enforces first-order Bellman consistency by training the value function to match both Bellman targets and their derivatives.

Reference-Free Sampling-Based Model Predictive Control

Fabian Schramm, Pierre Fabre, Nicolas Perrin-Gilbert, Justin Carpentier

Published in IEEE International Conference on Robotics and Automation (ICRA), 2025

We present a sampling-based model predictive control framework that enables emergent locomotion without relying on handcrafted gait patterns or predefined contact sequences.

Leveraging Randomized Smoothing for Optimal Control of Nonsmooth Dynamical Systems

Quentin Le Lidec, Fabian Schramm, Louis Montaut, Cordelia Schmid, Ivan Laptev, Justin Carpentier

Published in Nonlinear Analysis: Hybrid Systems, International Federation of Automatic Control (IFAC) journal, 2024

This paper presents randomized smoothing to tackle non-smoothness issues commonly encountered in optimal control and provides key insights on the interplay between Reinforcement Learning and Optimal Control.

Reactive Stepping for Humanoid Robots using Reinforcement Learning: Application to Standing Push Recovery on the Exoskeleton Atalante

Alexis Duburcq, Fabian Schramm, Guilhem Boéris, Nicolas Bredeche, Yann Chevaleyre

Published in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022

Video

This paper presents a reinforcement learning framework capable of learning robust standing push recovery for bipedal robots with a smooth out-of-the-box transfer to reality, requiring only instantaneous proprioceptive observations.