MARLEL

Mathematics of Reinforcement Learning and ExperimentaL sciences

Exploring the mathematics of adaptive experimentation.

The Mathematics of Reinforcement Learning and ExperimentaL sciences (MARLEL) group is an international and interdisciplinary research network that has grown organically from long-term collaborations between co-authors and colleagues. While not formally affiliated with a specific institution, the group brings together researchers who share a common vision and scientific affinity, and who have been working together over the years on advancing both the theory and the practice of reinforcement learning.

As its name suggests, the group’s central focus is to advance the core mathematical foundations of Reinforcement Learning, with particular attention to the interplay between sequential decision theory, optimization, and statistical learning.

Beyond its theoretical dimension, the group also promotes cross-disciplinary dialogue between reinforcement learning and the experimental sciences, in particular life sciences—such as agriculture, biology, and medicine—where learning-by-interaction paradigms naturally align with the experimental process itself. Through these connections, the group aims to develop methods that are both mathematically rigorous and scientifically impactful.

From a broader philosophical standpoint, the group views Reinforcement Learning as the mathematics of interaction and decision-making, a fundamental complement to the generative side of Artificial Intelligence. Whereas Generative AI focuses on modeling and creating data from existing knowledge, Decisional AI, embodied by reinforcement learning, addresses the active process of exploring, experimenting, and optimizing in uncertain environments. This makes it a natural theoretical counterpart to experimental science itself, where knowledge emerges from iterative cycles of hypothesis, intervention, and observation.

By deepening the mathematical understanding of these adaptive processes, the group seeks to build bridges between formal theory and empirical inquiry, fostering a view of intelligence not only as a capacity to infer from data, but as an ability to learn through action: to decide, test, and evolve knowledge through experimentation.

Collaborators


Research topics

This group targets producing high-quality contributions on theoretical, algorithmic, and applied aspects of online learning, bandit theory, and reinforcement learning, with a particular emphasis on models that incorporate structure, hierarchy, uncertainty, or complex feedback, in order to advance our understanding of how agents can learn efficiently from sequential interaction, limited information, or structured environments. Our main focus of interests include, but are not limited to, the following list of topics:

1. Foundations of Reinforcement Learning Theory

2. Online Learning and Bandit Theory

3. Offline, Continual, and Transfer Reinforcement Learning

4. Planning, Hierarchical Models, and Representation Learning

5. Robustness, Safety, and Privacy

6. Learning in Structured or Application-Driven Domains


Publications of the group