« Le Séminaire Palaisien » | Luiz Chamon & Charlotte Dion-Blanc
Chaque session du séminaire est divisée en deux présentations scientifiques de 40 minutes chacune : 30 minutes d’exposé et 10 minutes de questions. Luiz Chamon & Charlotte Dion-Blanc animeront la session de janvier 2026 !
Inscriptions gratuites mais obligatoires, dans la limite des places disponibles. Un buffet sera servi à l'issue du séminaire.
Machine learning (ML) and artificial intelligence (AI) now automate entire systems rather than individual tasks. As such, ML/AI models are no longer responsible for a single top-line metric (e.g., prediction accuracy), but must face a growing set of potentially conflicting system requirements, such as robustness, fairness, safety, and alignment with prior knowledge. These challenges are exacerbated in uncertain, data-driven settings and further complicated by the scale and heterogeneity of modern ML/AI applications that involve from static, discriminative models (e.g., neural network classifiers) to dynamic, generative models (e.g., Langevin diffusions used for sampling). This keynote defines WHAT constitutes a requirement and explains WHY incorporating them into learning is critical. It then shows HOW to do so using constrained learning and illustrates WHEN and WHERE this approach is effective by presenting use cases in ML for science, safe reinforcement learning, and sampling. Ultimately, this talk aims to show you (WHO) that constrained learning is key to building trustworthy ML/AI systems, enabling a shift from a paradigm of artificial intelligence that is supposed to implicitly emerge from data to one of engineered intelligence that explicitly does what we want.
In this talk, I will present some results on supervised classification for continuous-time processes. The idea is to be able to propose some consistent classifier, for different type of temporal data.
The methodology relies on mimicking the Bayes rule in the multiclass classification setting, in order to derive a classifier that is consistent as the number of labeled observations increases. I will first consider the case of repeated observations of paths of stochastic differential equations driven by Brownian motion, with different labels.
In this setting, I use a plug-in technique to estimate the conditional probability that the label equals k given the observation, and I will provide some cases in which the optimal rate of convergence is achieved. I will then move on to the case of interacting particle systems of McKean–Vlasov type. This framework generalizes the previous example and raises additional challenges.
The third case I will discuss concerns multivariate Hawkes processes with
K classes, distinguished by the parameters of their intensity functions. Here, the observations are event-time data. The multivariate Hawkes process is a nice way to model interactive agents for which some events are recorded continuously. To derive a consistent classification rule in this context, I will present an Empirical Risk Minimization strategy with a refitting step using a Lasso criterion.
The main novelty of the presented results is that they do not rely on the asymptotic properties of the underlying processes, and that the considered models are high-dimensional.