TOUTELIA 2021 : Statistical Physics, Probability and AI
de
jeudi 16 décembre 2021 (09:00)
à
vendredi 17 décembre 2021 (15:00)
lundi 13 décembre 2021
mardi 14 décembre 2021
mercredi 15 décembre 2021
jeudi 16 décembre 2021
14:00
Talk by Yann Ollivier
Talk by Yann Ollivier
14:00 - 15:30
This is the conference talk. Yann Ollivier will also give a colloquim talk the next day. Abstract: Markov decision processes are a model for several artificial intelligence problems, such as games (chess, Go...) or robotics. At each timestep, an agent has to choose an action, then receives a reward, and then the agent's environment changes (deterministically or stochastically) in response to the agent's action. The agent's goal is to adjust its actions to maximize its total reward. In principle, the optimal behavior can be obtained by dynamic programming or optimal control techniques, although practice is another story. Here we consider a more complex problem: learn all optimal behaviors for all possible reward functions in a given environment. Ideally, such a "controllable agent" could be given a description of a task (reward function, such as "you get +10 for reaching here but -1 for going through there") and immediately perform the optimal behavior for that task. This requires a good understanding of the mapping from a reward function to the associated optimal behavior. We will present our recent theoretical and empirical results in this direction. There exists a particular "map" of a Markov decision process, on which near-optimal behaviors for all reward functions can be read directly by an algebraic formula. Moreover, this "map" is learnable by standard deep learning techniques from random interactions with the environment.
15:30
Talk by Sandrine Péché
Talk by Sandrine Péché
15:30 - 16:30
We consider a random matrix model M = YY∗ where Y = (f(WX)_ij) and W and X are large rectangular matrices with iid entries.The function f is called the activation function in certain neural networks. Pennington and Worah have identified the empirical eigenvalue distribution of such random matrices in the Gaussian case (W and X). We extend their result to a wider class of distributions for a certain class of activation functions. This is joint work with Lucas Benigni.
vendredi 17 décembre 2021
09:00
Talk by Franck Gabriel
Talk by Franck Gabriel
09:00 - 10:30
In recent years, randomness has become more important in machine learning. Through two examples, we will see that it can be used to [1] "select" well-behaved regions of parameters and [2] provide an easier optimization problem. [1] In deep learning, I will present the NTK regime, where, by considering a "wide" random initialization, it can be shown that neural networks with large width converge, and the dynamics of the output function can be described. [2] In kernel methods, instead of looking for an optimal function in the RKHS, one can look for an optimal function in a random vector space: this is the random feature method. After explaining why it provides an approximation to kernel methods, I will present the implicit bias that finite sampling induces on the output function.
10:30
Talk by Marc Lelarge
Talk by Marc Lelarge
10:30 - 12:00
Abstract: Geometric deep learning is an attempt for geometric unification of a broad class of machine learning problems from the perspectives of symmetry and invariance. In this talk, I will present some advances of geometric deep learning applied to combinatorial structures. I will focus on various classes of graph neural networks that have been shown to be successful in a wide range of applications with graph structured data.
12:00
Lunch break
Lunch break
12:00 - 14:00
14:00
COLLOQUIUM by Yann Ollivier
COLLOQUIUM by Yann Ollivier
14:00 - 15:00
Résumé : "Les problèmes de raisonnement inductif ou d'extrapolation comme "deviner la suite d'une série de nombres", ou plus généralement, "comprendre la structure cachée dans des observations", sont fondamentaux si l'on veut un jour construire une intelligence artificielle. On a parfois l'impression que ces problèmes ne sont pas mathématiquement bien définis. Or il existe une théorie mathématique rigoureuse du raisonnement inductif et de l'extrapolation, basée sur la théorie de l'information. Cette théorie est très élégante, mais difficile à appliquer. En pratique aujourd'hui, ce sont les réseaux de neurones qui donnent les meilleurs résultats sur toute une série de problèmes concrets d'induction et d'apprentissage (vision, reconnaissance de la parole, récemment le jeu de Go ou les voitures sans pilote...) Je ferai le point sur quelques-uns des principes mathématiques sous-jacents et sur leur lien avec la théorie de l'information."