
Activez les alertes d’offres d’emploi par e-mail !
Générez un CV personnalisé en quelques minutes
Décrochez un entretien et gagnez plus. En savoir plus
A leading research institute in France is seeking a postdoctoral researcher to develop theoretical frameworks for analyzing generalization and memorization in diffusion models. The ideal candidate should have strong English proficiency, coding skills in Python, and some knowledge of French. The position offers competitive working conditions including partial reimbursement for transport costs and the possibility of teleworking.
Inria, the French national research institute for the digital sciences.
The position will be in the framework of the ERC Starting Grant DYNASTY (Dynamics‑Aware Theory of Deep Learning).
The position might include traveling to conferences for paper presentation. Travel expenses will be covered within the limits of the scale in force.
The rapid success of diffusion models [1, 2, 3], which have achieved state‑of‑the‑art performance across diverse domains, motivates the need for a theoretical understanding of the mechanisms underpinning their strong capabilities. The unique structure of these models, as well as recent evidence [4, 5] indicating they forgo the advantageous properties of benign overfitting, suggests that diffusion models are a fundamental divergence from traditional deep learning paradigms. This suggests that existing generalisation theories are insufficient and highlights the need for a bespoke, algorithm‑dependent framework to capture the phenomena present in these models.
We are seeking a postdoctoral researcher to develop theoretical frameworks for analysing generalisation and memorisation in diffusion models. The project's central goal is to move beyond algorithm‑independent bounds and develop rigorous theory that unpacks the generalisation properties inherent to the training and sampling processes. A key component of this analysis will be to precisely characterise the mechanisms driving memorisation. This theoretical work will serve as the foundation for developing well‑founded algorithms that target the task of preventing data copying (e.g. [6, 7]) as well as the problem of data attribution (e.g. [8, 9]), allowing for the precise measurement of individual training examples' influence on model outputs.