Thèse Prompts Préservant la Confidentialité pour les 'Large Language Models' H/F - Doctorat.Gouv.Fr
- CDD
- Doctorat.Gouv.Fr
Les missions du poste
Établissement : Institut Polytechnique de Paris École polytechnique École doctorale : Ecole Doctorale de l'Institut Polytechnique de Paris Laboratoire de recherche : Centre INRIA Saclay - Île-de-France Direction de la thèse : Catuscia PALAMIDESSI ORCID 0000000345977002 Début de la thèse : 2026-10-01 Date limite de candidature : 2026-08-31T23:59:59 Large Language Models (LLMs) such as ChatGPT [1] and LLaMA [2] have achieved considerable success and are now widely used across multiple domains. Excelling at generating human-like text and solving complex tasks, they have become central to the current AI landscape and are driving significant progress in both research and industry. Their widespread deployment underscores their transformative impact on both technology and society. However, such large adoption introduces privacy issues. Users typically interact with LLMs through prompts that may contain private and sensitive information, which can be collected and used by the LLM owner [3, 4]. Previous research has mainly focused on private training or fine-tuning of LLMs, which protects the training data of the LLM owner. In contrast, little attention has been given to the perspective of users who interact with trained or fine-tuned LLMs after deployment and share private information through their prompts. The general framework that provides theoretical privacy guarantees is Differential Privacy (DP) [5], which operates by adding random noise.To submit your candidature, please send an email to ****@****.** and to ****@****.** The pipeline for applying metric privacy to user prompts works as follows:
1. Embedding transformation: Given a prompt x, we first transform it into an embedding vector v = E(x) using a pretrained embedding model.
2. Noise addition: Noise is added directly in the embedding space according to a Laplace-like distribution consistent with metric privacy
3. Reconstruction: The noisy embedding is then decoded back into a sanitized prompt. This reconstruction process aims to preserve as much semantic content as possible while ensuring the targeted privacy guarantee.
4. Privacy budget management and reuse: Sanitized prompts may be reused or combined with information from other sanitized prompts. By the post-processing property of metric privacy, these operations do not affect privacy guarantees. In addition, a user-facing monitoring component will indicate when a cumulative privacy threshold is approached, helping users manage their interactions while maintaining privacy.
Challenges
1. Reconstruction from embeddings: While the pipeline transforms prompts to embeddings and back, only a few works have considered the reverse process from embedding space to discrete word space [7, 8]. Moreover, none have addressed reconstruction from a noisy embedding representation. It is therefore necessary to fine-tune the pretrained embedding model so that it can both encode prompts and reconstruct them from noisy embeddings while preserving as much semantic content as possible. Importantly, this fine-tuning should be performed on public data to avoid extra privacy costs.
2. Choice of distance metric: A central component of metric privacy is the distance function, which determines how much noise is injected and how privacy guarantees propagate through the embedding space. The metric must meaningfully reflect semantic similarity while matching the geometry of embeddings. A poorly chosen metric may either destroy utility (too much noise) or weaken privacy (too little noise). Identifying and validating an
appropriate distance is therefore a critical challenge.
3. Utility improvement: Even though metric privacy allows for a better optimization of the privacy-utility tradeoff compared to classical differential privacy, a noticeable utility gap remains. Noisy embeddings inevitably introduce semantic drift, and the reconstructed prompts may still lose important information. Improving utility thus requires exploring better noise distributions, improved decoding strategies, structure-aware distances, or regularization schemes that make the embedding space more robust to perturbations. Designing such enhancements while preserving formal privacy guarantees is a significant and open problem.
4. Privacy budget management: As users interact with the model, more sensitive information is shared, which increases the privacy budget and weakens privacy guarantees. A key challenge is how to optimally reuse sanitized prompts, either as they are or by combining information from multiple sanitized prompts, so as to maximize utility and reduce redundant noise injection, while ensuring that the cumulative privacy budget is not exceeded. Designing strategies for effective reuse and combination of sanitized prompts, and for alerting users when a threshold is approached, is critical for enabling more interactions with the LLM
without compromising privacy.
Le profil recherché
Master's in Computer Science or in Computer Engineering, and fluence in English.