Challenges in Augmenting Large Language Models with Private Data

Pre-FPO Presentation
Aug 17, 2023, 9:00 am10:30 am
Zoom (see abstract)



Event Description

Quantifying privacy risks in LLMs is an important research question. We take a step towards answering this question by defining a real-world threat model wherein an entity seeks to augment a foundation model LLM with private data they possess, via methods such as fine-tuning or in-context learning. The entity also seeks to improve the quality of their LLM outputs over time by learning from human feedback. We examine attacks on this system in the form of poisoning attacks, and defenses in the form of systems that provide differential privacy guarantees.


Adviser: Prateek Mittal