Extracting Training Data from Large Language Models

Using GPT 2 model, they demonstrate that it is possible to have Training data leakage in language models.

Not all memorization is bad. Unintended “Eidetic memorization” of rare examples is a problem.

Yannic Kilcher’s explanation on YouTube


Training data leakage in language models < >