1. Carlini
N.
Tramer
F.
Wallace
E.
Jagielski
M.
Herbert-Voss
A.
Lee
K.
... & Raffel
C. (2021). Extracting Training Data from Large Language Models. In 30th USENIX Security Symposium (USENIX Security 21). This paper empirically demonstrates that language models can memorize and regurgitate verbatim text sequences from their training data
including unique and private information
confirming the high risk of personal data disclosure.
2. National Institute of Standards and Technology (NIST). (2023). AI Risk Management Framework (AI RMF 1.0). In Section 4.2
"MAP
" the framework emphasizes identifying risks to individuals' rights and safety
stating
"AI systems can also affect privacy by
for example
re-identifying individuals from what was thought to be anonymized data..." This highlights privacy breaches as a primary risk category.
3. Stanford University. (2023). CS224N: NLP with Deep Learning
Lecture on Ethics in NLP. Course materials discuss the risks of training models on private data
noting that models can memorize sensitive information like names
phone numbers
and addresses
which can then be extracted
posing a direct privacy risk.