1. O'Neil
C. (2016). Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Crown. (While a book
its concepts are foundational in university courses on AI ethics). Chapter 1 discusses how zip codes are used as proxies for race
leading to discriminatory outcomes in models for insurance
credit
and parole.
2. National Institute of Standards and Technology (NIST). (2023). AI Risk Management Framework (AI RMF 1.0). NIST AI 100-1. Section 4.3.2
"Systemic and Human Biases
" discusses how proxy variables that correlate with protected classes can introduce harmful bias into AI systems.
3. Barocas
S.
& Selbst
A. D. (2016). Big Data's Disparate Impact. California Law Review
104(3)
671-732. Page 677 discusses how seemingly neutral variables
such as zip codes
can function as proxies for protected classes like race
leading to discriminatory outcomes that may violate anti-discrimination laws.
4. Hardt
M. (2018). CS 294: Fairness in Machine Learning. University of California
Berkeley. Lecture 1 notes discuss the sources of bias and discrimination in machine learning
including the use of proxy variables like zip codes in credit scoring as a classic example of redlining. Available from UC Berkeley course materials.