— Ch. 1 · Origins And Historical Context —
Fairness (machine learning).
~5 min read · Ch. 1 of 5
The year 1964 marked a turning point in American history when the U.S. Civil Rights Act passed into law. This legislation sparked intense debate within the scientific community about how to measure fairness and bias in decision-making processes. Researchers spent the next decade trying to define what it meant for an algorithm or human judge to be fair. By the end of the 1970s, these discussions had largely vanished from mainstream academic discourse. The competing notions of fairness left little room for clarity on which definition should take precedence. A sharp resurgence occurred only after 2016 when ProPublica released a report claiming that COMPAS software used in US courts was racially biased. This report reignited decades-old questions about whether automated systems could ever truly be impartial. Tech companies began releasing tools like IBM's Python libraries and Google's guidelines to detect and reduce bias. Facebook introduced Fairness Flow as a tool to identify discrimination in their AI programs. Critics argued these efforts remained insufficient because employees rarely used them and the tools did not cover all programs.
Language And Cultural Bias
ChatGPT describes liberalism from an Anglo-American perspective while ignoring valid viewpoints from other cultures. When queried about political ideologies, the model emphasizes human rights and equality but omits aspects like opposing state intervention found in Vietnamese perspectives. It also fails to mention limitations of government power prevalent in Chinese thought. Japanese, Korean, French, and German corpora remain absent from its responses despite being multilingual chatbots. Luo et al. demonstrated that current large language models present Anglo-American views as truth systematically. They downplay non-English perspectives as irrelevant or noise during training processes. English-centric data causes this systematic deviation in sampling information. The true coverage of topics and views available in repositories gets distorted by this approach. A query about liberalism yields answers reflecting only one cultural framework. Other political perspectives embedded in different languages disappear entirely from generated text. This blindness extends beyond politics into gender roles where nurses are associated with women and engineers with men. LinkedIn profiles analyzed through Natural Language Processing methods reveal similar patterns of exclusion.