Explainability of human ethical algorithms
October 10, 2023 — October 10, 2023
bounded compute
collective knowledge
cooperation
culture
ethics
gene
incentive mechanisms
institutions
machine learning
neuron
sparser than thou
utility
wonk
Can we explain the ethical algorithms that humans run, in the same way that we attempt to explain black-box ML algorithms?
I’m thinking of things like influence functions, which determine what examples we reason from, as in (Grosse et al. 2023), but also reasoning from features would be interesting.
1 Disgust
Interesting starting point. See disgust.
2 Theory of moral sentiments
e.g. Haidt (2013).
3 References
DellaPosta, Shi, and Macy. 2015. “Why Do Liberals Drink Lattes?” American Journal of Sociology.
Feinberg, Antonenko, Willer, et al. 2014. “Gut Check: Reappraisal of Disgust Helps Explain Liberal–Conservative Differences on Issues of Purity.” Emotion.
Grosse, Bae, Anil, et al. 2023. “Studying Large Language Model Generalization with Influence Functions.”
Haidt. 2013. The Righteous Mind: Why Good People Are Divided by Politics and Religion.
Inbar, Pizarro, and Bloom. 2009. “Conservatives Are More Easily Disgusted Than Liberals.” Cognition and Emotion.
Inbar, Pizarro, Iyer, et al. 2012. “Disgust Sensitivity, Political Conservatism, and Voting.” Social Psychological and Personality Science.
Moral Sentiments and Material Interests: The Foundations of Cooperation in Economic Life. 2006.
Storr. 2021. The Status Game: On Human Life and How to Play It.