How can I use a background in the social sciences to help with AI alignment?
Nora Ammann, in the post AI alignment
Some examples of approaches, taken from the post:
-
Steve Byrnes’s research on brain-like AGI
safety asks how we can align artificial general intelligence if it’s built on the same principles as the human brain, drawing analogies with neuroscience. -
John Wentworth studies agent-like systems in nature to understand agency in general.
-
Andrew Critch’s research on multipolar takeoffs and “robust agent-agnostic processes” relates to concepts from sociology.
-
Discussions of mesa-optimization use human evolution as a source of analogies.
Principles of Intelligent Behavior in Biological and Social Systems (PIBBSS) is a group that runs a summer research fellowship and has recommendations for books and videos.
Other such research agendas exist. You can consider these as examples of what alignment-relevant research with varying amounts of math and computer science could look like:
-
An Open Agency Architecture for Safe Transformative AI is an AI alignment paradigm aimed at ending the acute risk period without creating worse risks.
-
Learning Normativity: A Research Agenda aims to develop ways for agents
to learn norms like languages and values in the absence of perfect feedback.AgentView full definitionA system that can be understood as taking actions towards achieving a goal.
-
What Should AI Owe To Us? Accountable and Aligned AI Systems via Contractualist AI Alignment tries to ground AI alignment in pluralist and contractualist norms.
-
Political Economy of Reinforcement Learning (PERLS) is a workshop studying the societal implications of reinforcement learning systems.
-
The Alignment of Complex Systems Research Group studies connections between AI alignment and complex systems theory.
See also the EA Forum post Social scientists interested in AI safety A research field about how to prevent risks from advanced artificial intelligence.