Take AISafety.info’s 3 minute survey to help inform our strategy and priorities

Take the survey
Alignment research

Current techniques
Benchmarks and evals
Prosaic alignment
Interpretability
Agent foundations
Other alignment approaches
Organizations and agendas
Researchers

What is prosaic alignment?

Prosaic AI alignment is an approach to alignment research that assumes that future artificial general intelligence

(AGI) will be developed "prosaically" - i.e., without "reveal[ing] any fundamentally new ideas about the nature of intelligence or turn[ing] up any 'unknown unknowns.'" In other words, it assumes the AI techniques we're already using are sufficient to produce AGI if scaled far enough. Because of this assumption, prosaic alignment research is often relatively concrete and empirical, evaluating the effectiveness of proposed alignment techniques by trying them out on existing AI systems.

Some examples of prosaic alignment proposals are debate, imitating humans, preference learning, and iterated distillation and amplification.

Keep Reading

Continue with the next entry in "Alignment research"
What is "externalized reasoning oversight"?
Next
Or jump to a related question


AISafety.info

We’re a global team of specialists and volunteers from various backgrounds who want to ensure that the effects of future AI are beneficial rather than catastrophic.

© AISafety.info, 2022—2025

Aisafety.info is an Ashgro Inc Project. Ashgro Inc (EIN: 88-4232889) is a 501(c)(3) Public Charity incorporated in Delaware.