Alignment Wiki is a comprehensive encyclopedia of AI alignment research, theories, organizations, and key figures. Our goal is to provide neutral, well-sourced information about the field of AI alignment.
As AI systems become more capable, ensuring they remain beneficial and aligned with human values becomes increasingly important. Alignment Wiki aims to make the field more accessible by documenting:
Alignment Wiki strives to be neutral and encyclopedic. We present different perspectives in the field fairly and avoid advocating for particular approaches. All content should be well-sourced and verifiable.
Alignment Wiki is a moderated wiki. Anyone can suggest edits, which are reviewed by moderators before publication. This approach maintains quality while allowing broad participation.
To contribute:
For questions, suggestions, or issues, please reach out through GitHub or submit feedback via the edit suggestion system.
Content on Alignment Wiki is available under the Creative Commons Attribution-ShareAlike 4.0 International License (CC BY-SA 4.0).