Researching how AI systems can navigate complex ethical scenarios while respecting diverse human values.
Advancing AI aligned with diverse human values.
Join our mission to advance AI alignment research through pluralistic values.
Lead technical research on AI safety and alignment, developing novel approaches to value learning, preference modeling, and robustness in AI systems.
Build and maintain experimental infrastructure for AI safety research, focusing on scalable systems for testing alignment theories and value learning approaches.
Investigate foundational questions in AI alignment and safety, combining theoretical analysis with empirical insights to develop new frameworks for value learning and preference modeling.