How can we keep AI within reasonable limits?
Our Security and Safety research focuses on ensuring that AI systems remain beneficial, controllable, and aligned with human values. We develop frameworks and mechanisms to prevent AI systems from causing unintended harm.
This critical research area addresses how to build robust safeguards, establish clear boundaries, and maintain human oversight over increasingly powerful AI systems.
Ensuring AI systems pursue intended goals and values
Identifying and mitigating potential AI-related risks
Formal methods for proving AI system safety
Maintaining meaningful human control over AI decisions