research

Our research spans six key areas in AI security.

AI Agent Security

Securing autonomous AI agents against adversarial manipulation, prompt injection, and unintended behaviors in real-world deployments.


Interpretable AI Security

Leveraging interpretability and explainability techniques to understand, diagnose, and mitigate vulnerabilities in AI systems.


Usable Security of AI

Designing intuitive security mechanisms and interfaces that help users safely interact with and configure AI systems.


AI Misuse Measurement

Developing methods to quantify and detect malicious applications of AI, from deepfakes to automated cyber attacks.


AI Society Security

Investigating the societal impacts of AI security risks, including misinformation, surveillance, and governance challenges.


AI for Security

Applying AI techniques to strengthen cybersecurity defenses, including threat detection, vulnerability analysis, and automated response.