research
Our research spans six key areas in AI security.
AI Agent Security
Securing autonomous AI agents against adversarial manipulation, prompt injection, and unintended behaviors in real-world deployments.
Interpretable AI Security
Leveraging interpretability and explainability techniques to understand, diagnose, and mitigate vulnerabilities in AI systems.
Usable Security of AI
Designing intuitive security mechanisms and interfaces that help users safely interact with and configure AI systems.
AI Misuse Measurement
Developing methods to quantify and detect malicious applications of AI, from deepfakes to automated cyber attacks.
AI Society Security
Investigating the societal impacts of AI security risks, including misinformation, surveillance, and governance challenges.
AI for Security
Applying AI techniques to strengthen cybersecurity defenses, including threat detection, vulnerability analysis, and automated response.