AI-driven security and software engineering

Modern society depends on critical infrastructure run by large, complex software, but software is expensive to develop and difficult to verify. AI will change how we build, verify, and maintain software. Towards this goal, we are making progress on:

Trustworthy AI

While machine learning has made great strides, it isn't ready for many scenarios because of concerns over its security, privacy, and robustness. We are making progress towards trustworthy AI, including:

  • Attack against safety-aligned LLMs (NeurIPS 2024)
  • A benchmark for evaluating transfer-based attacks (NeurIPS 2023)
  • Improving adversarial transferability via intermediate-level perturbation decay (NeurIPS 2023)
  • Use squeeze training for improving adversarial robustness (ICLR 2023)
  • Attack Bayesian models to improve transferability of adversarial examples (ICLR 2023)