The Secure and Private AI (SPY) Lab conducts research on the security, privacy and trustworthiness of machine learning systems.
We often approach these problems from an adversarial perspective, by designing attacks that probe the worst-case performance of a system to ultimately understand and improve its safety.
Jun 5, 2024 | Our papers Stealing part of a production language model and Position: Considerations for Differentially Private Learning with Large-Scale Public Pretraining obtained oral presentations at ICML 2024. |
---|---|
Apr 5, 2024 | The paper Evading Black-box Classifiers Without Breaking Eggs has been awarded as Distinguished Paper Runner-Up at IEEE SaTML 2024. |
Mar 12, 2024 | We have reverse-engineered the (secret) Claude 3 tokenizer by inspecting the generation stream. Check our blog post, code and Twitter thread. |
Feb 22, 2024 | Lukas Fluri was awarded an ETH Medal for his Master’s Thesis “Evaluating Superhuman Models with Consistency Checks”. Congrats! |