Options
2025
Conference Paper not in Proceedings
Title
LipShiFT: A Certifiably Robust Shift-based Vision Transformer
Title Supplement
Paper presented at the ICLR 2025 Workshop: VerifAI: AI Verification in the Wild, Singapore, April 27, 2025
Abstract
Deriving tight Lipschitz bounds for transformer-based architectures presents a significant challenge. The large input sizes and high-dimensional attention modules typically prove to be crucial bottlenecks during the training process and leads to sub-optimal results. Our research highlights practical constraints of these methods in vision tasks. We find that Lipschitz-based margin training acts as a strong regularizer while restricting weights in successive layers of the model. Focusing on a Lipschitz continuous variant of the ShiftViT model, we address significant training challenges for transformer-based architectures under norm-constrained input setting. We provide an upper bound estimate for the Lipschitz constants of this model using the l2 norm on common image classification datasets. Ultimately, we demonstrate that our method scales to larger models and advances the state-of-the-art in certified robustness for transformer-based architectures.
Author(s)
Open Access
File(s)
Rights
CC BY 4.0: Creative Commons Attribution
Language
English