TY - JOUR
T1 - Fine-grained Local Sensitivity Analysis of Standard Dot-Product Self-Attention
AU - Havens, Aaron
AU - Araujo, Alexandre
AU - Zhang, Huan
AU - Hu, Bin
N1 - A. Havens and B. Hu are generously supported by the AFOSR award FA9550-23-1-0732 and the NSF award CAREER-2048168. H. Zhang was supported by the AI2050 program at Schmidt Sciences and NSF SLES award (2331967).
PY - 2024
Y1 - 2024
N2 - Self-attention has been widely used in various machine learning models, such as vision transformers. The standard dot-product self-attention is arguably the most popular structure, and there is a growing interest in understanding the mathematical properties of such attention mechanisms. This paper presents a fine-grained local sensitivity analysis of the standard dot-product self-attention, leading to new non-vacuous certified robustness results for vision transformers. Despite the well-known fact that dot-product self-attention is not (globally) Lipschitz, we develop new theoretical analysis of Local Fine-grained Attention Sensitivity (LoFAST) quantifying the effect of input feature perturbations on the attention output. Our analysis reveals that the local sensitivity of dot-product self-attention to ℓ2 perturbations can actually be controlled by several key quantities associated with the attention weight matrices and the unperturbed input. We empirically validate our theoretical findings by computing non-vacuous certified ℓ2-robustness for vision transformers on CIFAR-10 and SVHN datasets. The code for LoFAST is available at https://github.com/AaronHavens/LoFAST.
AB - Self-attention has been widely used in various machine learning models, such as vision transformers. The standard dot-product self-attention is arguably the most popular structure, and there is a growing interest in understanding the mathematical properties of such attention mechanisms. This paper presents a fine-grained local sensitivity analysis of the standard dot-product self-attention, leading to new non-vacuous certified robustness results for vision transformers. Despite the well-known fact that dot-product self-attention is not (globally) Lipschitz, we develop new theoretical analysis of Local Fine-grained Attention Sensitivity (LoFAST) quantifying the effect of input feature perturbations on the attention output. Our analysis reveals that the local sensitivity of dot-product self-attention to ℓ2 perturbations can actually be controlled by several key quantities associated with the attention weight matrices and the unperturbed input. We empirically validate our theoretical findings by computing non-vacuous certified ℓ2-robustness for vision transformers on CIFAR-10 and SVHN datasets. The code for LoFAST is available at https://github.com/AaronHavens/LoFAST.
UR - http://www.scopus.com/inward/record.url?scp=85203790611&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85203790611&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85203790611
SN - 2640-3498
VL - 235
SP - 17680
EP - 17696
JO - Proceedings of Machine Learning Research
JF - Proceedings of Machine Learning Research
T2 - 41st International Conference on Machine Learning, ICML 2024
Y2 - 21 July 2024 through 27 July 2024
ER -