TY - GEN
T1 - Manipulating Machine Learning
T2 - 39th IEEE Symposium on Security and Privacy, SP 2018
AU - Jagielski, Matthew
AU - Oprea, Alina
AU - Biggio, Battista
AU - Liu, Chang
AU - Nita-Rotaru, Cristina
AU - Li, Bo
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2018/7/23
Y1 - 2018/7/23
N2 - As machine learning becomes widely used for automated decisions, attackers have strong incentives to manipulate the results and models generated by machine learning algorithms. In this paper, we perform the first systematic study of poisoning attacks and their countermeasures for linear regression models. In poisoning attacks, attackers deliberately influence the training data to manipulate the results of a predictive model. We propose a theoretically-grounded optimization framework specifically designed for linear regression and demonstrate its effectiveness on a range of datasets and models. We also introduce a fast statistical attack that requires limited knowledge of the training process. Finally, we design a new principled defense method that is highly resilient against all poisoning attacks. We provide formal guarantees about its convergence and an upper bound on the effect of poisoning attacks when the defense is deployed. We evaluate extensively our attacks and defenses on three realistic datasets from health care, loan assessment, and real estate domains.
AB - As machine learning becomes widely used for automated decisions, attackers have strong incentives to manipulate the results and models generated by machine learning algorithms. In this paper, we perform the first systematic study of poisoning attacks and their countermeasures for linear regression models. In poisoning attacks, attackers deliberately influence the training data to manipulate the results of a predictive model. We propose a theoretically-grounded optimization framework specifically designed for linear regression and demonstrate its effectiveness on a range of datasets and models. We also introduce a fast statistical attack that requires limited knowledge of the training process. Finally, we design a new principled defense method that is highly resilient against all poisoning attacks. We provide formal guarantees about its convergence and an upper bound on the effect of poisoning attacks when the defense is deployed. We evaluate extensively our attacks and defenses on three realistic datasets from health care, loan assessment, and real estate domains.
KW - adversarial machine learning
KW - poisoning attacks
KW - robust linear regression
UR - http://www.scopus.com/inward/record.url?scp=85050640069&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85050640069&partnerID=8YFLogxK
U2 - 10.1109/SP.2018.00057
DO - 10.1109/SP.2018.00057
M3 - Conference contribution
AN - SCOPUS:85050640069
T3 - Proceedings - IEEE Symposium on Security and Privacy
SP - 19
EP - 35
BT - Proceedings - 2018 IEEE Symposium on Security and Privacy, SP 2018
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 21 May 2018 through 23 May 2018
ER -