TY - CONF
T1 - DP-OPT
T2 - 12th International Conference on Learning Representations, ICLR 2024
AU - Hong, Junyuan
AU - Wang, Jiachen T.
AU - Zhang, Chenhui
AU - Li, Zhangheng
AU - Li, Bo
AU - Wang, Zhangyang
N1 - The work of Z. Wang is in part supported by the National Science Foundation under Grant IIS-2212176. This work is partially supported by the National Science Foundation under grant No. 1910100, No. 2046726, No. 2229876, DARPA GARD, the National Aeronautics and Space Administration (NASA) under grant No. 80NSSC20M0229, the Alfred P. Sloan Fellowship, the Michael Hammer Fellowship at MIT, and Princeton's Gordon Y. S. Wu Fellowship. Portions of this research were conducted with the advanced computing resources provided by Texas A&M High Performance Research Computing1, a composable computing cluster (He et al., 2023). We also want to thank anonymous reviewers for their constructive suggestions and comments.
PY - 2024
Y1 - 2024
N2 - Large Language Models (LLMs) have emerged as dominant tools for various tasks, particularly when tailored for a specific target by prompt tuning. Nevertheless, concerns surrounding data privacy present obstacles due to the tuned prompts' dependency on sensitive private information. A practical solution is to host a local LLM and optimize a soft prompt privately using data. Yet, hosting a local model becomes problematic when model ownership is protected. Alternative methods, like sending data to the model's provider for training, intensify these privacy issues facing an untrusted provider. In this paper, we present a novel solution called Differentially-Private Offsite Prompt Tuning (DP-OPT) to address this challenge. Our approach involves tuning a discrete prompt on the client side and then applying it to the desired cloud models. We demonstrate that prompts suggested by LLMs themselves can be transferred without compromising performance significantly. To ensure that the prompts do not leak private information, we introduce the first private prompt generation mechanism, by a differentially-private (DP) ensemble of in-context learning with private demonstrations. With DP-OPT, generating privacy-preserving prompts by Vicuna-7b can yield competitive performance compared to non-private in-context learning on GPT3.5 or local private prompt tuning. Codes are available at https://github.com/VITA-Group/DP-OPT.
AB - Large Language Models (LLMs) have emerged as dominant tools for various tasks, particularly when tailored for a specific target by prompt tuning. Nevertheless, concerns surrounding data privacy present obstacles due to the tuned prompts' dependency on sensitive private information. A practical solution is to host a local LLM and optimize a soft prompt privately using data. Yet, hosting a local model becomes problematic when model ownership is protected. Alternative methods, like sending data to the model's provider for training, intensify these privacy issues facing an untrusted provider. In this paper, we present a novel solution called Differentially-Private Offsite Prompt Tuning (DP-OPT) to address this challenge. Our approach involves tuning a discrete prompt on the client side and then applying it to the desired cloud models. We demonstrate that prompts suggested by LLMs themselves can be transferred without compromising performance significantly. To ensure that the prompts do not leak private information, we introduce the first private prompt generation mechanism, by a differentially-private (DP) ensemble of in-context learning with private demonstrations. With DP-OPT, generating privacy-preserving prompts by Vicuna-7b can yield competitive performance compared to non-private in-context learning on GPT3.5 or local private prompt tuning. Codes are available at https://github.com/VITA-Group/DP-OPT.
UR - http://www.scopus.com/inward/record.url?scp=85200542482&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85200542482&partnerID=8YFLogxK
M3 - Paper
AN - SCOPUS:85200542482
Y2 - 7 May 2024 through 11 May 2024
ER -