TY - GEN
T1 - Learning and Evaluating a Differentially Private Pre-trained Language Model
AU - Hoory, Shlomo
AU - Feder, Amir
AU - Tendler, Avichai
AU - Cohen, Alon
AU - Erell, Sofia
AU - Laish, Itay
AU - Nakhost, Hootan
AU - Stemmer, Uri
AU - Benjamini, Ayelet
AU - Hassidim, Avinatan
AU - Matias, Yossi
N1 - Publisher Copyright:
© 2021 Association for Computational Linguistics.
PY - 2021/1/1
Y1 - 2021/1/1
N2 - Contextual language models have led to significantly better results, especially when pretrained on the same data as the downstream task. While this additional pre-training usually improves performance, it can lead to information leakage and therefore risks the privacy of individuals mentioned in the training data. One method to guarantee the privacy of such individuals is to train a differentially-private language model, but this usually comes at the expense of model performance. Also, in the absence of a differentially private vocabulary training, it is not possible to modify the vocabulary to fit the new data, which might further degrade results. In this work we bridge these gaps, and provide guidance to future researchers and practitioners on how to improve privacy while maintaining good model performance. We introduce a novel differentially private word-piece algorithm, which allows training a tailored domain-specific vocabulary while maintaining privacy. We then experiment with entity extraction tasks from clinical notes, and demonstrate how to train a differentially private pre-trained language model (i.e., BERT) with a privacy guarantee of ? = 1.1 and with only a small degradation in performance. Finally, as it is hard to tell given a privacy parameter ? what was the effect on the trained representation, we present experiments showing that the trained model does not memorize private information.
AB - Contextual language models have led to significantly better results, especially when pretrained on the same data as the downstream task. While this additional pre-training usually improves performance, it can lead to information leakage and therefore risks the privacy of individuals mentioned in the training data. One method to guarantee the privacy of such individuals is to train a differentially-private language model, but this usually comes at the expense of model performance. Also, in the absence of a differentially private vocabulary training, it is not possible to modify the vocabulary to fit the new data, which might further degrade results. In this work we bridge these gaps, and provide guidance to future researchers and practitioners on how to improve privacy while maintaining good model performance. We introduce a novel differentially private word-piece algorithm, which allows training a tailored domain-specific vocabulary while maintaining privacy. We then experiment with entity extraction tasks from clinical notes, and demonstrate how to train a differentially private pre-trained language model (i.e., BERT) with a privacy guarantee of ? = 1.1 and with only a small degradation in performance. Finally, as it is hard to tell given a privacy parameter ? what was the effect on the trained representation, we present experiments showing that the trained model does not memorize private information.
UR - http://www.scopus.com/inward/record.url?scp=85128307387&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85128307387
T3 - Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021
SP - 1178
EP - 1189
BT - Findings of the Association for Computational Linguistics, Findings of ACL
A2 - Moens, Marie-Francine
A2 - Huang, Xuanjing
A2 - Specia, Lucia
A2 - Yih, Scott Wen-Tau
PB - Association for Computational Linguistics (ACL)
T2 - 2021 Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021
Y2 - 7 November 2021 through 11 November 2021
ER -