Abstract
Providing privacy for natural language text data remains a largely open problem, despite its great practical importance. The current state of the art is manual redaction of sensitive words such as names, addresses etc. In this paper we propose viewing a corpus of text as a probability distribution over sequences of words. A sentence is then one realization from this distribution and redacting words changes the probability distribution. We use the Renyi-divergence divergence as a measure of the distance between two redacted datasets. We show that if enough words are redacted then sensitive redacted text can be made be statistically indistinguishable from non-sensitive redacted text. This can be used to develop efficient redaction strategies, that minimise the amount of redaction while meeting a privacy target.
This work was supported by Science Foundation Ireland grant 16/IA/4610.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Adding noise to an embedding perturbs it to nearby words, the way in which words are mapped to be close together (or far apart) therefore directly affects the output of the word-level DP sanitisation process.
- 2.
See Appendix https://anonymous.4open.science/r/appendix_repo-F4CC for more details.
- 3.
The choice of embedding will, in general, affect the estimated divergence. This can be mitigated by calculating the divergence for many different embeddings and using the worst-case (i.e. largest) value. However, we found the impact to be relatively minor in practice, see Section-6.3, and SentenceBERT [15] to work well.
- 4.
We select the range to be large enough that \(D_{\alpha }\) no longer increases as we increase \(\alpha \).
- 5.
- 6.
Data can be downloaded by following the instructions in the repository https://github.com/xuqiongkai/PATR.
- 7.
- 8.
- 9.
In particular, the DP analysis ignores correlations between the words in a sentence and so may greatly underestimate the information release. The impact of correlations on DP is well known and was first noted by [9].
- 10.
- 11.
Training code can be found at: https://github.com/pytorch/examples/tree/main/word_language_model.
- 12.
And one of the major deficiencies of all approaches tied to a single up front choice of embedding, such as word-level DP approaches.
- 13.
- 14.
The embedding vector of each word in a sentence is calculated, and the mean of these vectors is used as the sentence embedding.
References
Abadi, M., et al.: Deep learning with differential privacy. In: Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, pp. 308–318. CCS ’16, Association for Computing Machinery, New York, NY, USA (2016). https://doi.org/10.1145/2976749.2978318
Bosch, N., Crues, R., Shaik, N., Paquette, L.: Hello, [REDACTED]: protecting student privacy in analyses of online discussion forums. Grantee Submission (2020)
Brown, H., Lee, K., Mireshghallah, F., Shokri, R., Tramèr, F.: What does it mean for a language model to preserve privacy? In: 2022 ACM Conference on Fairness, Accountability, and Transparency, pp. 2280–2292. FAccT ’22, Association for Computing Machinery, New York, NY, USA (2022). https://doi.org/10.1145/3531146.3534642
Bun, M., Steinke, T.: Concentrated differential privacy: simplifications, extensions, and lower bounds (2016)
Chen, S., et al.: A customized text sanitization mechanism with differential privacy. In: Findings of the Association for Computational Linguistics: ACL 2023, pp. 5747–5758. Association for Computational Linguistics, Toronto, Canada (2023). https://doi.org/10.18653/v1/2023.findings-acl.355
Doudalis, S., Kotsogiannis, I., Haney, S., Machanavajjhala, A., Mehrotra, S.: One-sided differential privacy (2017)
Dwork, C., McSherry, F., Nissim, K., Smith, A.: Calibrating noise to sensitivity in private data analysis. In: Halevi, S., Rabin, T. (eds.) Theory of Cryptography, pp. 265–284. Springer, Berlin, Heidelberg (2006)
Feyisetan, O., Balle, B., Drake, T., Diethe, T.: Privacy- and utility-preserving textual analysis via calibrated multivariate perturbations. In: Proceedings of the 13th International Conference on Web Search and Data Mining, pp. 178–186. WSDM ’20, Association for Computing Machinery, New York, NY, USA (2020). https://doi.org/10.1145/3336191.3371856
Kifer, D., Machanavajjhala, A.: No free lunch in data privacy. In: Proceedings of the 2011 ACM SIGMOD International Conference on Management of Data, pp. 193–204. SIGMOD ’11, Association for Computing Machinery, New York, NY, USA (2011). https://doi.org/10.1145/1989323.1989345
Mattern, J., Weggenmann, B., Kerschbaum, F.: The limits of word level differential privacy. In: Findings of the Association for Computational Linguistics: NAACL 2022, pp. 867–881. Association for Computational Linguistics, Seattle, United States (2022). https://doi.org/10.18653/v1/2022.findings-naacl.65
Mironov, I.: Rényi differential privacy. In: 2017 IEEE 30th Computer Security Foundations Symposium (CSF). IEEE (2017). https://doi.org/10.1109/csf.2017.11
Murugadoss, K., et al.: Building a best-in-class automated de-identification tool for electronic health records through ensemble learning. medRxiv (2021). https://doi.org/10.1101/2020.12.22.20248270
Noshad, M., Moon, K.R., Sekeh, S.Y., Hero, A.O.: Direct estimation of information divergence using nearest neighbor ratios. In: 2017 IEEE International Symposium on Information Theory (ISIT). IEEE (2017). https://doi.org/10.1109/isit.2017.8006659
Pennington, J., Socher, R., Manning, C.D.: Glove: global vectors for word representation. In: Empirical Methods in Natural Language Processing (EMNLP), pp. 1532–1543 (2014). http://www.aclweb.org/anthology/D14-1162
Reimers, N., Gurevych, I.: Sentence-BERT: sentence embeddings using Siamese BERT-networks (2019)
Shi, W., Shea, R., Chen, S., Zhang, C., Jia, R., Yu, Z.: Just fine-tune twice: selective differential privacy for large language models. In: Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pp. 6327–6340. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates (2022). https://aclanthology.org/2022.emnlp-main.425
Shokri, R., Shmatikov, V.: Privacy-preserving deep learning. In: Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, pp. 1310–1321. CCS ’15, Association for Computing Machinery, New York, NY, USA (2015). https://doi.org/10.1145/2810103.2813687
Voigt, R., Jurgens, D., Prabhakaran, V., Jurafsky, D., Tsvetkov, Y.: RtGender: a corpus for studying differential responses to gender. In: Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018). European Language Resources Association (ELRA), Miyazaki, Japan (2018). https://aclanthology.org/L18-1445
Wen, Z., Lu, X.H., Reddy, S.: MeDAL: medical abbreviation disambiguation dataset for natural language understanding pretraining. In: Proceedings of the 3rd Clinical Natural Language Processing Workshop. Association for Computational Linguistics (2020). https://doi.org/10.18653/v1/2020.clinicalnlp-1.15
Yue, X., Du, M., Wang, T., Li, Y., Sun, H., Chow, S.S.M.: Differential privacy for text analytics via natural text sanitization. In: Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pp. 3853–3866. Association for Computational Linguistics (2021). https://doi.org/10.18653/v1/2021.findings-acl.337
Zhao, X., Li, L., Wang, Y.X.: Provably confidential language modelling. In: Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 943–955. Association for Computational Linguistics, Seattle, United States (2022). https://doi.org/10.18653/v1/2022.naacl-main.69
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2025 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Gusain, V., Leith, D. (2025). Plausible Deniability of Redacted Text. In: Garcia-Alfaro, J., et al. Computer Security. ESORICS 2024 International Workshops. ESORICS 2024. Lecture Notes in Computer Science, vol 15263. Springer, Cham. https://doi.org/10.1007/978-3-031-82349-7_4
Download citation
DOI: https://doi.org/10.1007/978-3-031-82349-7_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-82348-0
Online ISBN: 978-3-031-82349-7
eBook Packages: Computer ScienceComputer Science (R0)