The Unreasonable Ineffectiveness of Nucleus Sampling on Mitigating Text Memorization
Borec, Luka and Sadler, Philipp and Schlangen, David
This work analyses the text memorization behavior of large language models (LLMs) when subjected to nucleus sampling. Stochastic decoding methods like nucleus sampling are typically applied to overcome issues such as monotonous and repetitive text generation, which are often observed with maximization-based decoding techniques. We hypothesize that nucleus sampling might also reduce the occurrence of memorization patterns, because it could lead to the selection of tokens outside the memorized sequence. To test this hypothesis we create a diagnostic dataset with a known distribution of duplicates that gives us some control over the likelihood of memorisation of certain parts of the training data. Our analysis of two GPT-Neo models fine-tuned on this dataset interestingly shows that (i) an increase of the nucleus size reduces memorization only modestly, and (ii) even when models do not engage in “hard” memorization – a verbatim reproduction of training samples – they may still display “soft” memorization whereby they generate outputs that echo the training data but without a complete one-by-one resemblance.
In Proceedings of the 17th International Natural Language Generation Conference , 2024[PDF]
@inproceedings{Borec-2024, title = {The Unreasonable Ineffectiveness of Nucleus Sampling on Mitigating Text Memorization}, author = {Borec, Luka and Sadler, Philipp and Schlangen, David}, editor = {Mahamood, Saad and Minh, Nguyen Le and Ippolito, Daphne}, booktitle = {Proceedings of the 17th International Natural Language Generation Conference}, month = sep, year = {2024}, address = {Tokyo, Japan}, publisher = {Association for Computational Linguistics}, url = {https://aclanthology.org/2024.inlg-main.30}, pages = {358--370} }