no code implementations • ACL 2022 • Clara Meister, Gian Wiher, Tiago Pimentel, Ryan Cotterell
When generating natural language from neural probabilistic models, high probability does not always coincide with high quality.
no code implementations • 31 Mar 2022 • Clara Meister, Gian Wiher, Tiago Pimentel, Ryan Cotterell
Specifically, we posit that human-like language should contain an amount of information (quantified as negative log-probability) that is close to the entropy of the distribution over natural strings.
no code implementations • 29 Mar 2022 • Gian Wiher, Clara Meister, Ryan Cotterell
For example, the nature of the diversity-quality trade-off in language generation is very task-specific; the length bias often attributed to beam search is not constant across tasks.
3 code implementations • 1 Feb 2022 • Clara Meister, Tiago Pimentel, Gian Wiher, Ryan Cotterell
Automatic and human evaluations show that, in comparison to nucleus and top-k sampling, locally typical sampling offers competitive performance (in both abstractive summarization and story generation) in terms of quality while consistently reducing degenerate repetitions.