Language models generate texts by successively predicting probability distributions for next tokens given past ones. A growing field of interest tries to leverage external information in the decoding process so that the generated texts have desired properties, such as being more natural, non toxic, faithful, or having a specific writing style. A solution is to use a classifier at each generation step, resulting in a cooperative environment where the classifier guides the decoding of the language model distribution towards relevant texts for the task at hand. In this paper, we examine three families of (transformer-based) discriminators for this specific task of cooperative decoding: bidirectional, left-to-right and generative ones. We evaluate the pros and cons of these different types of discriminators for cooperative generation, exploring respective accuracy on classification tasks along with their impact on the resulting sample quality and computational performances. We also provide the code of a batched implementation of the powerful cooperative decoding strategy used for our experiments, the Monte Carlo Tree Search, working with each discriminator for Natural Language Generation.

Which Discriminator for Cooperative Text Generation? / Chaffin, Antoine; Scialom, Thomas; Lamprier, Sylvain; Staiano, Jacopo; Piwowarski, Benjamin; Kijak, Ewa; Claveau, Vincent. - (2022), pp. 2360-2365. (Intervento presentato al convegno SIGIR 2022 tenutosi a Madrid nel 11th-15th July) [10.1145/3477495.3531858].

Which Discriminator for Cooperative Text Generation?

Staiano, Jacopo;
2022-01-01

Abstract

Language models generate texts by successively predicting probability distributions for next tokens given past ones. A growing field of interest tries to leverage external information in the decoding process so that the generated texts have desired properties, such as being more natural, non toxic, faithful, or having a specific writing style. A solution is to use a classifier at each generation step, resulting in a cooperative environment where the classifier guides the decoding of the language model distribution towards relevant texts for the task at hand. In this paper, we examine three families of (transformer-based) discriminators for this specific task of cooperative decoding: bidirectional, left-to-right and generative ones. We evaluate the pros and cons of these different types of discriminators for cooperative generation, exploring respective accuracy on classification tasks along with their impact on the resulting sample quality and computational performances. We also provide the code of a batched implementation of the powerful cooperative decoding strategy used for our experiments, the Monte Carlo Tree Search, working with each discriminator for Natural Language Generation.
2022
SIGIR 2022: Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval
New York, NY, USA
Association for Computing Machinery, Inc
9781450387323
Chaffin, Antoine; Scialom, Thomas; Lamprier, Sylvain; Staiano, Jacopo; Piwowarski, Benjamin; Kijak, Ewa; Claveau, Vincent
Which Discriminator for Cooperative Text Generation? / Chaffin, Antoine; Scialom, Thomas; Lamprier, Sylvain; Staiano, Jacopo; Piwowarski, Benjamin; Kijak, Ewa; Claveau, Vincent. - (2022), pp. 2360-2365. (Intervento presentato al convegno SIGIR 2022 tenutosi a Madrid nel 11th-15th July) [10.1145/3477495.3531858].
File in questo prodotto:
File Dimensione Formato  
3477495.3531858.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.14 MB
Formato Adobe PDF
1.14 MB Adobe PDF   Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/362927
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? 1
  • OpenAlex ND
social impact