Generative Adversarial Networks (GANs) have known a tremendous success for many continuous generation tasks, especially in the field of image generation. However, for discrete outputs such as language, optimizing GANs remains an open problem with many instabilities, as no gradient can be properly back-propagated from the discriminator output to the generator parameters. An alternative is to learn the generator network via reinforcement learning, using the discriminator signal as a reward, but such a technique suffers from moving rewards and vanishing gradient problems. Finally, it often falls short compared to direct maximum-likelihood approaches. In this paper, we introduce Generative Cooperative Networks, in which the discriminator architecture is cooperatively used along with the generation policy to output samples of realistic texts for the task at hand. We give theoretical guarantees of convergence for our approach, and study various efficient decoding schemes to empirically achieve state-of-the-art results in two main NLG tasks.

Generative Cooperative Networks for Natural Language Generation / Lamprier, Sylvain; Scialom, Thomas; Chaffin, Antoine; Claveau, Vincent; Kijak, Ewa; Staiano, Jacopo; Piwowarski, Benjamin. - 162:(2022), pp. 11891-11905. (Intervento presentato al convegno ICML 2022 tenutosi a Baltimore, Maryland nel 17th-23rd July 2022).

Generative Cooperative Networks for Natural Language Generation

Staiano, Jacopo;
2022-01-01

Abstract

Generative Adversarial Networks (GANs) have known a tremendous success for many continuous generation tasks, especially in the field of image generation. However, for discrete outputs such as language, optimizing GANs remains an open problem with many instabilities, as no gradient can be properly back-propagated from the discriminator output to the generator parameters. An alternative is to learn the generator network via reinforcement learning, using the discriminator signal as a reward, but such a technique suffers from moving rewards and vanishing gradient problems. Finally, it often falls short compared to direct maximum-likelihood approaches. In this paper, we introduce Generative Cooperative Networks, in which the discriminator architecture is cooperatively used along with the generation policy to output samples of realistic texts for the task at hand. We give theoretical guarantees of convergence for our approach, and study various efficient decoding schemes to empirically achieve state-of-the-art results in two main NLG tasks.
2022
Proceedings of the 39th International Conference on Machine Learning
Cambridge, MA
PMLR
Lamprier, Sylvain; Scialom, Thomas; Chaffin, Antoine; Claveau, Vincent; Kijak, Ewa; Staiano, Jacopo; Piwowarski, Benjamin
Generative Cooperative Networks for Natural Language Generation / Lamprier, Sylvain; Scialom, Thomas; Chaffin, Antoine; Claveau, Vincent; Kijak, Ewa; Staiano, Jacopo; Piwowarski, Benjamin. - 162:(2022), pp. 11891-11905. (Intervento presentato al convegno ICML 2022 tenutosi a Baltimore, Maryland nel 17th-23rd July 2022).
File in questo prodotto:
File Dimensione Formato  
lamprier22a.pdf

accesso aperto

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 623.01 kB
Formato Adobe PDF
623.01 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/363046
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 10
  • ???jsp.display-item.citation.isi??? 1
  • OpenAlex ND
social impact