The primary goal of this FBK’s systems submission to the IWSLT 2022 offline and simultaneous speech translation tasks is to reduce model training costs without sacrificing translation quality. As such, we first question the need of ASR pre-training, showing that it is not essential to achieve competitive results. Second, we focus on data filtering, showing that a simple method that looks at the ratio between source and target characters yields a quality improvement of 1 BLEU. Third, we compare different methods to reduce the detrimental effect of the audio segmentation mismatch between training data manually segmented at sentence level and inference data that is automatically segmented. Towards the same goal of training cost reduction, we participate in the simultaneous task with the same model trained for offline ST. The effectiveness of our lightweight training strategy is shown by the high score obtained on the MuST-C en-de corpus (26.7 BLEU) and is confirmed in high-resource data conditions by a 1.6 BLEU improvement on the IWSLT2020 test set over last year’s winning system.

Efficient yet Competitive Speech Translation: FBK@IWSLT2022 / Gaido, Marco; Papi, Sara; Fucci, Dennis; Fiameni, Giuseppe; Negri, Matteo; Turchi, Marco. - (2022), pp. 177-189. (Intervento presentato al convegno 19th International Conference on Spoken Language Translation, IWSLT 2022 tenutosi a Dublin, Ireland (in-person and online) nel 26-27, May 2022) [10.18653/v1/2022.iwslt-1.13].

Efficient yet Competitive Speech Translation: FBK@IWSLT2022

Gaido, Marco;Papi, Sara;Fucci, Dennis;Turchi, Marco
2022-01-01

Abstract

The primary goal of this FBK’s systems submission to the IWSLT 2022 offline and simultaneous speech translation tasks is to reduce model training costs without sacrificing translation quality. As such, we first question the need of ASR pre-training, showing that it is not essential to achieve competitive results. Second, we focus on data filtering, showing that a simple method that looks at the ratio between source and target characters yields a quality improvement of 1 BLEU. Third, we compare different methods to reduce the detrimental effect of the audio segmentation mismatch between training data manually segmented at sentence level and inference data that is automatically segmented. Towards the same goal of training cost reduction, we participate in the simultaneous task with the same model trained for offline ST. The effectiveness of our lightweight training strategy is shown by the high score obtained on the MuST-C en-de corpus (26.7 BLEU) and is confirmed in high-resource data conditions by a 1.6 BLEU improvement on the IWSLT2020 test set over last year’s winning system.
2022
Proceedings of the 19th International Conference on Spoken Language Translation (IWSLT 2022)
Dublin, Ireland (in-person and online)
Association for Computational Linguistics (ACL)
978-1-955917-41-4
Gaido, Marco; Papi, Sara; Fucci, Dennis; Fiameni, Giuseppe; Negri, Matteo; Turchi, Marco
Efficient yet Competitive Speech Translation: FBK@IWSLT2022 / Gaido, Marco; Papi, Sara; Fucci, Dennis; Fiameni, Giuseppe; Negri, Matteo; Turchi, Marco. - (2022), pp. 177-189. (Intervento presentato al convegno 19th International Conference on Spoken Language Translation, IWSLT 2022 tenutosi a Dublin, Ireland (in-person and online) nel 26-27, May 2022) [10.18653/v1/2022.iwslt-1.13].
File in questo prodotto:
File Dimensione Formato  
2022.iwslt-1.13.pdf

accesso aperto

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Creative commons
Dimensione 443.9 kB
Formato Adobe PDF
443.9 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/369995
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 12
  • ???jsp.display-item.citation.isi??? 3
social impact