In this paper, we present MT-GAN-BERT, i.e., a BERT-based architecture for faceted classification tasks. It aims to reduce the requirements of Transformers both in terms of the amount of annotated data and the computational cost required at classification time. First, MT-GAN-BERT enables semi-supervised learning in BERT-based architectures based on Generative Adversarial Learning. Second, it implements a Multi-task Learning approach to solve multiple tasks simultaneously. A single BERTbased model is used to encode the input examples, while multiple linear layers are used to implement the classification steps, with a significant reduction of the computational costs. Experimental evaluations against six classification tasks involved in detecting abusive languages in Italian suggest that MT-GAN-BERT represents a sustainable solution that generally improves the raw adoption of multiple BERT-based models with lighter requirements in terms of annotated data and computational costs.

Breazzano, C., Croce, D., Basili, R. (2021). MT-GAN-BERT: Multi-Task and Generative Adversarial Learning for sustainable Language Processing. In NL4AI 2021: fifth workshop on natural language for Artificial Intelligence: proceedings of the fifth workshop on natural language for Artificial Intelligence (NL4AI 2021) co-located with 20th International Conference of the Italian Association for Artificial Intelligence (AI*IA 2021). CEUR-WS.

MT-GAN-BERT: Multi-Task and Generative Adversarial Learning for sustainable Language Processing

Croce D.;Basili R.
2021-01-01

Abstract

In this paper, we present MT-GAN-BERT, i.e., a BERT-based architecture for faceted classification tasks. It aims to reduce the requirements of Transformers both in terms of the amount of annotated data and the computational cost required at classification time. First, MT-GAN-BERT enables semi-supervised learning in BERT-based architectures based on Generative Adversarial Learning. Second, it implements a Multi-task Learning approach to solve multiple tasks simultaneously. A single BERTbased model is used to encode the input examples, while multiple linear layers are used to implement the classification steps, with a significant reduction of the computational costs. Experimental evaluations against six classification tasks involved in detecting abusive languages in Italian suggest that MT-GAN-BERT represents a sustainable solution that generally improves the raw adoption of multiple BERT-based models with lighter requirements in terms of annotated data and computational costs.
5th Workshop on Natural Language for Artificial Intelligence, NL4AI 2021
Online event
2021
5
Rilevanza internazionale
2021
Settore INF/01
Settore ING-INF/05
English
BERT
Generative adversarial learning
Multi-task learning
Semi supervised learning
Sustainable NLP
Intervento a convegno
Breazzano, C., Croce, D., Basili, R. (2021). MT-GAN-BERT: Multi-Task and Generative Adversarial Learning for sustainable Language Processing. In NL4AI 2021: fifth workshop on natural language for Artificial Intelligence: proceedings of the fifth workshop on natural language for Artificial Intelligence (NL4AI 2021) co-located with 20th International Conference of the Italian Association for Artificial Intelligence (AI*IA 2021). CEUR-WS.
Breazzano, C; Croce, D; Basili, R
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2108/359270
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 3
  • ???jsp.display-item.citation.isi??? ND
social impact