This paper explores the correlation between linguistic diversity, sentiment analysis and transformer model architectures. We aim to investigate how different English variations impact transformer-based models for irony detection. To conduct our study, we used the EPIC corpus to extract five diverse English variation-specific datasets and applied the KEN pruning algorithm on five different architectures. Our results reveal several similarities between optimal subnetworks, which provide insights into the linguistic variations that share strong resemblances and those that exhibit greater dissimilarities. We discovered that optimal subnetworks across models share at least 60% of their parameters, emphasizing the significance of parameter values in capturing and interpreting linguistic variations. This study highlights the inherent structural similarities between models trained on different variants of the same language and also the critical role of parameter values in capturing these nuances.

Mastromattei, M., Zanzotto, F.m. (2024). Linguistic fingerprint in transformer models: how language variation influences parameter selection in irony detection. In Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives) @ LREC-COLING 2024 (pp.123-130). ELRA; ICCL.

Linguistic fingerprint in transformer models: how language variation influences parameter selection in irony detection

Mastromattei M.;Zanzotto F. M.
2024-01-01

Abstract

This paper explores the correlation between linguistic diversity, sentiment analysis and transformer model architectures. We aim to investigate how different English variations impact transformer-based models for irony detection. To conduct our study, we used the EPIC corpus to extract five diverse English variation-specific datasets and applied the KEN pruning algorithm on five different architectures. Our results reveal several similarities between optimal subnetworks, which provide insights into the linguistic variations that share strong resemblances and those that exhibit greater dissimilarities. We discovered that optimal subnetworks across models share at least 60% of their parameters, emphasizing the significance of parameter values in capturing and interpreting linguistic variations. This study highlights the inherent structural similarities between models trained on different variants of the same language and also the critical role of parameter values in capturing these nuances.
3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives) at LREC-COLING 2024
Torino, Italy
2024
3
Rilevanza internazionale
2024
Settore IINF-05/A - Sistemi di elaborazione delle informazioni
English
Explainable models
Irony detection
Language variation
Model optimization
Intervento a convegno
Mastromattei, M., Zanzotto, F.m. (2024). Linguistic fingerprint in transformer models: how language variation influences parameter selection in irony detection. In Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives) @ LREC-COLING 2024 (pp.123-130). ELRA; ICCL.
Mastromattei, M; Zanzotto, Fm
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2108/389004
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact