Language learning systems usually generalize linguistic observations into rules and patterns that are statistical models of higher level semantic inferences. When the availability of training data is scarce, lexical information can be limited by data sparseness effects and generalization is thus needed. Distributional models represent lexical semantic information in terms of the basic co-occurrences between words in large-scale text collections. As recent works already address, the definition of proper distributional models as well as methods able to express the meaning of phrases or sentences as operations on lexical representations is a complex problem, and a still largely open issue. In this paper, a perspective centered on Convolution Kernels is discussed and the formulation of a Partial Tree Kernel that integrates syntactic information and lexical generalization is studied. Moreover a large scale investigation of different representation spaces, each capturing a different linguistic relation, is provided.

Croce, D., Filice, S., Basili, R. (2015). Distributional models for lexical semantics: An investigation of different representations for natural language learning. In Studies in Computational Intelligence (pp. 115-134). Springer Verlag [10.1007/978-3-319-14206-7_6].

Distributional models for lexical semantics: An investigation of different representations for natural language learning

CROCE, DANILO;BASILI, ROBERTO
2015-03-01

Abstract

Language learning systems usually generalize linguistic observations into rules and patterns that are statistical models of higher level semantic inferences. When the availability of training data is scarce, lexical information can be limited by data sparseness effects and generalization is thus needed. Distributional models represent lexical semantic information in terms of the basic co-occurrences between words in large-scale text collections. As recent works already address, the definition of proper distributional models as well as methods able to express the meaning of phrases or sentences as operations on lexical representations is a complex problem, and a still largely open issue. In this paper, a perspective centered on Convolution Kernels is discussed and the formulation of a Partial Tree Kernel that integrates syntactic information and lexical generalization is studied. Moreover a large scale investigation of different representation spaces, each capturing a different linguistic relation, is provided.
mar-2015
Settore ING-INF/05 - SISTEMI DI ELABORAZIONE DELLE INFORMAZIONI
Settore INF/01 - INFORMATICA
English
Rilevanza internazionale
Articolo scientifico in atti di convegno
Distributional lexical semantics; Kernel methods; Question classification; Artificial Intelligence
http://www.springer.com/series/7092
Croce, D., Filice, S., Basili, R. (2015). Distributional models for lexical semantics: An investigation of different representations for natural language learning. In Studies in Computational Intelligence (pp. 115-134). Springer Verlag [10.1007/978-3-319-14206-7_6].
Croce, D; Filice, S; Basili, R
Contributo in libro
File in questo prodotto:
File Dimensione Formato  
331972_1_En_6_Chapter_Author.pdf

solo utenti autorizzati

Licenza: Copyright dell'editore
Dimensione 416.56 kB
Formato Adobe PDF
416.56 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2108/124017
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact