Abstract
In this chapter, we present the vector space model and some ways to further process such a representation: With feature hashing, random indexing, latent semantic analysis, non-negative matrix factorization, explicit semantic analysis and word embedding, a word or a text may be associated with a distributed semantic representation. Deep learning, explicit semantic networks and auxiliary non-linguistic information provide further means for creating distributed representations from linguistic data. We point to a few of the methods and datasets used to evaluate the many different algorithms that create a semantic representation, and we also point to some of the problems associated with distributed representations.
Original language | English |
---|---|
Title of host publication | Statistical Semantics : Methods and Applications |
Publisher | Springer |
Publication date | 2020 |
Pages | 11-31 |
Chapter | 2 |
ISBN (Print) | 978-3-030-37249-1 |
ISBN (Electronic) | 978-3-030-37250-7 |
DOIs | |
Publication status | Published - 2020 |