Combining embedding methods for a word intrusion task

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

27 Downloads (Pure)

Abstract

We report a new baseline for a Danish word intrusion task by combining pre-trained offthe-shelf word, subword and knowledge graph embedding models. We test fastText, Byte-Pair Encoding, BERT and the knowledge graph embedding in Wembedder, finding fastText as the individual model with the superior performance, while a simple combination of the fastText with other models can slightly improve the accuracy of finding the odd-one-out words in the word
intrusion task.
Original languageEnglish
Title of host publicationProceedings of the 15th Conference on Natural Language Processing
PublisherAssociation for Computational Linguistics
Publication date2019
Pages237-240
Publication statusPublished - 2019
Event15th Conference on Natural Language Processing - Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, Germany
Duration: 9 Oct 201911 Oct 2019
https://2019.konvens.org/

Conference

Conference15th Conference on Natural Language Processing
LocationFriedrich-Alexander-Universität Erlangen-Nürnberg
Country/TerritoryGermany
CityErlangen
Period09/10/201911/10/2019
Internet address

Cite this