Deep Dive into Word Sense Disambiguation with LSTM

Research output: Chapter in Book / Report / Conference proceedingConference contributionAcademicpeer-review

Abstract

LSTM-based language models have been shown effective in Word Sense Disambiguation
(WSD). In particular, the technique proposed by Yuan et al. (2016) returned state-of-the-art performance
in several benchmarks, but neither the training data nor the source code was released.
This paper presents the results of a reproduction study and analysis of this technique using only
openly available datasets (GigaWord, SemCor, OMSTI) and software (TensorFlow). Our study
showed that similar results can be obtained with much less data than hinted at by Yuan et al.
(2016). Detailed analyses shed light on the strengths and weaknesses of this method. First,
adding more unannotated training data is useful, but is subject to diminishing returns. Second,
the model can correctly identify both popular and unpopular meanings. Finally, the limited sense
coverage in the annotated datasets is a major limitation. All code and trained models are made
freely available.
LanguageEnglish
Title of host publicationProceedings of the International Conference on Computational Linguistics (COLING 2018)
PublisherInternational Conference on Computational Linguistics (COLING)
Pages354-365
Number of pages12
ISBN (Print)9781948087506
Publication statusPublished - 2018
Event27th International Conference on Computational Linguistics COLING 2018 - Santa Fe, NM
Duration: 20 Aug 201826 Aug 2018
Conference number: 27

Conference

Conference27th International Conference on Computational Linguistics COLING 2018
Abbreviated titleCOLING 2018
CitySanta Fe, NM
Period20/08/1826/08/18

Cite this

Le, M. N., Postma, M. C., Urbani, J., & Vossen, P. T. J. M. (2018). Deep Dive into Word Sense Disambiguation with LSTM. In Proceedings of the International Conference on Computational Linguistics (COLING 2018) (pp. 354-365). International Conference on Computational Linguistics (COLING).
Le, M.N. ; Postma, M.C. ; Urbani, J. ; Vossen, P.T.J.M. / Deep Dive into Word Sense Disambiguation with LSTM. Proceedings of the International Conference on Computational Linguistics (COLING 2018). International Conference on Computational Linguistics (COLING), 2018. pp. 354-365
@inproceedings{533d9c98290d4e5f95e268fce05abde2,
title = "Deep Dive into Word Sense Disambiguation with LSTM",
abstract = "LSTM-based language models have been shown effective in Word Sense Disambiguation(WSD). In particular, the technique proposed by Yuan et al. (2016) returned state-of-the-art performancein several benchmarks, but neither the training data nor the source code was released.This paper presents the results of a reproduction study and analysis of this technique using onlyopenly available datasets (GigaWord, SemCor, OMSTI) and software (TensorFlow). Our studyshowed that similar results can be obtained with much less data than hinted at by Yuan et al.(2016). Detailed analyses shed light on the strengths and weaknesses of this method. First,adding more unannotated training data is useful, but is subject to diminishing returns. Second,the model can correctly identify both popular and unpopular meanings. Finally, the limited sensecoverage in the annotated datasets is a major limitation. All code and trained models are madefreely available.",
author = "M.N. Le and M.C. Postma and J. Urbani and P.T.J.M. Vossen",
year = "2018",
language = "English",
isbn = "9781948087506",
pages = "354--365",
booktitle = "Proceedings of the International Conference on Computational Linguistics (COLING 2018)",
publisher = "International Conference on Computational Linguistics (COLING)",

}

Le, MN, Postma, MC, Urbani, J & Vossen, PTJM 2018, Deep Dive into Word Sense Disambiguation with LSTM. in Proceedings of the International Conference on Computational Linguistics (COLING 2018). International Conference on Computational Linguistics (COLING), pp. 354-365, 27th International Conference on Computational Linguistics COLING 2018, Santa Fe, NM, 20/08/18.

Deep Dive into Word Sense Disambiguation with LSTM. / Le, M.N.; Postma, M.C.; Urbani, J.; Vossen, P.T.J.M.

Proceedings of the International Conference on Computational Linguistics (COLING 2018). International Conference on Computational Linguistics (COLING), 2018. p. 354-365.

Research output: Chapter in Book / Report / Conference proceedingConference contributionAcademicpeer-review

TY - GEN

T1 - Deep Dive into Word Sense Disambiguation with LSTM

AU - Le, M.N.

AU - Postma, M.C.

AU - Urbani, J.

AU - Vossen, P.T.J.M.

PY - 2018

Y1 - 2018

N2 - LSTM-based language models have been shown effective in Word Sense Disambiguation(WSD). In particular, the technique proposed by Yuan et al. (2016) returned state-of-the-art performancein several benchmarks, but neither the training data nor the source code was released.This paper presents the results of a reproduction study and analysis of this technique using onlyopenly available datasets (GigaWord, SemCor, OMSTI) and software (TensorFlow). Our studyshowed that similar results can be obtained with much less data than hinted at by Yuan et al.(2016). Detailed analyses shed light on the strengths and weaknesses of this method. First,adding more unannotated training data is useful, but is subject to diminishing returns. Second,the model can correctly identify both popular and unpopular meanings. Finally, the limited sensecoverage in the annotated datasets is a major limitation. All code and trained models are madefreely available.

AB - LSTM-based language models have been shown effective in Word Sense Disambiguation(WSD). In particular, the technique proposed by Yuan et al. (2016) returned state-of-the-art performancein several benchmarks, but neither the training data nor the source code was released.This paper presents the results of a reproduction study and analysis of this technique using onlyopenly available datasets (GigaWord, SemCor, OMSTI) and software (TensorFlow). Our studyshowed that similar results can be obtained with much less data than hinted at by Yuan et al.(2016). Detailed analyses shed light on the strengths and weaknesses of this method. First,adding more unannotated training data is useful, but is subject to diminishing returns. Second,the model can correctly identify both popular and unpopular meanings. Finally, the limited sensecoverage in the annotated datasets is a major limitation. All code and trained models are madefreely available.

UR - https://www.researchgate.net/publication/327500586_A_Deep_Dive_into_Word_Sense_Disambiguation_with_LSTM

UR - https://github.com/cltl/LSTM-WSD

UR - https://aclanthology.info/events/coling-2018

M3 - Conference contribution

SN - 9781948087506

SP - 354

EP - 365

BT - Proceedings of the International Conference on Computational Linguistics (COLING 2018)

PB - International Conference on Computational Linguistics (COLING)

ER -

Le MN, Postma MC, Urbani J, Vossen PTJM. Deep Dive into Word Sense Disambiguation with LSTM. In Proceedings of the International Conference on Computational Linguistics (COLING 2018). International Conference on Computational Linguistics (COLING). 2018. p. 354-365