Proceedings of the 13th ESSLLI Student Session pages:143-152
European Summer School in Logic, Language and Information (ESSLLI) edition:20 location:Hamburg, Germany date:4-15 August 2008
Word Space Models provide a convenient way of modelling word meaning in terms of a word’s contexts in a corpus. This paper investigates the influence of the type of context features on the kind of semantic information that the models capture. In particular, we make a distinction between semantic similarity and semantic relatedness. It is shown that the strictness of the context definition correlates with the models’ ability to identify semantically similar words: syntactic approaches perform
better than bag-of-word models, and small context windows are better than larger ones. For semantic relatedness, however, syntactic features and small context windows
are at a clear disadvantage. Second-order bag-of-word models perform below average across the board.