首页    期刊浏览 2024年11月26日 星期二
登录注册

文章基本信息

  • 标题:Measuring Semantic Similarity between Words Using Web Pages
  • 本地全文:下载
  • 作者:T.Sujatha ; Ramesh Naidu G ; P.Suresh B
  • 期刊名称:International Journal of Soft Computing & Engineering
  • 电子版ISSN:2231-2307
  • 出版年度:2012
  • 卷号:2
  • 期号:3
  • 页码:213-216
  • 出版社:International Journal of Soft Computing & Engineering
  • 摘要:Semantic similarity measures play an important role in the extraction of semantic relations. Semantic similarity measures are widely used in Natural Language Processing (NLP) and Information Retrieval (IR). The work proposed here uses web based metrics to compute the semantic similarity between words or terms and also compares with the state-of-the-art. For a computer to decide the semantic similarity, it should understand the semantics of the words. Computer being a syntactic machine, it cannot understand the semantics. So always an attempt is made to represent the semantics as syntax. There are various methods proposed to find the semantic similarity between words. Some of these methods have used the precompiled databases like WordNet, and Brown Corpus. Some are based on Web Search Engine. The approach presented here is altogether different from these methods. It makes use of snippets returned by the Wikipedia or any encyclopedia such as Britannica Encyclopedia. The snippets are preprocessed for stop word removal and stemming. For suffix removal an algorithm by M. F. Porter is referred. Luhn’s Idea is used for extraction of significant words from the preprocessed snippets. Similarity measures proposed here are based on the five different association measures in Information retrieval, namely simple matching, Dice, Jaccard, Overlap, Cosine coefficient. Performance of these methods is evaluated using Miller and Charle’s benchmark dataset. It gives higher correlation value of 0.80 than some of the existing methods.
  • 关键词:Semantic Similarity; Wikipedia; Web Search Engine;Natural Language Processing; Information Retrieval; Web;Mining
国家哲学社会科学文献中心版权所有