Copyright Statement: This is an open access article licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, even commercially as long as the original work is properly cited.
Digital Object Identifier (DOI) : 10.14569/IJACSA.2014.050802
Article Published in International Journal of Advanced Computer Science and Applications(IJACSA), Volume 5 Issue 8, 2014.
Abstract: This paper puts forward a two layers computing method to calculate semantic similarity of Chinese word. Firstly, using Latent Dirichlet Allocation (LDA) subject model to generate subject spatial domain. Then mapping word into topic space and forming topic distribution which is used to calculate semantic similarity of word(the first layer computing). Finally, using semantic dictionary "HowNet" to deeply excavate semantic similarity of word (the second layer computing). This method not only overcomes the problem that it’s not specific enough merely using LDA to calculate semantic similarity of word, but also solves the problems such as new words (haven’t been added in dictionary) and without considering specific context when calculating semantic similarity based on semantic dictionary "HowNet". By experimental comparison, this thesis proves feasibility,availability and advantages of the calculation method.
Liqiang Pan, Pu Zhang and Anping Xiong, “Semantic Similarity Calculation of Chinese Word” International Journal of Advanced Computer Science and Applications(IJACSA), 5(8), 2014. http://dx.doi.org/10.14569/IJACSA.2014.050802