What NLP tools to use to match phrases having similar meaning or semantics
Solution 1
When Latent Semantic Analysis refers to a "document", it basically means any set of words that is longer than 1. You can use it to compute the similarity between a document and another document, between a word and another word, or between a word and a document. So you could certainly use it for your chosen application.
Other algorithms that may be useful include:
- Random indexing ( https://www.sics.se/~mange/papers/RI_intro.pdf ) is easy enough to implement oneself without too much difficulty. There is also an implementation within https://code.google.com/p/airhead-research/ , but it's in Java, not Python.
- Topic modeling ( http://psiexp.ss.uci.edu/research/papers/SteyversGriffithsLSABookFormatted.pdf ) - Python implementation at http://radimrehurek.com/gensim/tutorial.html
- DISSECT ( http://clic.cimec.unitn.it/composes/toolkit/introduction.html ) - Python implementation at http://clic.cimec.unitn.it/composes/toolkit/installation.html
- BEAGLE ( http://www.indiana.edu/~clcl/BEAGLE/Jones_Mewhort_PR.pdf ) - Python implementation at https://github.com/mike-lawrence/wikiBEAGLE
Solution 2
If you have a big corpus, where these words occur, available, you can train a model to represent each word as vector. For instance, you can use deep learning via word2vec’s "skip-gram and CBOW models", they are implemented in the gensim software package
In the word2vec model, each word is represented by a vector, you can then measure the semantic similarity between two words by measuring the cosine of the vectors representing th words. Semantic similar words should have a high cosine similarity, for instance:
model.similarity('cheap','inexpensive') = 0.8
(The value is made up, just for illustration.)
Also, from my experiments, summing a relatively small number of words (i.e., up to 3 or 4 words) preserves the semantics, for instance:
vector1 = model['cheap']+model['health']+model['insurance']
vector2 = model['low']+model['cost']+model['medical']+model['insurance']
similarity(vector1,vector2) = 0.7
(Again, just for illustration.)
You can use this semantic similarity measure between words as a measure to generate your clusters.
Solution 3
I'd start by taking a look at Wordnet. It will give you real synonyms and other word relations for hundreds of thousands of terms. Since you tagged the nltk
: It provides bindings for Wordnet, and you can use it as the basis for domain-specific solutions.
Still in the NLTK, check out the discussion of the method similar()
in the introduction to the NLTK book, and the class nltk.text.ContextIndex
that it's based on. (All pretty simple still, but it might be all you really need).
Arun Shyam
Updated on June 09, 2022Comments
-
Arun Shyam almost 2 years
I am working on a project which requires me to match a phrase or keyword with a set of similar keywords. I need to perform semantic analysis for the same.
an example:
Relevant QT
cheap health insurance
affordable health insurance
low cost medical insurance
health plan for less
inexpensive health coverageCommon Meaning
low cost health insurance
Here the the word under Common Meaning column should match the under Relevant QT column. I looked at a bunch of tools and techniques to do the same. S-Match seemed very promising, but I have to work in Python, not in Java. Also Latent Semantic Analysis looks good but I think its more for document classification based upon a Keyword rather than keyword matching. I am somewhat familiar with NLTK. Could someone provide some insight on what direction I should proceed and what tools I should use for the same?