Here we use word embedding instead of hot encoding, using a Word2Vec Learner node. The hidden layer size is set to 10, therefore producing an embedding with very small dimensionality.
The output of the Word2Vec Learner node is a model. The Vocabulary Extractor node extracts the words from the model vocabulary and provides their embedding in form of collections. Collection items are isolated using a Split Collection column node and the distances between word emebedding vectors are calculated.
At the end, n selected words are visualized on a scatter plot, to show proximity of same semantic words across different embedding coordinates. The String input node allows to insert one selected word and retrieve all word distances from that word. Smaller distances should correspond to closer words in context or in meaning.
Workflow
Distances on Word Embeddings
External resources
Used extensions & nodes
Created with KNIME Analytics Platform version 4.1.0 Note: Not all extensions may be displayed.
- Go to item
KNIME Textprocessing - Deeplearning4J Integration (64bit only)
KNIME AG, Zurich, Switzerland
Version 4.1.0
- Go to item
- Go to item
- Go to item
- Go to item
- Go to item
- Go to item
Legal
By using or downloading the workflow, you agree to our terms and conditions.