Actions

Algebra with Words: Difference between revisions

From Algolit

 
(8 intermediate revisions by one other user not shown)
Line 1: Line 1:
{|
+
by Algolit
|-
 
| Type: || Algoliterary exploration
 
|-
 
| Technique: || Word embeddings, word2vec
 
|-
 
| Developed by: || Radim Rehurek and Petr Sojka & Algolit
 
|}
 
  
 
Word embeddings are language modelling techniques that through multiple mathematical operations of counting and ordering, plot words into a multi-dimensional vector space. When embedding words, they transform from being distinct symbols into mathematical objects that can be multiplied, divided, added or substracted.
 
Word embeddings are language modelling techniques that through multiple mathematical operations of counting and ordering, plot words into a multi-dimensional vector space. When embedding words, they transform from being distinct symbols into mathematical objects that can be multiplied, divided, added or substracted.
  
While distributing the words along the many diagonal lines of the vector space, the visibility of their new geometrical placements disappears. However, what is gained are multiple, simultaneous ways of ordering. Algebraic operations make the relations between vectors graspable again.
+
By distributing the words along the many diagonal lines of the multi-dimensional vector space, their new geometrical placements become impossible to perceive by humans. However, what is gained are multiple, simultaneous ways of ordering. Algebraic operations make the relations between vectors graspable again.
  
This exploration is using [https://radimrehurek.com/gensim/index.html gensim], an open source vector space and topic modelling toolkit implemented in Python, to manipulate text according to the mathematic relationships which emerge between the words, once they have been plotted in a vector space.
+
This installation uses [https://radimrehurek.com/gensim/index.html Gensim], an open-source vector space and topic-modelling toolkit implemented in the programming language Python. It allows to manipulate the text using the mathematical relationships that emerge between the words, once they have been plotted in a vector space.
  
[[Category:Data_Workers]][[Category:Data_Workers_EN]]
+
------------------------------------------
 +
Concept & interface: Cristina Cochior
 +
 
 +
Technique: word embeddings, word2vec
 +
 
 +
Original model: Radim Rehurek and Petr Sojka

Latest revision as of 19:23, 19 March 2019

by Algolit

Word embeddings are language modelling techniques that through multiple mathematical operations of counting and ordering, plot words into a multi-dimensional vector space. When embedding words, they transform from being distinct symbols into mathematical objects that can be multiplied, divided, added or substracted.

By distributing the words along the many diagonal lines of the multi-dimensional vector space, their new geometrical placements become impossible to perceive by humans. However, what is gained are multiple, simultaneous ways of ordering. Algebraic operations make the relations between vectors graspable again.

This installation uses Gensim, an open-source vector space and topic-modelling toolkit implemented in the programming language Python. It allows to manipulate the text using the mathematical relationships that emerge between the words, once they have been plotted in a vector space.


Concept & interface: Cristina Cochior

Technique: word embeddings, word2vec

Original model: Radim Rehurek and Petr Sojka