{"id":16701,"date":"2020-07-12T15:25:58","date_gmt":"2020-07-12T09:55:58","guid":{"rendered":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/"},"modified":"2024-09-02T17:17:15","modified_gmt":"2024-09-02T11:47:15","slug":"word-embedding","status":"publish","type":"post","link":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/","title":{"rendered":"What is Word Embedding | Word2Vec | GloVe"},"content":{"rendered":"\n<ol class=\"wp-block-list\">\n<li><a href=\"#sh1\">What is Word Embedding?<\/a><\/li>\n\n\n\n<li><a href=\"#sh2\">Why Word Embedding is used?<\/a><\/li>\n\n\n\n<li><a href=\"#sh3\">What is Word2Vec<\/a>\n<ol class=\"wp-block-list\">\n<li><a href=\"#sh31\">Continuous Bag-of-Words, or CBOW model.<\/a><\/li>\n\n\n\n<li><a href=\"#sh32\">Skip-Gram Model.<\/a><\/li>\n<\/ol>\n<\/li>\n\n\n\n<li><a href=\"#sh4\">GloVe<\/a><\/li>\n<\/ol>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"what-is-word-embedding\"><strong>What is Word Embedding?<\/strong><\/h2>\n\n\n\n<p>Humans have always excelled at understanding languages. It is easy for humans to understand the relationship between words but for computers, this task may not be simple. For example, we humans understand the words like king and queen, man and woman, tiger and tigress have a certain type of relation between them but how can a computer figure this out?<br><\/p>\n\n\n\n<p>Word embeddings are basically a form of word representation that bridges the human understanding of language to that of a machine. They have learned representations of text in an n-dimensional space where words that have the same meaning have a similar representation. Meaning that two similar words are represented by almost similar vectors that are very closely placed in a vector space. These are essential for solving most Natural language processing problems.<br><\/p>\n\n\n<figure class=\"wp-block-image aligncenter size-large is-resized zoomable\" data-full=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we1.png\"><img decoding=\"async\" width=\"283\" height=\"285\" src=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we1.png\" alt=\"word embedding\" class=\"wp-image-16766\" style=\"width:283px;height:285px\" srcset=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we1.png 283w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we1-150x150.png 150w\" sizes=\"(max-width: 283px) 100vw, 283px\" \/><figcaption class=\"wp-element-caption\"> <em>Similar words are closely placed in vector space<\/em> <\/figcaption><\/figure>\n\n\n\n<p>Thus when using word embeddings, all individual words are represented as real-valued vectors in a predefined vector space. Each word is mapped to one vector and the vector values are learned in a way that resembles a neural network.<\/p>\n\n\n\n<p>Word2Vec is one of the most popular technique to learn word embeddings using shallow neural network. It was developed by Tomas Mikolov in 2013 at Google.<\/p>\n\n\n<figure class=\"wp-block-image size-large zoomable\" data-full=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/June-29-banner-for-neural-network-1-1.png\"><a href=\"https:\/\/www.mygreatlearning.com\/academy\/learn-for-free\/courses\/introduction-to-neural-networks-and-deep-learning\" target=\"_blank\" rel=\"noreferrer noopener\"><img decoding=\"async\" width=\"1000\" height=\"242\" src=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/June-29-banner-for-neural-network-1-1.png\" alt=\"\" class=\"wp-image-16705\" srcset=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/June-29-banner-for-neural-network-1-1.png 1000w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/June-29-banner-for-neural-network-1-1-300x73.png 300w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/June-29-banner-for-neural-network-1-1-768x186.png 768w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/June-29-banner-for-neural-network-1-1-696x168.png 696w\" sizes=\"(max-width: 1000px) 100vw, 1000px\" \/><\/a><\/figure>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"why-word-embeddings-are-used\"><strong>Why Word Embeddings are used?<\/strong><br><\/h2>\n\n\n\n<p>As we know the machine learning models cannot process text so we need to figure out a way to convert these textual data into numerical data. Previously techniques like <a href=\"https:\/\/www.mygreatlearning.com\/blog\/bag-of-words\/\" target=\"_blank\" rel=\"noreferrer noopener\" aria-label=\"Bag of Words (opens in a new tab)\">Bag of Words<\/a> and <a href=\"https:\/\/www.mygreatlearning.com\/blog\/bag-of-words\/\" target=\"_blank\" rel=\"noreferrer noopener\" aria-label=\"TF-IDF (opens in a new tab)\">TF-IDF<\/a> have been discussed that can help achieve use this task. Apart from this, we can use two more techniques such as one-hot encoding, or we can use unique numbers to represent words in a vocabulary. The latter approach is more efficient than one-hot encoding as instead of a sparse vector, we now have a dense one. Thus this approach even works when our vocabulary is large.<br><\/p>\n\n\n\n<p>In the below example, we assume we have a small vocabulary containing just four words, using the two techniques we represent the sentence \u2018Come sit down\u2019.<br><\/p>\n\n\n<figure class=\"wp-block-image aligncenter size-large zoomable\" data-full=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we2.png\"><img decoding=\"async\" width=\"449\" height=\"176\" src=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we2.png\" alt=\"word embedding\" class=\"wp-image-16767\" srcset=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we2.png 449w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we2-300x118.png 300w\" sizes=\"(max-width: 449px) 100vw, 449px\" \/><figcaption class=\"wp-element-caption\"> <em>One-hot encoding vs integer encoding<\/em> <\/figcaption><\/figure>\n\n\n\n<p>However, the integer-encoding is arbitrary as it does not capture any relationship between words. It can be challenging for a model to interpret, for example, a linear classifier learns a single weight for each feature. Because there is no relationship between the similarity of any two words and the similarity of their encodings, this feature-weight combination is not meaningful.<br><\/p>\n\n\n\n<p>Thus by using word embeddings, words that are close in meaning are grouped near to one another in vector space. For example, while representing a word such as frog, the nearest neighbour of a frog would be frogs, toads, Litoria. This implies that it is alright for a classifier to not see the word Litoria and only frog during training, and the classifier would not be thrown off when it sees Litoria during testing because the two-word vectors are similar. Also, word embeddings learn relationships. Vector differences between a pair of words can be added to another word vector to find the analogous word. For example, \u201cman\u201d -\u201cwoman\u201d + \u201cqueen\u201d \u2248 \u201cking\u201d.<br><\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"what-is-word2vec\"><strong>What is word2Vec?<\/strong><\/h2>\n\n\n\n<p>Word2vec is a method to efficiently create word embeddings by using a two-layer neural network. It was developed by Tomas Mikolov, et al. at Google in 2013 as a response to make the neural-network-based training of the embedding more efficient and since then has become the de facto standard for developing pre-trained word embedding.<\/p>\n\n\n\n<p>The input of word2vec is a text corpus and its output is a set of vectors known as feature vectors that represent words in that corpus. While Word2vec is not a deep neural network, it turns text into a numerical form that deep neural networks can understand.<\/p>\n\n\n\n<p>The Word2Vec objective function causes the words that have a similar context to have similar embeddings. Thus in this vector space, these words are really close. Mathematically, the cosine of the angle (Q) between such vectors should be close to 1, i.e. angle close to 0.<\/p>\n\n\n<figure class=\"wp-block-image aligncenter size-large zoomable\" data-full=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we3.png\"><img decoding=\"async\" width=\"243\" height=\"213\" src=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we3.png\" alt=\"word embedding\" class=\"wp-image-16768\"><figcaption class=\"wp-element-caption\"> <em>For Frog and Litoria, the angle q tends to zero<\/em> <\/figcaption><\/figure>\n\n\n\n<p>Word2vec is not a single algorithm but a combination of two techniques \u2013 CBOW(Continuous bag of words) and Skip-gram model. Both of these are shallow neural networks which map word(s) to the target variable which is also a word(s). Both of these techniques learn weights which act as word vector representations.&nbsp;<\/p>\n\n\n<figure class=\"wp-block-image size-large td-caption-align-https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4.png zoomable\" data-full=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4.png\"><img decoding=\"async\" width=\"1024\" height=\"624\" src=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4-1024x624.png\" alt=\"Word2Vec\" class=\"wp-image-16769\" srcset=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4-1024x624.png 1024w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4-300x183.png 300w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4-768x468.png 768w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4-1536x936.png 1536w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4-696x424.png 696w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4-1068x651.png 1068w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4-689x420.png 689w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we4.png 1542w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><figcaption class=\"wp-element-caption\"> <em> Word2Vec Training Models taken from &ldquo;Efficient Estimation of Word Representations in Vector Space&rdquo;, 2013<\/em> <\/figcaption><\/figure>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"continuous-bag-of-words-model-cbow\"><strong>Continuous Bag-of-Words model&nbsp; (CBOW)<\/strong><br><\/h3>\n\n\n\n<p>CBOW predicts the probability of a word to occur given the words surrounding it. We can consider a single word or a group of words. But for simplicity, we will take a single context word and try to predict a single target word.<br><\/p>\n\n\n\n<p>The English language contains almost 1.2 million words, making it impossible to include so many words in our example. So I \u2018ll consider a small example in which we have only four words i.e. live, home, they and at. For simplicity, we will consider that the corpus contains only one sentence, that being, \u2018They live at home\u2019.<\/p>\n\n\n<figure class=\"wp-block-image size-large zoomable\" data-full=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we5.png\"><img decoding=\"async\" width=\"582\" height=\"492\" src=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we5.png\" alt=\"Continuos Bag of words\" class=\"wp-image-16770\" srcset=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we5.png 582w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we5-300x254.png 300w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we5-497x420.png 497w\" sizes=\"(max-width: 582px) 100vw, 582px\" \/><\/figure>\n\n\n\n<p>First, we convert each word into a one-hot encoding form. Also, we\u2019ll not consider all the words in the sentence but ll only take certain words that are in a window. For example for a window size equal to three, we only consider three words in a sentence. The middle word is to be predicted and the surrounding two words are fed into the neural network as context. The window is then slid and the process is repeated again.<\/p>\n\n\n\n<p>Finally, after training the network repeatedly by sliding the window a shown above, we get weights which we use to get the embeddings as shown below.<\/p>\n\n\n<figure class=\"wp-block-image aligncenter size-large zoomable\" data-full=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we6.png\"><img decoding=\"async\" width=\"520\" height=\"267\" src=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we6.png\" alt=\"word embedding\" class=\"wp-image-16771\" srcset=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we6.png 520w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we6-300x154.png 300w\" sizes=\"(max-width: 520px) 100vw, 520px\" \/><\/figure>\n\n\n\n<p>Usually, we take a window size of around 8-10 words and have a vector size of 300.<\/p>\n\n\n\n<h3 class=\"wp-block-heading\" id=\"skip-gram-model\"><strong>Skip-gram model<\/strong><br><\/h3>\n\n\n\n<p>The Skip-gram model architecture usually tries to achieve the reverse of what the CBOW model does. It tries to predict the source context words (surrounding words) given a target word (the centre word)<\/p>\n\n\n\n<p>The working of the skip-gram model is quite similar to the CBOW but there is just a difference in the architecture of its neural network and the way the weight matrix is generated&nbsp; as shown in the figure below:<\/p>\n\n\n<figure class=\"wp-block-image size-large zoomable\" data-full=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we7.png\"><img decoding=\"async\" width=\"582\" height=\"492\" src=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we7.png\" alt=\"skip-gram model\" class=\"wp-image-16772\" srcset=\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we7.png 582w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we7-300x254.png 300w, https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/we7-497x420.png 497w\" sizes=\"(max-width: 582px) 100vw, 582px\" \/><\/figure>\n\n\n\n<p>After obtaining the weight matrix, the steps to get word embedding is same as CBOW.<br><\/p>\n\n\n\n<p>So now which one of the two algorithms should we use for implementing word2vec? Turns out for large corpus with higher dimensions, it is better to use skip-gram but is slow to train. Whereas CBOW is better for small corpus and is faster to train too.<br><\/p>\n\n\n\n<h2 class=\"wp-block-heading\" id=\"glove\"><strong>GloVe<\/strong><\/h2>\n\n\n\n<p>GloVe (Global Vectors for Word Representation) is an alternate method to create word embeddings. It is based on matrix factorization techniques on the word-context matrix. A large matrix of co-occurrence information is constructed and you count each \u201cword\u201d (the rows), and how frequently we see this word in some \u201ccontext\u201d (the columns) in a large corpus. Usually, we scan our corpus in the following manner: for each term, we look for context terms within some area defined by a window-size before the term and a window-size after the term. Also, we give less weight for more distant words.<\/p>\n\n\n\n<p>The number of \u201ccontexts\u201d is, of course, large, since it is essentially combinatorial in size. So then we factorize this matrix to yield a lower-dimensional matrix, where each row now yields a vector representation for each word. In general, this is done by minimizing a \u201creconstruction loss\u201d. This loss tries to find the lower-dimensional representations which can explain most of the variance in the high-dimensional data.<br><\/p>\n\n\n\n<p>In practice, we use both GloVe and Word2Vec to convert our text into embeddings and both exhibit comparable performances. Although in real applications we train our model over Wikipedia text with a window size around 5- 10. The number of words in the corpus is around 13 million, hence it takes a huge amount of time and resources to generate these embeddings. To avoid this we can use the pre-trained word vectors that are already trained and we can easily use them. Here are the links to download pre-trained<a href=\"http:\/\/vectors.nlpl.eu\/repository\/\"> Word2Vec<\/a> or <a href=\"https:\/\/nlp.stanford.edu\/projects\/glove\/\">GloVe<\/a>.<br><\/p>\n\n\n\n<p>This brings us to the end of this article where we learned about word embedding and some popular techniques to implement them.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>What is Word Embedding? Humans have always excelled at understanding languages. It is easy for humans to understand the relationship between words but for computers, this task may not be simple. For example, we humans understand the words like king and queen, man and woman, tiger and tigress have a certain type of relation between [&hellip;]<\/p>\n","protected":false},"author":41,"featured_media":16702,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"_uag_custom_page_level_css":"","site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"default","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","ast-disable-related-posts":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"set","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"var(--ast-global-color-4)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"footnotes":""},"categories":[2],"tags":[],"content_type":[],"class_list":["post-16701","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-artificial-intelligence"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO Premium plugin v26.6 (Yoast SEO v27.0) - https:\/\/yoast.com\/product\/yoast-seo-premium-wordpress\/ -->\n<title>What is Word Embedding | Word2Vec | GloVe<\/title>\n<meta name=\"description\" content=\"Wha is Word Embedding or Text: We convert text into Word Embeddings so that the Machine learning algorithms can process it.Word2Vec and GloVe are pioneers when it comes to Word Embedding\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"What is Word Embedding | Word2Vec | GloVe\" \/>\n<meta property=\"og:description\" content=\"Wha is Word Embedding or Text: We convert text into Word Embeddings so that the Machine learning algorithms can process it.Word2Vec and GloVe are pioneers when it comes to Word Embedding\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/\" \/>\n<meta property=\"og:site_name\" content=\"Great Learning Blog: Free Resources what Matters to shape your Career!\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/GreatLearningOfficial\/\" \/>\n<meta property=\"article:published_time\" content=\"2020-07-12T09:55:58+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2024-09-02T11:47:15+00:00\" \/>\n<meta property=\"og:image\" content=\"http:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1000\" \/>\n\t<meta property=\"og:image:height\" content=\"700\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Great Learning Editorial Team\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@https:\/\/twitter.com\/Great_Learning\" \/>\n<meta name=\"twitter:site\" content=\"@Great_Learning\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Great Learning Editorial Team\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"8 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#article\",\"isPartOf\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/\"},\"author\":{\"name\":\"Great Learning Editorial Team\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/person\/6f993d1be4c584a335951e836f2656ad\"},\"headline\":\"What is Word Embedding | Word2Vec | GloVe\",\"datePublished\":\"2020-07-12T09:55:58+00:00\",\"dateModified\":\"2024-09-02T11:47:15+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/\"},\"wordCount\":1418,\"commentCount\":1,\"publisher\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#organization\"},\"image\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg\",\"articleSection\":[\"AI and Machine Learning\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/\",\"url\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/\",\"name\":\"What is Word Embedding | Word2Vec | GloVe\",\"isPartOf\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#primaryimage\"},\"image\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#primaryimage\"},\"thumbnailUrl\":\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg\",\"datePublished\":\"2020-07-12T09:55:58+00:00\",\"dateModified\":\"2024-09-02T11:47:15+00:00\",\"description\":\"Wha is Word Embedding or Text: We convert text into Word Embeddings so that the Machine learning algorithms can process it.Word2Vec and GloVe are pioneers when it comes to Word Embedding\",\"breadcrumb\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#primaryimage\",\"url\":\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg\",\"contentUrl\":\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg\",\"width\":1000,\"height\":700,\"caption\":\"word embedding\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Blog\",\"item\":\"https:\/\/www.mygreatlearning.com\/blog\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"AI and Machine Learning\",\"item\":\"https:\/\/www.mygreatlearning.com\/blog\/artificial-intelligence\/\"},{\"@type\":\"ListItem\",\"position\":3,\"name\":\"What is Word Embedding | Word2Vec | GloVe\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#website\",\"url\":\"https:\/\/www.mygreatlearning.com\/blog\/\",\"name\":\"Great Learning Blog\",\"description\":\"Learn, Upskill &amp; Career Development Guide and Resources\",\"publisher\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#organization\"},\"alternateName\":\"Great Learning\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/www.mygreatlearning.com\/blog\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#organization\",\"name\":\"Great Learning\",\"url\":\"https:\/\/www.mygreatlearning.com\/blog\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/logo\/image\/\",\"url\":\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2022\/06\/GL-Logo.jpg\",\"contentUrl\":\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2022\/06\/GL-Logo.jpg\",\"width\":900,\"height\":900,\"caption\":\"Great Learning\"},\"image\":{\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/logo\/image\/\"},\"sameAs\":[\"https:\/\/www.facebook.com\/GreatLearningOfficial\/\",\"https:\/\/x.com\/Great_Learning\",\"https:\/\/www.instagram.com\/greatlearningofficial\/\",\"https:\/\/www.linkedin.com\/school\/great-learning\/\",\"https:\/\/in.pinterest.com\/greatlearning12\/\",\"https:\/\/www.youtube.com\/user\/beaconelearning\/\"],\"description\":\"Great Learning is a leading global ed-tech company for professional training and higher education. It offers comprehensive, industry-relevant, hands-on learning programs across various business, technology, and interdisciplinary domains driving the digital economy. These programs are developed and offered in collaboration with the world's foremost academic institutions.\",\"email\":\"info@mygreatlearning.com\",\"legalName\":\"Great Learning Education Services Pvt. Ltd\",\"foundingDate\":\"2013-11-29\",\"numberOfEmployees\":{\"@type\":\"QuantitativeValue\",\"minValue\":\"1001\",\"maxValue\":\"5000\"}},{\"@type\":\"Person\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/person\/6f993d1be4c584a335951e836f2656ad\",\"name\":\"Great Learning Editorial Team\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2022\/02\/unnamed.webp\",\"contentUrl\":\"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2022\/02\/unnamed.webp\",\"caption\":\"Great Learning Editorial Team\"},\"description\":\"The Great Learning Editorial Staff includes a dynamic team of subject matter experts, instructors, and education professionals who combine their deep industry knowledge with innovative teaching methods. Their mission is to provide learners with the skills and insights needed to excel in their careers, whether through upskilling, reskilling, or transitioning into new fields.\",\"sameAs\":[\"https:\/\/www.mygreatlearning.com\/\",\"https:\/\/in.linkedin.com\/school\/great-learning\/\",\"https:\/\/x.com\/https:\/\/twitter.com\/Great_Learning\",\"https:\/\/www.youtube.com\/channel\/UCObs0kLIrDjX2LLSybqNaEA\"],\"award\":[\"Best EdTech Company of the Year 2024\",\"Education Economictimes Outstanding Education\/Edtech Solution Provider of the Year 2024\",\"Leading E-learning Platform 2024\"],\"url\":\"https:\/\/www.mygreatlearning.com\/blog\/author\/greatlearning\/\"}]}<\/script>\n<!-- \/ Yoast SEO Premium plugin. -->","yoast_head_json":{"title":"What is Word Embedding | Word2Vec | GloVe","description":"Wha is Word Embedding or Text: We convert text into Word Embeddings so that the Machine learning algorithms can process it.Word2Vec and GloVe are pioneers when it comes to Word Embedding","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/","og_locale":"en_US","og_type":"article","og_title":"What is Word Embedding | Word2Vec | GloVe","og_description":"Wha is Word Embedding or Text: We convert text into Word Embeddings so that the Machine learning algorithms can process it.Word2Vec and GloVe are pioneers when it comes to Word Embedding","og_url":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/","og_site_name":"Great Learning Blog: Free Resources what Matters to shape your Career!","article_publisher":"https:\/\/www.facebook.com\/GreatLearningOfficial\/","article_published_time":"2020-07-12T09:55:58+00:00","article_modified_time":"2024-09-02T11:47:15+00:00","og_image":[{"width":1000,"height":700,"url":"http:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg","type":"image\/jpeg"}],"author":"Great Learning Editorial Team","twitter_card":"summary_large_image","twitter_creator":"@https:\/\/twitter.com\/Great_Learning","twitter_site":"@Great_Learning","twitter_misc":{"Written by":"Great Learning Editorial Team","Est. reading time":"8 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#article","isPartOf":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/"},"author":{"name":"Great Learning Editorial Team","@id":"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/person\/6f993d1be4c584a335951e836f2656ad"},"headline":"What is Word Embedding | Word2Vec | GloVe","datePublished":"2020-07-12T09:55:58+00:00","dateModified":"2024-09-02T11:47:15+00:00","mainEntityOfPage":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/"},"wordCount":1418,"commentCount":1,"publisher":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/#organization"},"image":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#primaryimage"},"thumbnailUrl":"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg","articleSection":["AI and Machine Learning"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/","url":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/","name":"What is Word Embedding | Word2Vec | GloVe","isPartOf":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/#website"},"primaryImageOfPage":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#primaryimage"},"image":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#primaryimage"},"thumbnailUrl":"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg","datePublished":"2020-07-12T09:55:58+00:00","dateModified":"2024-09-02T11:47:15+00:00","description":"Wha is Word Embedding or Text: We convert text into Word Embeddings so that the Machine learning algorithms can process it.Word2Vec and GloVe are pioneers when it comes to Word Embedding","breadcrumb":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#primaryimage","url":"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg","contentUrl":"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg","width":1000,"height":700,"caption":"word embedding"},{"@type":"BreadcrumbList","@id":"https:\/\/www.mygreatlearning.com\/blog\/word-embedding\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Blog","item":"https:\/\/www.mygreatlearning.com\/blog\/"},{"@type":"ListItem","position":2,"name":"AI and Machine Learning","item":"https:\/\/www.mygreatlearning.com\/blog\/artificial-intelligence\/"},{"@type":"ListItem","position":3,"name":"What is Word Embedding | Word2Vec | GloVe"}]},{"@type":"WebSite","@id":"https:\/\/www.mygreatlearning.com\/blog\/#website","url":"https:\/\/www.mygreatlearning.com\/blog\/","name":"Great Learning Blog","description":"Learn, Upskill &amp; Career Development Guide and Resources","publisher":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/#organization"},"alternateName":"Great Learning","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.mygreatlearning.com\/blog\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/www.mygreatlearning.com\/blog\/#organization","name":"Great Learning","url":"https:\/\/www.mygreatlearning.com\/blog\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/logo\/image\/","url":"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2022\/06\/GL-Logo.jpg","contentUrl":"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2022\/06\/GL-Logo.jpg","width":900,"height":900,"caption":"Great Learning"},"image":{"@id":"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/GreatLearningOfficial\/","https:\/\/x.com\/Great_Learning","https:\/\/www.instagram.com\/greatlearningofficial\/","https:\/\/www.linkedin.com\/school\/great-learning\/","https:\/\/in.pinterest.com\/greatlearning12\/","https:\/\/www.youtube.com\/user\/beaconelearning\/"],"description":"Great Learning is a leading global ed-tech company for professional training and higher education. It offers comprehensive, industry-relevant, hands-on learning programs across various business, technology, and interdisciplinary domains driving the digital economy. These programs are developed and offered in collaboration with the world's foremost academic institutions.","email":"info@mygreatlearning.com","legalName":"Great Learning Education Services Pvt. Ltd","foundingDate":"2013-11-29","numberOfEmployees":{"@type":"QuantitativeValue","minValue":"1001","maxValue":"5000"}},{"@type":"Person","@id":"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/person\/6f993d1be4c584a335951e836f2656ad","name":"Great Learning Editorial Team","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/www.mygreatlearning.com\/blog\/#\/schema\/person\/image\/","url":"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2022\/02\/unnamed.webp","contentUrl":"https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2022\/02\/unnamed.webp","caption":"Great Learning Editorial Team"},"description":"The Great Learning Editorial Staff includes a dynamic team of subject matter experts, instructors, and education professionals who combine their deep industry knowledge with innovative teaching methods. Their mission is to provide learners with the skills and insights needed to excel in their careers, whether through upskilling, reskilling, or transitioning into new fields.","sameAs":["https:\/\/www.mygreatlearning.com\/","https:\/\/in.linkedin.com\/school\/great-learning\/","https:\/\/x.com\/https:\/\/twitter.com\/Great_Learning","https:\/\/www.youtube.com\/channel\/UCObs0kLIrDjX2LLSybqNaEA"],"award":["Best EdTech Company of the Year 2024","Education Economictimes Outstanding Education\/Edtech Solution Provider of the Year 2024","Leading E-learning Platform 2024"],"url":"https:\/\/www.mygreatlearning.com\/blog\/author\/greatlearning\/"}]}},"uagb_featured_image_src":{"full":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg",1000,700,false],"thumbnail":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032-150x150.jpg",150,150,true],"medium":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032-300x210.jpg",300,210,true],"medium_large":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032-768x538.jpg",768,538,true],"large":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg",1000,700,false],"1536x1536":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg",1000,700,false],"2048x2048":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg",1000,700,false],"web-stories-poster-portrait":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg",640,448,false],"web-stories-publisher-logo":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg",96,67,false],"web-stories-thumbnail":["https:\/\/www.mygreatlearning.com\/blog\/wp-content\/uploads\/2020\/07\/shutterstock_1195373032.jpg",150,105,false]},"uagb_author_info":{"display_name":"Great Learning Editorial Team","author_link":"https:\/\/www.mygreatlearning.com\/blog\/author\/greatlearning\/"},"uagb_comment_info":2,"uagb_excerpt":"What is Word Embedding? Humans have always excelled at understanding languages. It is easy for humans to understand the relationship between words but for computers, this task may not be simple. For example, we humans understand the words like king and queen, man and woman, tiger and tigress have a certain type of relation between&hellip;","_links":{"self":[{"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/posts\/16701","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/users\/41"}],"replies":[{"embeddable":true,"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/comments?post=16701"}],"version-history":[{"count":9,"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/posts\/16701\/revisions"}],"predecessor-version":[{"id":97723,"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/posts\/16701\/revisions\/97723"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/media\/16702"}],"wp:attachment":[{"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/media?parent=16701"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/categories?post=16701"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/tags?post=16701"},{"taxonomy":"content_type","embeddable":true,"href":"https:\/\/www.mygreatlearning.com\/blog\/wp-json\/wp\/v2\/content_type?post=16701"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}