Concept information
Término preferido
word embedding
Definición
- Process by which words are mapped into real-valued vectors. (ARTES)
Concepto genérico
Conceptos específicos
Etiquetas alternativas
- word embeddings
Nota
- Each word in the vocabulary is represented by a vector w ∈ RD, where D is the dimension fixed in advance. One of the major advantages of representing words as vectors is the fact that standard similarity measures such as cosine similarity or Euclidean distance can be used, enabling semantic distances to be calculated between words. Contrary to what we may be led to think by the recent popularity surge for word embeddings, the use of compact, vectorial word representations is by no means new, and the theoretical underpinnings can be traced back at least to the 1950s and the theory of distributional semantics. The distributional hypothesis, the idea that you can define a word by the company it keeps (Harris, 1954), popularised in the 1950s by philosophers and linguists such as Harris (1954), Firth (1957) and Wittgenstein (1953), has been influential in the way textual input is represented in NLP [...) (Bawden, Rachel, Going beyond the sentence : Contextual Machine Translation of Dialogue, Université Paris-Saclay, 2018)
En otras lenguas
-
francés
-
plongement de mot
-
plongement de mots
-
représentation distribuée de mots
URI
http://data.loterre.fr/ark:/67375/8LP-M1NFTGZ7-1
{{label}}
{{#each values }} {{! loop through ConceptPropertyValue objects }}
{{#if prefLabel }}
{{/if}}
{{/each}}
{{#if notation }}{{ notation }} {{/if}}{{ prefLabel }}
{{#ifDifferentLabelLang lang }} ({{ lang }}){{/ifDifferentLabelLang}}
{{#if vocabName }}
{{ vocabName }}
{{/if}}