Store besparelser
Hurtig levering
Gemte
Log ind
0
Kurv
Kurv

Embeddings in Natural Language Processing

- Theory and Advances in Vector Representations of Meaning

Embeddings in Natural Language Processing

- Theory and Advances in Vector Representations of Meaning
Tjek vores konkurrenters priser
Embeddings have undoubtedly been one of the most influential research areas in Natural Language Processing (NLP). Encoding information into a low-dimensional vector representation, which is easily integrable in modern machine learning models, has played a central role in the development of NLP. Embedding techniques initially focused on words, but the attention soon started to shift to other forms: from graph structures, such as knowledge bases, to other types of textual content, such as sentences and documents.This book provides a high-level synthesis of the main embedding techniques in NLP, in the broad sense. The book starts by explaining conventional word vector space models and word embeddings (e.g., Word2Vec and GloVe) and then moves to other types of embeddings, such as word sense, sentence and document, and graph embeddings. The book also provides an overview of recent developments in contextualized representations (e.g., ELMo and BERT) and explains their potential in NLP.Throughout the book, the reader can find both essential information for understanding a certain topic from scratch and a broad overview of the most successful techniques developed in the literature.
Tjek vores konkurrenters priser
Normalpris
kr 526
Fragt: 39 kr
6 - 8 hverdage
20 kr
Pakkegebyr
God 4 anmeldelser på
Tjek vores konkurrenters priser
Embeddings have undoubtedly been one of the most influential research areas in Natural Language Processing (NLP). Encoding information into a low-dimensional vector representation, which is easily integrable in modern machine learning models, has played a central role in the development of NLP. Embedding techniques initially focused on words, but the attention soon started to shift to other forms: from graph structures, such as knowledge bases, to other types of textual content, such as sentences and documents.This book provides a high-level synthesis of the main embedding techniques in NLP, in the broad sense. The book starts by explaining conventional word vector space models and word embeddings (e.g., Word2Vec and GloVe) and then moves to other types of embeddings, such as word sense, sentence and document, and graph embeddings. The book also provides an overview of recent developments in contextualized representations (e.g., ELMo and BERT) and explains their potential in NLP.Throughout the book, the reader can find both essential information for understanding a certain topic from scratch and a broad overview of the most successful techniques developed in the literature.
Produktdetaljer
Sprog: Engelsk
Sider: 157
ISBN-13: 9783031010491
Indbinding: Paperback
Udgave:
ISBN-10: 3031010493
Udg. Dato: 13 nov 2020
Længde: 0mm
Bredde: 191mm
Højde: 235mm
Forlag: Springer International Publishing AG
Oplagsdato: 13 nov 2020
Forfatter(e) Jose Camacho-Collados, Mohammad Taher Pilehvar


Kategori Naturligt sprog og maskinoversættelse


ISBN-13 9783031010491


Sprog Engelsk


Indbinding Paperback


Sider 157


Udgave


Længde 0mm


Bredde 191mm


Højde 235mm


Udg. Dato 13 nov 2020


Oplagsdato 13 nov 2020


Forlag Springer International Publishing AG