978-981-99-1600-9.pdf

This book provides an overview of the recent advances in representation learning theory, algorithms, and applications for natural language processing (NLP), ranging from word embeddings to pre-trained language models. It is divided into four parts. Part I presents the representation learning techniq...

Πλήρης περιγραφή

Λεπτομέρειες βιβλιογραφικής εγγραφής
Γλώσσα:English
Έκδοση: Springer Nature 2023
Διαθέσιμο Online:https://link.springer.com/978-981-99-1600-9
id oapen-20.500.12657-76271
record_format dspace
spelling oapen-20.500.12657-762712023-09-14T03:36:11Z Representation Learning for Natural Language Processing Liu, Zhiyuan Lin, Yankai Sun, Maosong Deep Learning Representation Learning Knowledge Representation Word Representation Document Representation Big Data Machine Learning Natural Language Processing Artificial Intelligence bic Book Industry Communication::U Computing & information technology::UY Computer science::UYQ Artificial intelligence::UYQL Natural language & machine translation bic Book Industry Communication::C Language::CF linguistics::CFX Computational linguistics bic Book Industry Communication::U Computing & information technology::UY Computer science::UYQ Artificial intelligence bic Book Industry Communication::U Computing & information technology::UN Databases::UNF Data mining This book provides an overview of the recent advances in representation learning theory, algorithms, and applications for natural language processing (NLP), ranging from word embeddings to pre-trained language models. It is divided into four parts. Part I presents the representation learning techniques for multiple language entries, including words, sentences and documents, as well as pre-training techniques. Part II then introduces the related representation techniques to NLP, including graphs, cross-modal entries, and robustness. Part III then introduces the representation techniques for the knowledge that are closely related to NLP, including entity-based world knowledge, sememe-based linguistic knowledge, legal domain knowledge and biomedical domain knowledge. Lastly, Part IV discusses the remaining challenges and future research directions. The theories and algorithms of representation learning presented can also benefit other related domains such as machine learning, social network analysis, semantic Web, information retrieval, data mining and computational biology. This book is intended for advanced undergraduate and graduate students, post-doctoral fellows, researchers, lecturers, and industrial engineers, as well as anyone interested in representation learning and natural language processing. As compared to the first edition, the second edition (1) provides a more detailed introduction to representation learning in Chapter 1; (2) adds four new chapters to introduce pre-trained language models, robust representation learning, legal knowledge representation learning and biomedical knowledge representation learning; (3) updates recent advances in representation learning in all chapters; and (4) corrects some errors in the first edition. The new contents will be approximately 50%+ compared to the first edition. This is an open access book. 2023-09-13T19:48:07Z 2023-09-13T19:48:07Z 2023 book ONIX_20230913_9789819916009_38 9789819916009 9789819915996 https://library.oapen.org/handle/20.500.12657/76271 eng application/pdf n/a 978-981-99-1600-9.pdf https://link.springer.com/978-981-99-1600-9 Springer Nature Springer Nature Singapore 10.1007/978-981-99-1600-9 10.1007/978-981-99-1600-9 6c6992af-b843-4f46-859c-f6e9998e40d5 e06840e4-106f-423f-bba7-d034dee9cf25 9789819916009 9789819915996 Springer Nature Singapore 521 Singapore [...] Tsinghua University THU open access
institution OAPEN
collection DSpace
language English
description This book provides an overview of the recent advances in representation learning theory, algorithms, and applications for natural language processing (NLP), ranging from word embeddings to pre-trained language models. It is divided into four parts. Part I presents the representation learning techniques for multiple language entries, including words, sentences and documents, as well as pre-training techniques. Part II then introduces the related representation techniques to NLP, including graphs, cross-modal entries, and robustness. Part III then introduces the representation techniques for the knowledge that are closely related to NLP, including entity-based world knowledge, sememe-based linguistic knowledge, legal domain knowledge and biomedical domain knowledge. Lastly, Part IV discusses the remaining challenges and future research directions. The theories and algorithms of representation learning presented can also benefit other related domains such as machine learning, social network analysis, semantic Web, information retrieval, data mining and computational biology. This book is intended for advanced undergraduate and graduate students, post-doctoral fellows, researchers, lecturers, and industrial engineers, as well as anyone interested in representation learning and natural language processing. As compared to the first edition, the second edition (1) provides a more detailed introduction to representation learning in Chapter 1; (2) adds four new chapters to introduce pre-trained language models, robust representation learning, legal knowledge representation learning and biomedical knowledge representation learning; (3) updates recent advances in representation learning in all chapters; and (4) corrects some errors in the first edition. The new contents will be approximately 50%+ compared to the first edition. This is an open access book.
title 978-981-99-1600-9.pdf
spellingShingle 978-981-99-1600-9.pdf
title_short 978-981-99-1600-9.pdf
title_full 978-981-99-1600-9.pdf
title_fullStr 978-981-99-1600-9.pdf
title_full_unstemmed 978-981-99-1600-9.pdf
title_sort 978-981-99-1600-9.pdf
publisher Springer Nature
publishDate 2023
url https://link.springer.com/978-981-99-1600-9
_version_ 1799945258094559232