Generalized Domain Adaptation for Sequence Labeling in Natural Language Processing

Generalized Domain Adaptation for Sequence Labeling in Natural Language Processing

Author: Min Xiao

Publisher:

Published: 2016

Total Pages: 100

ISBN-13:

DOWNLOAD EBOOK

Sequence labeling tasks have been widely studied in the natural language processing area, such as part-of-speech tagging, syntactic chunking, dependency parsing, and etc. Most of those systems are developed on a large amount of labeled training data via supervised learning. However, manually collecting labeled training data is too time-consuming and expensive. As an alternative, to alleviate the issue of label scarcity, domain adaptation has recently been proposed to train a statistical machine learning model in a target domain where there is no enough labeled training data by exploiting existing free labeled training data in a different but related source domain. The natural language processing community has witnessed the success of domain adaptation in a variety of sequence labeling tasks. Though the labeled training data in the source domain are available and free, however, they are not exactly as and can be very different from the test data in the target domain. Thus, simply applying naive supervised machine learning algorithms without considering domain differences may not fulfill the purpose. In this dissertation, we developed several novel representation learning approaches to address domain adaptation for sequence labeling in natural language processing. Those representation learning techniques aim to induce latent generalizable features to bridge domain divergence to enable cross-domain prediction. We first tackle a semi-supervised domain adaptation scenario where the target domain has a small amount of labeled training data and propose a distributed representation learning approach based on a probabilistic neural language model. We then relax the assumption of the availability of labeled training data in the target domain and study an unsupervised domain adaptation scenario where the target domain has only unlabeled training data, and give a task-informative representation learning approach based on dynamic dependency networks. Both works are developed in the setting where different domains contain sentences in different genres. We then extend and generalize domain adaptation into a more challenging scenario where different domains contain sentences in different languages and propose two cross-lingual representation learning approaches, one is based on deep neural networks with auxiliary bilingual word pairs and the other is based on annotation projection with auxiliary parallel sentences. All four specific learning scenarios are extensively evaluated with different sequence labeling tasks. The empirical results demonstrate the effectiveness of those generalized domain adaptation techniques for sequence labeling in natural language processing.


Biomedical Natural Language Processing

Biomedical Natural Language Processing

Author: Kevin Bretonnel Cohen

Publisher: John Benjamins Publishing Company

Published: 2014-02-15

Total Pages: 174

ISBN-13: 9027271062

DOWNLOAD EBOOK

Biomedical Natural Language Processing is a comprehensive tour through the classic and current work in the field. It discusses all subjects from both a rule-based and a machine learning approach, and also describes each subject from the perspective of both biological science and clinical medicine. The intended audience is readers who already have a background in natural language processing, but a clear introduction makes it accessible to readers from the fields of bioinformatics and computational biology, as well. The book is suitable as a reference, as well as a text for advanced courses in biomedical natural language processing and text mining.


Spoken Language Understanding

Spoken Language Understanding

Author: Gokhan Tur

Publisher: John Wiley & Sons

Published: 2011-05-03

Total Pages: 443

ISBN-13: 1119993946

DOWNLOAD EBOOK

Spoken language understanding (SLU) is an emerging field in between speech and language processing, investigating human/ machine and human/ human communication by leveraging technologies from signal processing, pattern recognition, machine learning and artificial intelligence. SLU systems are designed to extract the meaning from speech utterances and its applications are vast, from voice search in mobile devices to meeting summarization, attracting interest from both commercial and academic sectors. Both human/machine and human/human communications can benefit from the application of SLU, using differing tasks and approaches to better understand and utilize such communications. This book covers the state-of-the-art approaches for the most popular SLU tasks with chapters written by well-known researchers in the respective fields. Key features include: Presents a fully integrated view of the two distinct disciplines of speech processing and language processing for SLU tasks. Defines what is possible today for SLU as an enabling technology for enterprise (e.g., customer care centers or company meetings), and consumer (e.g., entertainment, mobile, car, robot, or smart environments) applications and outlines the key research areas. Provides a unique source of distilled information on methods for computer modeling of semantic information in human/machine and human/human conversations. This book can be successfully used for graduate courses in electronics engineering, computer science or computational linguistics. Moreover, technologists interested in processing spoken communications will find it a useful source of collated information of the topic drawn from the two distinct disciplines of speech processing and language processing under the new area of SLU.


Practical Natural Language Processing

Practical Natural Language Processing

Author: Sowmya Vajjala

Publisher: "O'Reilly Media, Inc."

Published: 2020-06-17

Total Pages: 456

ISBN-13: 1492054003

DOWNLOAD EBOOK

Many books and courses tackle natural language processing (NLP) problems with toy use cases and well-defined datasets. But if you want to build, iterate, and scale NLP systems in a business setting and tailor them for particular industry verticals, this is your guide. Software engineers and data scientists will learn how to navigate the maze of options available at each step of the journey. Through the course of the book, authors Sowmya Vajjala, Bodhisattwa Majumder, Anuj Gupta, and Harshit Surana will guide you through the process of building real-world NLP solutions embedded in larger product setups. You’ll learn how to adapt your solutions for different industry verticals such as healthcare, social media, and retail. With this book, you’ll: Understand the wide spectrum of problem statements, tasks, and solution approaches within NLP Implement and evaluate different NLP applications using machine learning and deep learning methods Fine-tune your NLP solution based on your business problem and industry vertical Evaluate various algorithms and approaches for NLP product tasks, datasets, and stages Produce software solutions following best practices around release, deployment, and DevOps for NLP systems Understand best practices, opportunities, and the roadmap for NLP from a business and product leader’s perspective


Transfer Learning and Robustness for Natural Language Processing

Transfer Learning and Robustness for Natural Language Processing

Author: Di Jin (Ph.D.)

Publisher:

Published: 2020

Total Pages: 217

ISBN-13:

DOWNLOAD EBOOK

Teaching machines to understand human language is one of the most elusive and long-standing challenges in Natural Language Processing (NLP). Driven by the fast development of deep learning, state-of-the-art NLP models have already achieved human-level performance in various large benchmark datasets, such as SQuAD, SNLI, and RACE. However, when these strong models are deployed to real-world applications, they often show poor generalization capability in two situations: 1. There is only a limited amount of data available for model training; 2. Deployed models may degrade significantly in performance on noisy test data or natural/artificial adversaries. In short, performance degradation on low-resource tasks/datasets and unseen data with distribution shifts imposes great challenges to the reliability of NLP models and prevent them from being massively applied in the wild. This dissertation aims to address these two issues. Towards the first one, we resort to transfer learning to leverage knowledge acquired from related data in order to improve performance on a target low-resource task/dataset. Specifically, we propose different transfer learning methods for three natural language understanding tasks: multi-choice question answering, dialogue state tracking, and sequence labeling, and one natural language generation task: machine translation. These methods are based on four basic transfer learning modalities: multi-task learning, sequential transfer learning, domain adaptation, and cross-lingual transfer. We show experimental results to validate that transferring knowledge from related domains, tasks, and languages can improve the target task/dataset significantly. For the second issue, we propose methods to evaluate the robustness of NLP models on text classification and entailment tasks. On one hand, we reveal that although these models can achieve a high accuracy of over 90%, they still easily crash over paraphrases of original samples by changing only around 10% words to their synonyms. On the other hand, by creating a new challenge set using four adversarial strategies, we find even the best models for the aspect-based sentiment analysis task cannot reliably identify the target aspect and recognize its sentiment accordingly. On the contrary, they are easily confused by distractor aspects. Overall, these findings raise great concerns of robustness of NLP models, which should be enhanced to ensure their long-run stable service.


Representation Learning for Natural Language Processing

Representation Learning for Natural Language Processing

Author: Zhiyuan Liu

Publisher: Springer Nature

Published: 2020-07-03

Total Pages: 319

ISBN-13: 9811555737

DOWNLOAD EBOOK

This open access book provides an overview of the recent advances in representation learning theory, algorithms and applications for natural language processing (NLP). It is divided into three parts. Part I presents the representation learning techniques for multiple language entries, including words, phrases, sentences and documents. Part II then introduces the representation techniques for those objects that are closely related to NLP, including entity-based world knowledge, sememe-based linguistic knowledge, networks, and cross-modal entries. Lastly, Part III provides open resource tools for representation learning techniques, and discusses the remaining challenges and future research directions. The theories and algorithms of representation learning presented can also benefit other related domains such as machine learning, social network analysis, semantic Web, information retrieval, data mining and computational biology. This book is intended for advanced undergraduate and graduate students, post-doctoral fellows, researchers, lecturers, and industrial engineers, as well as anyone interested in representation learning and natural language processing.


Natural Language Processing and Chinese Computing

Natural Language Processing and Chinese Computing

Author: Jie Tang

Publisher: Springer Nature

Published: 2019-09-30

Total Pages: 850

ISBN-13: 303032236X

DOWNLOAD EBOOK

This two-volume set of LNAI 11838 and LNAI 11839 constitutes the refereed proceedings of the 8th CCF Conference on Natural Language Processing and Chinese Computing, NLPCC 2019, held in Dunhuang, China, in October 2019. The 85 full papers and 56 short papers presented were carefully reviewed and selected from 492 submissions. They are organized in the following topical sections: Conversational Bot/QA/IR; Knowledge graph/IE; Machine Learning for NLP; Machine Translation; NLP Applications; NLP for Social Network; NLP Fundamentals; Text Mining; Short Papers; Explainable AI Workshop; Student Workshop: Evaluation Workshop.


Clinical Text Mining

Clinical Text Mining

Author: Hercules Dalianis

Publisher: Springer

Published: 2018-05-14

Total Pages: 192

ISBN-13: 3319785036

DOWNLOAD EBOOK

This open access book describes the results of natural language processing and machine learning methods applied to clinical text from electronic patient records. It is divided into twelve chapters. Chapters 1-4 discuss the history and background of the original paper-based patient records, their purpose, and how they are written and structured. These initial chapters do not require any technical or medical background knowledge. The remaining eight chapters are more technical in nature and describe various medical classifications and terminologies such as ICD diagnosis codes, SNOMED CT, MeSH, UMLS, and ATC. Chapters 5-10 cover basic tools for natural language processing and information retrieval, and how to apply them to clinical text. The difference between rule-based and machine learning-based methods, as well as between supervised and unsupervised machine learning methods, are also explained. Next, ethical concerns regarding the use of sensitive patient records for research purposes are discussed, including methods for de-identifying electronic patient records and safely storing patient records. The book’s closing chapters present a number of applications in clinical text mining and summarise the lessons learned from the previous chapters. The book provides a comprehensive overview of technical issues arising in clinical text mining, and offers a valuable guide for advanced students in health informatics, computational linguistics, and information retrieval, and for researchers entering these fields.


Knowledge Graph and Semantic Computing: Knowledge Computing and Language Understanding

Knowledge Graph and Semantic Computing: Knowledge Computing and Language Understanding

Author: Xiaoyan Zhu

Publisher: Springer Nature

Published: 2020-01-03

Total Pages: 226

ISBN-13: 9811519560

DOWNLOAD EBOOK

This book constitutes the refereed proceedings of the 4th China Conference on Knowledge Graph and Semantic Computing, CCKS 2019, held in Hangzhou, China, in August 2019. The 18 revised full papers presented were carefully reviewed and selected from 140 submissions. The papers cover wide research fields including the knowledge graph, the semantic Web, linked data, NLP, information extraction, knowledge representation and reasoning.