You are on page 1of 53

Deep Learning in Natural Language

Processing 1st Edition Li Deng


Visit to download the full and correct content document:
https://textbookfull.com/product/deep-learning-in-natural-language-processing-1st-edit
ion-li-deng/
More products digital (pdf, epub, mobi) instant
download maybe you interests ...

Deep learning in natural language processing Deng

https://textbookfull.com/product/deep-learning-in-natural-
language-processing-deng/

Deep Learning for Natural Language Processing Develop


Deep Learning Models for Natural Language in Python
Jason Brownlee

https://textbookfull.com/product/deep-learning-for-natural-
language-processing-develop-deep-learning-models-for-natural-
language-in-python-jason-brownlee/

Python Natural Language Processing Advanced machine


learning and deep learning techniques for natural
language processing 1st Edition Jalaj Thanaki

https://textbookfull.com/product/python-natural-language-
processing-advanced-machine-learning-and-deep-learning-
techniques-for-natural-language-processing-1st-edition-jalaj-
thanaki/

Applied Natural Language Processing with Python:


Implementing Machine Learning and Deep Learning
Algorithms for Natural Language Processing 1st Edition
Taweh Beysolow Ii
https://textbookfull.com/product/applied-natural-language-
processing-with-python-implementing-machine-learning-and-deep-
learning-algorithms-for-natural-language-processing-1st-edition-
Deep Learning for Natural Language Processing (MEAP
V07) Stephan Raaijmakers

https://textbookfull.com/product/deep-learning-for-natural-
language-processing-meap-v07-stephan-raaijmakers/

Natural language processing with TensorFlow Teach


language to machines using Python s deep learning
library 1st Edition Thushan Ganegedara

https://textbookfull.com/product/natural-language-processing-
with-tensorflow-teach-language-to-machines-using-python-s-deep-
learning-library-1st-edition-thushan-ganegedara/

Natural Language Processing 1st Edition Jacob


Eisenstein

https://textbookfull.com/product/natural-language-processing-1st-
edition-jacob-eisenstein/

Machine Learning with PySpark: With Natural Language


Processing and Recommender Systems 1st Edition Pramod
Singh

https://textbookfull.com/product/machine-learning-with-pyspark-
with-natural-language-processing-and-recommender-systems-1st-
edition-pramod-singh/

Natural Language Processing in Artificial Intelligence


1st Edition Brojo Kishore Mishra

https://textbookfull.com/product/natural-language-processing-in-
artificial-intelligence-1st-edition-brojo-kishore-mishra/
Li Deng · Yang Liu Editors

Deep Learning
in Natural
Language
Processing
Deep Learning in Natural Language Processing
Li Deng Yang Liu

Editors

Deep Learning in Natural


Language Processing

123
Editors
Li Deng Yang Liu
AI Research at Citadel Tsinghua University
Chicago, IL Beijing
USA China

and

AI Research at Citadel
Seattle, WA
USA

ISBN 978-981-10-5208-8 ISBN 978-981-10-5209-5 (eBook)


https://doi.org/10.1007/978-981-10-5209-5
Library of Congress Control Number: 2018934459

© Springer Nature Singapore Pte Ltd. 2018


This work is subject to copyright. All rights are reserved by the Publisher, whether the whole or part
of the material is concerned, specifically the rights of translation, reprinting, reuse of illustrations,
recitation, broadcasting, reproduction on microfilms or in any other physical way, and transmission
or information storage and retrieval, electronic adaptation, computer software, or by similar or dissimilar
methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this
publication does not imply, even in the absence of a specific statement, that such names are exempt from
the relevant protective laws and regulations and therefore free for general use.
The publisher, the authors and the editors are safe to assume that the advice and information in this
book are believed to be true and accurate at the date of publication. Neither the publisher nor the
authors or the editors give a warranty, express or implied, with respect to the material contained herein or
for any errors or omissions that may have been made. The publisher remains neutral with regard to
jurisdictional claims in published maps and institutional affiliations.

Printed on acid-free paper

This Springer imprint is published by the registered company Springer Nature Singapore Pte Ltd.
part of Springer Nature
The registered company address is: 152 Beach Road, #21-01/04 Gateway East, Singapore 189721,
Singapore
Foreword

“Written by a group of the most active researchers in the field, led by Dr. Deng, an
internationally respected expert in both NLP and deep learning, this book provides
a comprehensive introduction to and up-to-date review of the state of art in applying
deep learning to solve fundamental problems in NLP. Further, the book is highly
timely, as demands for high-quality and up-to-date textbooks and research refer-
ences have risen dramatically in response to the tremendous strides in deep learning
applications to NLP. The book offers a unique reference guide for practitioners in
various sectors, especially the Internet and AI start-ups, where NLP technologies
are becoming an essential enabler and a core differentiator.”
Hongjiang Zhang (Founder, Sourcecode Capital; former CEO of KingSoft)
“This book provides a comprehensive introduction to the latest advances in deep
learning applied to NLP. Written by experienced and aspiring deep learning and
NLP researchers, it covers a broad range of major NLP applications, including
spoken language understanding, dialog systems, lexical analysis, parsing, knowl-
edge graph, machine translation, question answering, sentiment analysis, and social
computing.
The book is clearly structured and moves from major research trends, to the
latest deep learning approaches, to their limitations and promising future work.
Given its self-contained content, sophisticated algorithms, and detailed use cases,
the book offers a valuable guide for all readers who are working on or learning
about deep learning and NLP.”
Haifeng Wang (Vice President and Head of Research, Baidu; former President
of ACL)
“In 2011, at the dawn of deep learning in industry, I estimated that in most speech
recognition applications, computers still made 5 to 10 times more errors than human
subjects, and highlighted the importance of knowledge engineering in future
directions. Within only a handful of years since, deep learning has nearly closed the
gap in the accuracy of conversational speech recognition between human and
computers. Edited and written by Dr. Li Deng—a pioneer in the recent speech

v
vi Foreword

recognition revolution using deep learning—and his colleagues, this book elegantly
describes this part of the fascinating history of speech recognition as an important
subfield of natural language processing (NLP). Further, the book expands this
historical perspective from speech recognition to more general areas of NLP,
offering a truly valuable guide for the future development of NLP.
Importantly, the book puts forward a thesis that the current deep learning trend is
a revolution from the previous data-driven (shallow) machine learning era, although
ostensibly deep learning appears to be merely exploiting more data, more com-
puting power, and more complex models. Indeed, as the book correctly points out,
the current state of the art of deep learning technology developed for NLP appli-
cations, despite being highly successful in solving individual NLP tasks, has not
taken full advantage of rich world knowledge or human cognitive capabilities.
Therefore, I fully embrace the view expressed by the book’s editors and authors that
more advanced deep learning that seamlessly integrates knowledge engineering will
pave the way for the next revolution in NLP.
I highly recommend speech and NLP researchers, engineers, and students to read
this outstanding and timely book, not only to learn about the state of the art in NLP
and deep learning, but also to gain vital insights into what the future of the NLP
field will hold.”
Sadaoki Furui (President, Toyota Technological Institute at Chicago)
Preface

Natural language processing (NLP), which aims to enable computers to process


human languages intelligently, is an important interdisciplinary field crossing
artificial intelligence, computing science, cognitive science, information processing,
and linguistics. Concerned with interactions between computers and human lan-
guages, NLP applications such as speech recognition, dialog systems, information
retrieval, question answering, and machine translation have started to reshape the
way people identify, obtain, and make use of information.
The development of NLP can be described in terms of three major waves:
rationalism, empiricism, and deep learning. In the first wave, rationalist approaches
advocated the design of handcrafted rules to incorporate knowledge into NLP
systems based on the assumption that knowledge of language in the human mind is
fixed in advance by generic inheritance. In the second wave, empirical approaches
assume that rich sensory input and the observable language data in surface form are
required and sufficient to enable the mind to learn the detailed structure of natural
language. As a result, probabilistic models were developed to discover the regu-
larities of languages from large corpora. In the third wave, deep learning exploits
hierarchical models of nonlinear processing, inspired by biological neural systems
to learn intrinsic representations from language data, in ways that aim to simulate
human cognitive abilities.
The intersection of deep learning and natural language processing has resulted in
striking successes in practical tasks. Speech recognition is the first industrial NLP
application that deep learning has strongly impacted. With the availability of
large-scale training data, deep neural networks achieved dramatically lower
recognition errors than the traditional empirical approaches. Another prominent
successful application of deep learning in NLP is machine translation. End-to-end
neural machine translation that models the mapping between human languages
using neural networks has proven to improve translation quality substantially.
Therefore, neural machine translation has quickly become the new de facto tech-
nology in major commercial online translation services offered by large technology
companies: Google, Microsoft, Facebook, Baidu, and more. Many other areas of
NLP, including language understanding and dialog, lexical analysis and parsing,

vii
viii Preface

knowledge graph, information retrieval, question answering from text, social


computing, language generation, and text sentiment analysis, have also seen much
significant progress using deep learning, riding on the third wave of
NLP. Nowadays, deep learning is a dominating method applied to practically all
NLP tasks.
The main goal of this book is to provide a comprehensive survey on the recent
advances in deep learning applied to NLP. The book presents state of the art of
NLP-centric deep learning research, and focuses on the role of deep learning played
in major NLP applications including spoken language understanding, dialog sys-
tems, lexical analysis, parsing, knowledge graph, machine translation, question
answering, sentiment analysis, social computing, and natural language generation
(from images). This book is suitable for readers with a technical background in
computation, including graduate students, post-doctoral researchers, educators, and
industrial researchers and anyone interested in getting up to speed with the latest
techniques of deep learning associated with NLP.
The book is organized into eleven chapters as follows:
• Chapter 1: A Joint Introduction to Natural Language Processing and to Deep
Learning (Li Deng and Yang Liu)
• Chapter 2: Deep Learning in Conversational Language Understanding (Gokhan
Tur, Asli Celikyilmaz, Xiaodong He, Dilek Hakkani-Tür, and Li Deng)
• Chapter 3: Deep Learning in Spoken and Text-Based Dialog Systems
(Asli Celikyilmaz, Li Deng, and Dilek Hakkani-Tür)
• Chapter 4: Deep Learning in Lexical Analysis and Parsing (Wanxiang Che and
Yue Zhang)
• Chapter 5: Deep Learning in Knowledge Graph (Zhiyuan Liu and Xianpei Han)
• Chapter 6: Deep Learning in Machine Translation (Yang Liu and Jiajun Zhang)
• Chapter 7: Deep Learning in Question Answering (Kang Liu and Yansong Feng)
• Chapter 8: Deep Learning in Sentiment Analysis (Duyu Tang and Meishan
Zhang)
• Chapter 9: Deep Learning in Social Computing (Xin Zhao and Chenliang Li)
• Chapter 10: Deep Learning in Natural Language Generation from Images
(Xiaodong He and Li Deng)
• Chapter 11: Epilogue (Li Deng and Yang Liu)
Chapter 1 first reviews the basics of NLP as well as the main scope of NLP
covered in the following chapters of the book, and then goes in some depth into the
historical development of NLP summarized as three waves and future directions.
Subsequently, in Chaps. 2–10, an in-depth survey on the recent advances in deep
learning applied to NLP is organized into nine separate chapters, each covering a
largely independent application area of NLP. The main body of each chapter is
written by leading researchers and experts actively working in the respective field.
The origin of this book was the set of comprehensive tutorials given at the 15th
China National Conference on Computational Linguistics (CCL 2016) held in
October 2016 in Yantai, Shandong, China, where both of us, editors of this book,
Preface ix

were active participants and were taking leading roles. We thank our Springer’s
senior editor, Dr. Celine Lanlan Chang, who kindly invited us to create this book
and who has been providing much of timely assistance needed to complete this
book. We are grateful also to Springer’s Assistant Editor, Jane Li, for offering
invaluable help through various stages of manuscript preparation.
We thank all authors of Chaps. 2–10 who devoted their valuable time carefully
preparing the content of their chapters: Gokhan Tur, Asli Celikyilmaz, Dilek
Hakkani-Tur, Wanxiang Che, Yue Zhang, Xianpei Han, Zhiyuan Liu, Jiajun Zhang,
Kang Liu, Yansong Feng, Duyu Tang, Meishan Zhang, Xin Zhao, Chenliang Li,
and Xiaodong He. The authors of Chaps. 4–9 are CCL 2016 tutorial speakers. They
spent a considerable amount of time in updating their tutorial material with the
latest advances in the field since October 2016.
Further, we thank numerous reviewers and readers, Sadaoki Furui, Andrew Ng,
Fred Juang, Ken Church, Haifeng Wang, and Hongjiang Zhang, who not only gave
us much needed encouragements but also offered many constructive comments
which substantially improved earlier drafts of the book.
Finally, we give our appreciations to our organizations, Microsoft Research and
Citadel (for Li Deng) and Tsinghua University (for Yang Liu), who provided
excellent environments, supports, and encouragements that have been instrumental
for us to complete this book. Yang Liu is also supported by National Natural
Science Foundation of China (No.61522204, No.61432013, and No.61331013).

Seattle, USA Li Deng


Beijing, China Yang Liu
October 2017
Contents

1 A Joint Introduction to Natural Language Processing and to


Deep Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Li Deng and Yang Liu
2 Deep Learning in Conversational Language Understanding . . . . . . 23
Gokhan Tur, Asli Celikyilmaz, Xiaodong He, Dilek Hakkani-Tür
and Li Deng
3 Deep Learning in Spoken and Text-Based Dialog Systems . . . . . . 49
Asli Celikyilmaz, Li Deng and Dilek Hakkani-Tür
4 Deep Learning in Lexical Analysis and Parsing . . . . . . . . . . . . . . . 79
Wanxiang Che and Yue Zhang
5 Deep Learning in Knowledge Graph . . . . . . . . . . . . . . . . . . . . . . 117
Zhiyuan Liu and Xianpei Han
6 Deep Learning in Machine Translation . . . . . . . . . . . . . . . . . . . . 147
Yang Liu and Jiajun Zhang
7 Deep Learning in Question Answering . . . . . . . . . . . . . . . . . . . . . 185
Kang Liu and Yansong Feng
8 Deep Learning in Sentiment Analysis . . . . . . . . . . . . . . . . . . . . . . 219
Duyu Tang and Meishan Zhang
9 Deep Learning in Social Computing . . . . . . . . . . . . . . . . . . . . . . . . 255
Xin Zhao and Chenliang Li
10 Deep Learning in Natural Language Generation from Images . . . . 289
Xiaodong He and Li Deng
11 Epilogue: Frontiers of NLP in the Deep Learning Era . . . . . . . . . . 309
Li Deng and Yang Liu
Glossary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 327

xi
Contributors

Asli Celikyilmaz Microsoft Research, Redmond, WA, USA


Wanxiang Che Harbin Institute of Technology, Harbin, China
Li Deng Citadel, Seattle & Chicago, USA
Yansong Feng Peking University, Beijing, China
Dilek Hakkani-Tür Google, Mountain View, CA, USA
Xianpei Han Institute of Software, Chinese Academy of Sciences, Beijing, China
Xiaodong He Microsoft Research, Redmond, WA, USA
Chenliang Li Wuhan University, Wuhan, China
Kang Liu Institute of Automation, Chinese Academy of Sciences, Beijing, China
Yang Liu Tsinghua University, Beijing, China
Zhiyuan Liu Tsinghua University, Beijing, China
Duyu Tang Microsoft Research Asia, Beijing, China
Gokhan Tur Google, Mountain View, CA, USA
Jiajun Zhang Institute of Automation, Chinese Academy of Sciences, Beijing,
China
Meishan Zhang Heilongjiang University, Harbin, China
Yue Zhang Singapore University of Technology and Design, Singapore
Xin Zhao Renmin University of China, Beijing, China

xiii
Acronyms

AI Artificial intelligence
AP Averaged perceptron
ASR Automatic speech recognition
ATN Augmented transition network
BiLSTM Bidirectional long short-term memory
BiRNN Bidirectional recurrent neural network
BLEU Bilingual evaluation understudy
BOW Bag-of-words
CBOW Continuous bag-of-words
CCA Canonical correlation analysis
CCG Combinatory categorial grammar
CDL Collaborative deep learning
CFG Context free grammar
CYK Cocke–Younger–Kasami
CLU Conversational language understanding
CNN Convolutional neural network
CNNSM Convolutional neural network based semantic model
cQA Community question answering
CRF Conditional random field
CTR Collaborative topic regression
CVT Compound value typed
DA Denoising autoencoder
DBN Deep belief network
DCN Deep convex net
DNN Deep neural network
DSSM Deep structured semantic model
DST Dialog state tracking
EL Entity linking
EM Expectation maximization
FSM Finite state machine

xv
xvi Acronyms

GAN Generative adversarial network


GRU Gated recurrent unit
HMM Hidden Markov model
IE Information extraction
IRQA Information retrieval-based question answering
IVR Interactive voice response
KBQA Knowledge-based question answering
KG Knowledge graph
L-BFGS Limited-memory Broyden–Fletcher–Goldfarb–Shanno
LSI Latent semantic indexing
LSTM Long short-term memory
MC Machine comprehension
MCCNN Multicolumn convolutional neural network
MDP Markov decision process
MERT Minimum error rate training
METEOR Metric for evaluation of translation with explicit ordering
MIRA Margin infused relaxed algorithm
ML Machine learning
MLE Maximum likelihood estimation
MLP Multiple layer perceptron
MMI Maximum mutual information
M-NMF Modularized nonnegative matrix factorization
MRT Minimum risk training
MST Maximum spanning tree
MT Machine translation
MV-RNN Matrix-vector recursive neural network
NER Named entity recognition
NFM Neural factorization machine
NLG Natural language generation
NMT Neural machine translation
NRE Neural relation extraction
OOV Out-of-vocabulary
PA Passive aggressive
PCA Principal component analysis
PMI Point-wise mutual information
POS Part of speech
PV Paragraph vector
QA Question answering
RAE Recursive autoencoder
RBM Restricted Boltzmann machine
RDF Resource description framework
RE Relation extraction
RecNN Recursive neural network
RL Reinforcement learning
RNN Recurrent neural network
Acronyms xvii

ROUGE Recall-oriented understudy for gisting evaluation


RUBER Referenced metric and unreferenced metric blended evaluation routine
SDS Spoken dialog system
SLU Spoken language understanding
SMT Statistical machine translation
SP Semantic parsing
SRL Semantic role labeling
SRNN Segmental recurrent neural network
STAGG Staged query graph generation
SVM Support vector machine
UAS Unlabeled attachment score
UGC User-generated content
VIME Variational information maximizing exploration
VPA Virtual personal assistant
Chapter 1
A Joint Introduction to Natural
Language Processing and to Deep
Learning

Li Deng and Yang Liu

Abstract In this chapter, we set up the fundamental framework for the book. We
first provide an introduction to the basics of natural language processing (NLP) as an
integral part of artificial intelligence. We then survey the historical development of
NLP, spanning over five decades, in terms of three waves. The first two waves arose
as rationalism and empiricism, paving ways to the current deep learning wave. The
key pillars underlying the deep learning revolution for NLP consist of (1) distributed
representations of linguistic entities via embedding, (2) semantic generalization due
to the embedding, (3) long-span deep sequence modeling of natural language, (4)
hierarchical networks effective for representing linguistic levels from low to high,
and (5) end-to-end deep learning methods to jointly solve many NLP tasks. After
the survey, several key limitations of current deep learning technology for NLP are
analyzed. This analysis leads to five research directions for future advances in NLP.

1.1 Natural Language Processing: The Basics

Natural language processing (NLP) investigates the use of computers to process or to


understand human (i.e., natural) languages for the purpose of performing useful tasks.
NLP is an interdisciplinary field that combines computational linguistics, computing
science, cognitive science, and artificial intelligence. From a scientific perspective,
NLP aims to model the cognitive mechanisms underlying the understanding and pro-
duction of human languages. From an engineering perspective, NLP is concerned
with how to develop novel practical applications to facilitate the interactions between
computers and human languages. Typical applications in NLP include speech recog-
nition, spoken language understanding, dialogue systems, lexical analysis, parsing,
machine translation, knowledge graph, information retrieval, question answering,

L. Deng (B)
Citadel, Seattle & Chicago, USA
e-mail: l.deng@ieee.org
Y. Liu
Tsinghua University, Beijing, China
e-mail: liuyang2011@tsinghua.edu.cn

© Springer Nature Singapore Pte Ltd. 2018 1


L. Deng and Y. Liu (eds.), Deep Learning in Natural
Language Processing, https://doi.org/10.1007/978-981-10-5209-5_1
2 L. Deng and Y. Liu

sentiment analysis, social computing, natural language generation, and natural lan-
guage summarization. These NLP application areas form the core content of this
book.
Natural language is a system constructed specifically to convey meaning or seman-
tics, and is by its fundamental nature a symbolic or discrete system. The surface or
observable “physical” signal of natural language is called text, always in a sym-
bolic form. The text “signal” has its counterpart—the speech signal; the latter can
be regarded as the continuous correspondence of symbolic text, both entailing the
same latent linguistic hierarchy of natural language. From NLP and signal processing
perspectives, speech can be treated as “noisy” versions of text, imposing additional
difficulties in its need of “de-noising” when performing the task of understanding the
common underlying semantics. Chapters 2 and 3 as well as current Chap. 1 of this
book cover the speech aspect of NLP in detail, while the remaining chapters start
directly from text in discussing a wide variety of text-oriented tasks that exemplify
the pervasive NLP applications enabled by machine learning techniques, notably
deep learning.
The symbolic nature of natural language is in stark contrast to the continuous
nature of language’s neural substrate in the human brain. We will defer this discussion
to Sect. 1.6 of this chapter when discussing future challenges of deep learning in NLP.
A related contrast is how the symbols of natural language are encoded in several
continuous-valued modalities, such as gesture (as in sign language), handwriting
(as an image), and, of course, speech. On the one hand, the word as a symbol is
used as a “signifier” to refer to a concept or a thing in real world as a “signified”
object, necessarily a categorical entity. On the other hand, the continuous modalities
that encode symbols of words constitute the external signals sensed by the human
perceptual system and transmitted to the brain, which in turn operates in a continuous
fashion. While of great theoretical interest, the subject of contrasting the symbolic
nature of language versus its continuous rendering and encoding goes beyond the
scope of this book.
In the next few sections, we outline and discuss, from a historical perspective, the
development of general methodology used to study NLP as a rich interdisciplinary
field. Much like several closely related sub- and super-fields such as conversational
systems, speech recognition, and artificial intelligence, the development of NLP can
be described in terms of three major waves (Deng 2017; Pereira 2017), each of which
is elaborated in a separate section next.

1.2 The First Wave: Rationalism

NLP research in its first wave lasted for a long time, dating back to 1950s. In 1950,
Alan Turing proposed the Turing test to evaluate a computer’s ability to exhibit intelli-
gent behavior indistinguishable from that of a human (Turing 1950). This test is based
on natural language conversations between a human and a computer designed to gen-
erate human-like responses. In 1954, the Georgetown-IBM experiment demonstrated
1 A Joint Introduction to Natural Language Processing and to Deep Learning 3

the first machine translation system capable of translating more than 60 Russian sen-
tences into English.
The approaches, based on the belief that knowledge of language in the human
mind is fixed in advance by generic inheritance, dominated most of NLP research
between about 1960 and late 1980s. These approaches have been called rationalist
ones (Church 2007). The dominance of rationalist approaches in NLP was mainly
due to the widespread acceptance of arguments of Noam Chomsky for an innate
language structure and his criticism of N-grams (Chomsky 1957). Postulating that
key parts of language are hardwired in the brain at birth as a part of the human
genetic inheritance, rationalist approaches endeavored to design hand-crafted rules
to incorporate knowledge and reasoning mechanisms into intelligent NLP systems.
Up until 1980s, most notably successful NLP systems, such as ELIZA for simulating
a Rogerian psychotherapist and MARGIE for structuring real-world information into
concept ontologies, were based on complex sets of handwritten rules.
This period coincided approximately with the early development of artificial
intelligence, characterized by expert knowledge engineering, where domain experts
devised computer programs according to the knowledge about the (very narrow)
application domains they have (Nilsson 1982; Winston 1993). The experts designed
these programs using symbolic logical rules based on careful representations and
engineering of such knowledge. These knowledge-based artificial intelligence sys-
tems tend to be effective in solving narrow-domain problems by examining the
“head” or most important parameters and reaching a solution about the appropriate
action to take in each specific situation. These “head” parameters are identified in
advance by human experts, leaving the “tail” parameters and cases untouched. Since
they lack learning capability, they have difficulty in generalizing the solutions to new
situations and domains. The typical approach during this period is exemplified by
the expert system, a computer system that emulates the decision-making ability of a
human expert. Such systems are designed to solve complex problems by reasoning
about knowledge (Nilsson 1982). The first expert system was created in 1970s and
then proliferated in 1980s. The main “algorithm” used was the inference rules in the
form of “if-then-else” (Jackson 1998). The main strength of these first-generation
artificial intelligence systems is its transparency and interpretability in their (limited)
capability in performing logical reasoning. Like NLP systems such as ELIZA and
MARGIE, the general expert systems in the early days used hand-crafted expert
knowledge which was often effective in narrowly defined problems, although the
reasoning could not handle uncertainty that is ubiquitous in practical applications.
In specific NLP application areas of dialogue systems and spoken language under-
standing, to be described in more detail in Chaps. 2 and 3 of this book, such ratio-
nalistic approaches were represented by the pervasive use of symbolic rules and
templates (Seneff et al. 1991). The designs were centered on grammatical and onto-
logical constructs, which, while interpretable and easy to debug and update, had
experienced severe difficulties in practical deployment. When such systems worked,
they often worked beautifully; but unfortunately this happened just not very often
and the domains were necessarily limited.
4 L. Deng and Y. Liu

Likewise, speech recognition research and system design, another long-standing


NLP and artificial intelligence challenge, during this rationalist era were based
heavily on the paradigm of expert knowledge engineering, as elegantly analyzed
in (Church and Mercer 1993). During 1970s and early 1980s, the expert system
approach to speech recognition was quite popular (Reddy 1976; Zue 1985). How-
ever, the lack of abilities to learn from data and to handle uncertainty in reasoning was
acutely recognized by researchers, leading to the second wave of speech recognition,
NLP, and artificial intelligence described next.

1.3 The Second Wave: Empiricism

The second wave of NLP was characterized by the exploitation of data corpora and
of (shallow) machine learning, statistical or otherwise, to make use of such data
(Manning and Schtze 1999). As much of the structure of and theory about natural
language were discounted or discarded in favor of data-driven methods, the main
approaches developed during this era have been called empirical or pragmatic ones
(Church and Mercer 1993; Church 2014). With the increasing availability of machine-
readable data and steady increase of computational power, empirical approaches have
dominated NLP since around 1990. One of the major NLP conferences was even
named “Empirical Methods in Natural Language Processing (EMNLP)” to reflect
most directly the strongly positive sentiment of NLP researchers during that era
toward empirical approaches.
In contrast to rationalist approaches, empirical approaches assume that the human
mind only begins with general operations for association, pattern recognition, and
generalization. Rich sensory input is required to enable the mind to learn the detailed
structure of natural language. Prevalent in linguistics between 1920 and 1960, empiri-
cism has been undergoing a resurgence since 1990. Early empirical approaches to
NLP focused on developing generative models such as the hidden Markov model
(HMM) (Baum and Petrie 1966), the IBM translation models (Brown et al. 1993),
and the head-driven parsing models (Collins 1997) to discover the regularities of
languages from large corpora. Since late 1990s, discriminative models have become
the de facto approach in a variety of NLP tasks. Representative discriminative mod-
els and methods in NLP include the maximum entropy model (Ratnaparkhi 1997),
supporting vector machines (Vapnik 1998), conditional random fields (Lafferty et al.
2001), maximum mutual information and minimum classification error (He et al.
2008), and perceptron (Collins 2002).
Again, this era of empiricism in NLP was paralleled with corresponding approaches
in artificial intelligence as well as in speech recognition and computer vision. It came
about after clear evidence that learning and perception capabilities are crucial for
complex artificial intelligence systems but missing in the expert systems popular in
the previous wave. For example, when DARPA opened its first Grand Challenge for
autonomous driving, most vehicles then relied on the knowledge-based artificial intel-
ligence paradigm. Much like speech recognition and NLP, the autonomous driving and
1 A Joint Introduction to Natural Language Processing and to Deep Learning 5

computer vision researchers immediately realized the limitation of the knowledge-


based paradigm due to the necessity for machine learning with uncertainty handling
and generalization capabilities.
The empiricism in NLP and speech recognition in this second wave was based
on data-intensive machine learning, which we now call “shallow” due to the general
lack of abstractions constructed by many-layer or “deep” representations of data
which would come in the third wave to be described in the next section. In machine
learning, researchers do not need to concern with constructing precise and exact rules
as required for the knowledge-based NLP and speech systems during the first wave.
Rather, they focus on statistical models (Bishop 2006; Murphy 2012) or simple neural
networks (Bishop 1995) as an underlying engine. They then automatically learn or
“tune” the parameters of the engine using ample training data to make them handle
uncertainty, and to attempt to generalize from one condition to another and from one
domain to another. The key algorithms and methods for machine learning include EM
(expectation-maximization), Bayesian networks, support vector machines, decision
trees, and, for neural networks, backpropagation algorithm.
Generally speaking, the machine learning based NLP, speech, and other artificial
intelligence systems perform much better than the earlier, knowledge-based counter-
parts. Successful examples include almost all artificial intelligence tasks in machine
perception—speech recognition (Jelinek 1998), face recognition (Viola and Jones
2004), visual object recognition (Fei-Fei and Perona 2005), handwriting recognition
(Plamondon and Srihari 2000), and machine translation (Och 2003).
More specifically, in a core NLP application area of machine translation, as to be
described in detail in Chap. 6 of this book as well as in (Church and Mercer 1993), the
field has switched rather abruptly around 1990 from rationalistic methods outlined in
Sect. 1.2 to empirical, largely statistical methods. The availability of sentence-level
alignments in the bilingual training data made it possible to acquire surface-level
translation knowledge not by rules but from data directly, at the expense of discarding
or discounting structured information in natural languages. The most representative
work during this wave is that empowered by various versions of IBM translation
models (Brown et al. 1993). Subsequent developments during this empiricist era of
machine translation further significantly improved the quality of translation systems
(Och and Ney 2002; Och 2003; Chiang 2007; He and Deng 2012), but not at the
level of massive deployment in real world (which would come after the next, deep
learning wave).
In the dialogue and spoken language understanding areas of NLP, this empiri-
cist era was also marked prominently by data-driven machine learning approaches.
These approaches were well suited to meet the requirement for quantitative evalua-
tion and concrete deliverables. They focused on broader but shallow, surface-level
coverage of text and domains instead of detailed analyses of highly restricted text
and domains. The training data were used not to design rules for language under-
standing and response action from the dialogue systems but to learn parameters of
(shallow) statistical or neural models automatically from data. Such learning helped
reduce the cost of hand-crafted complex dialogue manager’s design, and helped
improve robustness against speech recognition errors in the overall spoken language
6 L. Deng and Y. Liu

understanding and dialogue systems; for a review, see He and Deng (2013). More
specifically, for the dialogue policy component of dialogue systems, powerful rein-
forcement learning based on Markov decision processes had been introduced during
this era; for a review, see Young et al. (2013). And for spoken language understand-
ing, the dominant methods moved from rule- or template-based ones during the first
wave to generative models like hidden Markov models (HMMs) (Wang et al. 2011)
to discriminative models like conditional random fields (Tur and Deng 2011).
Similarly, in speech recognition, over close to 30 years from early 1980 s to around
2010, the field was dominated by the (shallow) machine learning paradigm using the
statistical generative model based on the HMM integrated with Gaussian mixture
models, along with various versions of its generalization (Baker et al. 2009a, b;
Deng and O’Shaughnessy 2003; Rabiner and Juang 1993). Among many versions of
the generalized HMMs were statistical and neural-network-based hidden dynamic
models (Deng 1998; Bridle et al. 1998; Deng and Yu 2007). The former adopted EM
and switching extended Kalman filter algorithms for learning model parameters (Ma
and Deng 2004; Lee et al. 2004), and the latter used backpropagation (Picone et al.
1999). Both of them made extensive use of multiple latent layers of representations for
the generative process of speech waveforms following the long-standing framework
of analysis-by-synthesis in human speech perception. More significantly, inverting
this “deep” generative process to its counterpart of an end-to-end discriminative
process gave rise to the first industrial success of deep learning (Deng et al. 2010,
2013; Hinton et al. 2012), which formed a driving force of the third wave of speech
recognition and NLP that will be elaborated next.

1.4 The Third Wave: Deep Learning

While the NLP systems, including speech recognition, language understanding, and
machine translation, developed during the second wave performed a lot better and
with higher robustness than those during the first wave, they were far from human-
level performance and left much to desire. With a few exceptions, the (shallow)
machine learning models for NLP often did not have the capacity sufficiently large to
absorb the large amounts of training data. Further, the learning algorithms, methods,
and infrastructures were not powerful enough. All this changed several years ago,
giving rise to the third wave of NLP, propelled by the new paradigm of deep-structured
machine learning or deep learning (Bengio 2009; Deng and Yu 2014; LeCun et al.
2015; Goodfellow et al. 2016).
In traditional machine learning, features are designed by humans and feature
engineering is a bottleneck, requiring significant human expertise. Concurrently,
the associated shallow models lack the representation power and hence the ability
to form levels of decomposable abstractions that would automatically disentangle
complex factors in shaping the observed language data. Deep learning breaks away
the above difficulties by the use of deep, layered model structure, often in the form of
neural networks, and the associated end-to-end learning algorithms. The advances in
1 A Joint Introduction to Natural Language Processing and to Deep Learning 7

deep learning are one major driving force behind the current NLP and more general
artificial intelligence inflection point and are responsible for the resurgence of neural
networks with a wide range of practical, including business, applications (Parloff
2016).
More specifically, despite the success of (shallow) discriminative models in a
number of important NLP tasks developed during the second wave, they suffered from
the difficulty of covering all regularities in languages by designing features manually
with domain expertise. Besides the incompleteness problem, such shallow models
also face the sparsity problem as features usually only occur once in the training
data, especially for highly sparse high-order features. Therefore, feature design has
become one of the major obstacles in statistical NLP before deep learning comes
to rescue. Deep learning brings hope for addressing the human feature engineering
problem, with a view called “NLP from scratch” (Collobert et al. 2011), which was
in early days of deep learning considered highly unconventional. Such deep learning
approaches exploit the powerful neural networks that contain multiple hidden layers
to solve general machine learning tasks dispensing with feature engineering. Unlike
shallow neural networks and related machine learning models, deep neural networks
are capable of learning representations from data using a cascade of multiple layers of
nonlinear processing units for feature extraction. As higher level features are derived
from lower level features, these levels form a hierarchy of concepts.
Deep learning originated from artificial neural networks, which can be viewed as
cascading models of cell types inspired by biological neural systems. With the advent
of backpropagation algorithm (Rumelhart et al. 1986), training deep neural networks
from scratch attracted intensive attention in 1990s. In these early days, without large
amounts of training data and without proper design and learning methods, during
neural network training the learning signals vanish exponentially with the number
of layers (or more rigorously the depth of credit assignment) when propagated from
layer to layer, making it difficult to tune connection weights of deep neural networks,
especially the recurrent versions. Hinton et al. (2006) initially overcame this problem
by using unsupervised pretraining to first learn generally useful feature detectors.
Then, the network is further trained by supervised learning to classify labeled data.
As a result, it is possible to learn the distribution of a high-level representation using
low-level representations. This seminal work marks the revival of neural networks. A
variety of network architectures have since been proposed and developed, including
deep belief networks (Hinton et al. 2006), stacked auto-encoders (Vincent et al. 2010),
deep Boltzmann machines (Hinton and Salakhutdinov 2012), deep convolutional
neural works (Krizhevsky et al. 2012), deep stacking networks (Deng et al. 2012),
and deep Q-networks (Mnih et al. 2015). Capable of discovering intricate structures
in high-dimensional data, deep learning has since 2010 been successfully applied to
real-world tasks in artificial intelligence including notably speech recognition (Yu
et al. 2010; Hinton et al. 2012), image classification (Krizhevsky et al. 2012; He et al.
2016), and NLP (all chapters in this book). Detailed analyses and reviews of deep
learning have been provided in a set of tutorial survey articles (Deng 2014; LeCun
et al. 2015; Juang 2016).
8 L. Deng and Y. Liu

As speech recognition is one of core tasks in NLP, we briefly discuss it here due to
its importance as the first industrial NLP application in real world impacted strongly
by deep learning. Industrial applications of deep learning to large-scale speech recog-
nition started to take off around 2010. The endeavor was initiated with a collaboration
between academia and industry, with the original work presented at the 2009 NIPS
Workshop on Deep Learning for Speech Recognition and Related Applications. The
workshop was motivated by the limitations of deep generative models of speech, and
the possibility that the big-compute, big-data era warrants a serious exploration of
deep neural networks. It was believed then that pretraining DNNs using generative
models of deep belief nets based on the contrastive divergence learning algorithm
would overcome the main difficulties of neural nets encountered in the 1990s (Dahl
et al. 2011; Mohamed et al. 2009). However, early into this research at Microsoft, it
was discovered that without contrastive divergence pretraining, but with the use of
large amounts of training data together with the deep neural networks designed with
corresponding large, context-dependent output layers and with careful engineering,
dramatically lower recognition errors could be obtained than then-state-of-the-art
(shallow) machine learning systems (Yu et al. 2010, 2011; Dahl et al. 2012). This
finding was quickly verified by several other major speech recognition research
groups in North America (Hinton et al. 2012; Deng et al. 2013) and subsequently
overseas. Further, the nature of recognition errors produced by the two types of sys-
tems was found to be characteristically different, offering technical insights into how
to integrate deep learning into the existing highly efficient, run-time speech decod-
ing system deployed by major players in speech recognition industry (Yu and Deng
2015; Abdel-Hamid et al. 2014; Xiong et al. 2016; Saon et al. 2017). Nowadays,
backpropagation algorithm applied to deep neural nets of various forms is uniformly
used in all current state-of-the-art speech recognition systems (Yu and Deng 2015;
Amodei et al. 2016; Saon et al. 2017), and all major commercial speech recogni-
tion systems—Microsoft Cortana, Xbox, Skype Translator, Amazon Alexa, Google
Assistant, Apple Siri, Baidu and iFlyTek voice search, and more—are all based on
deep learning methods.
The striking success of speech recognition in 2010–2011 heralded the arrival of
the third wave of NLP and artificial intelligence. Quickly following the success of
deep learning in speech recognition, computer vision (Krizhevsky et al. 2012) and
machine translation (Bahdanau et al. 2015) were taken over by the similar deep
learning paradigm. In particular, while the powerful technique of neural embedding
of words was developed in as early as 2011 (Bengio et al. 2001), it is not until more
than 10 year later it was shown to be practically useful at a large and practically useful
scale (Mikolov et al. 2013) due to the availability of big data and faster computation.
In addition, a large number of other real-world NLP applications, such as image
captioning (Karpathy and Fei-Fei 2015; Fang et al. 2015; Gan et al. 2017), visual
question answering (Fei-Fei and Perona 2016), speech understanding (Mesnil et al.
2013), web search (Huang et al. 2013b), and recommendation systems, have been
made successful due to deep learning, in addition to many non-NLP tasks including
drug discovery and toxicology, customer relationship management, recommendation
systems, gesture recognition, medical informatics, advertisement, medical image
1 A Joint Introduction to Natural Language Processing and to Deep Learning 9

analysis, robotics, self-driving vehicles, board and eSports games (e.g., Atari, Go,
Poker, and the latest, DOTA2), and so on. For more details, see https://en.wikipedia.
org/wiki/deep_learning.
In more specific text-based NLP application areas, machine translation is perhaps
impacted the most by deep learning. Advancing from the shallow statistical machine
translation developed during the second wave of NLP, the current best machine
translation systems in real-world applications are based on deep neural networks. For
example, Google announced the first stage of its move to neural machine translation
in September 2016 and Microsoft made a similar announcement 2 months later.
Facebook has been working on the conversion to neural machine translation for
about a year, and by August 2017 it is at full deployment. Details of the deep learning
techniques in these state-of-the-art large-scale machine translation systems will be
reviewed in Chap. 6.
In the area of spoken language understanding and dialogue systems, deep learning
is also making a huge impact. The current popular techniques maintain and expand
the statistical methods developed during second-wave era in several ways. Like the
empirical, (shallow) machine learning methods, deep learning is also based on data-
intensive methods to reduce the cost of hand-crafted complex understanding and
dialogue management, to be robust against speech recognition errors under noise
environments and against language understanding errors, and to exploit the power
of Markov decision processes and reinforcement learning for designing dialogue
policy, e.g., (Gasic et al. 2017; Dhingra et al. 2017). Compared with the earlier
methods, deep neural network models and representations are much more powerful
and they make end-to-end learning possible. However, deep learning has not yet
solved the problems of interpretability and domain scalability associated with earlier
empirical techniques. Details of the deep learning techniques popular for current
spoken language understanding and dialogue systems as well as their challenges
will be reviewed in Chaps. 2 and 3.
Two important recent technological breakthroughs brought about in applying deep
learning to NLP problems are sequence-to-sequence learning (Sutskevar et al. 2014)
and attention modeling (Bahdanau et al. 2015). The sequence-to-sequence learning
introduces a powerful idea of using recurrent nets to carry out both encoding and
decoding in an end-to-end manner. While attention modeling was initially developed
to overcome the difficulty of encoding a long sequence, subsequent developments
significantly extended its power to provide highly flexible alignment of two arbitrary
sequences that can be learned together with neural network parameters. The key
concepts of sequence-to-sequence learning and of attention mechanism boosted the
performance of neural machine translation based on distributed word embedding over
the best system based on statistical learning and local representations of words and
phrases. Soon after this success, these concepts have also been applied successfully
to a number of other NLP-related tasks such as image captioning (Karpathy and
Fei-Fei 2015; Devlin et al. 2015), speech recognition (Chorowski et al. 2015), meta-
learning for program execution, one-shot learning, syntactic parsing, lip reading, text
understanding, summarization, and question answering and more.
10 L. Deng and Y. Liu

Setting aside their huge empirical successes, models of neural-network-based


deep learning are often simpler and easier to design than the traditional machine
learning models developed in the earlier wave. In many applications, deep learning
is performed simultaneously for all parts of the model, from feature extraction all
the way to prediction, in an end-to-end manner. Another factor contributing to the
simplicity of neural network models is that the same model building blocks (i.e., the
different types of layers) are generally used in many different applications. Using
the same building blocks for a large variety of tasks makes the adaptation of models
used for one task or data to another task or data relatively easy. In addition, software
toolkits have been developed to allow faster and more efficient implementation of
these models. For these reasons, deep neural networks are nowadays a prominent
method of choice for a large variety of machine learning and artificial intelligence
tasks over large datasets including, prominently, NLP tasks.
Although deep learning has proven effective in reshaping the processing of speech,
images, and videos in a revolutionary way, the effectiveness is less clear-cut in inter-
secting deep learning with text-based NLP despite its empirical successes in a number
of practical NLP tasks. In speech, image, and video processing, deep learning effec-
tively addresses the semantic gap problem by learning high-level concepts from raw
perceptual data in a direct manner. However, in NLP, stronger theories and structured
models on morphology, syntax, and semantics have been advanced to distill the under-
lying mechanisms of understanding and generation of natural languages, which have
not been as easily compatible with neural networks. Compared with speech, image,
and video signals, it seems less straightforward to see that the neural representations
learned from textual data can provide equally direct insights onto natural language.
Therefore, applying neural networks, especially those having sophisticated hierar-
chical architectures, to NLP has received increasing attention and has become the
most active area in both NLP and deep learning communities with highly visible
progresses made in recent years (Deng 2016; Manning and Socher 2017). Surveying
the advances and analyzing the future directions in deep learning for NLP form the
main motivation for us to write this chapter and to create this book, with the desire
for the NLP researchers to accelerate the research further in the current fast pace of
the progress.

1.5 Transitions from Now to the Future

Before analyzing the future dictions of NLP with more advanced deep learning, here
we first summarize the significance of the transition from the past waves of NLP to
the present one. We then discuss some clear limitations and challenges of the present
deep learning technology for NLP, to pave a way to examining further development
that would overcome these limitations for the next wave of innovations.
1 A Joint Introduction to Natural Language Processing and to Deep Learning 11

1.5.1 From Empiricism to Deep Learning: A Revolution

On the surface, the deep learning rising wave discussed in Sect. 1.4 in this chapter
appears to be a simple push of the second, empiricist wave of NLP (Sect. 1.3) into
an extreme end with bigger data, larger models, and greater computing power. After
all, the fundamental approaches developed during both waves are data-driven and
are based on machine learning and computation, and have dispensed with human-
centric “rationalistic” rules that are often brittle and costly to acquire in practical
NLP applications. However, if we analyze these approaches holistically and at a
deeper level, we can identify aspects of conceptual revolution moving from empiricist
machine learning to deep learning, and can subsequently analyze the future directions
of the field (Sect. 1.6). This revolution, in our opinion, is no less significant than the
revolution from the earlier rationalist wave to empiricist one as analyzed at the
beginning (Church and Mercer 1993) and at the end of the empiricist era (Charniak
2011).
Empiricist machine learning and linguistic data analysis during the second NLP
wave started in early 1990 s by crypto-analysts and computer scientists working
on natural language sources that are highly limited in vocabulary and application
domains. As we discussed in Sect. 1.3, surface-level text observations, i.e., words
and their sequences, are counted using discrete probabilistic models without relying
on deep structure in natural language. The basic representations were “one-hot” or
localist, where no semantic similarity between words was exploited. With restric-
tions in domains and associated text content, such structure-free representations and
empirical models are often sufficient to cover much of what needs to be covered.
That is, the shallow, count-based statistical models can naturally do well in limited
and specific NLP tasks. But when the domain and content restrictions are lifted for
more realistic NLP applications in real-world, count-based models would necessarily
become ineffective, no manner how many tricks of smoothing have been invented
in an attempt to mitigate the problem of combinatorial counting sparseness. This
is where deep learning for NLP truly shines—distributed representations of words
via embedding, semantic generalization due to the embedding, longer span deep
sequence modeling, and end-to-end learning methods have all contributed to beat-
ing empiricist, count-based methods in a wide range of NLP tasks as discussed in
Sect. 1.4.

1.5.2 Limitations of Current Deep Learning Technology

Despite the spectacular successes of deep learning in NLP tasks, most notably in
speech recognition/understanding, language modeling, and in machine translation,
there remain huge challenges. The current deep learning methods based on neu-
ral networks as a black box generally lack interpretability, even further away from
explainability, in contrast to the “rationalist” paradigm established during the first
12 L. Deng and Y. Liu

NLP wave where the rules devised by experts were naturally explainable. In practice,
however, it is highly desirable to explain the predictions from a seemingly “black-
box” model, not only for improving the model but for providing the users of the
prediction system with interpretations of the suggested actions to take (Koh and
Liang 2017).
In a number of applications, deep learning methods have proved to give recog-
nition accuracy close to or exceeding humans, but they require considerably more
training data, power consumption, and computing resources than humans. Also,
the accuracy results are statistically impressive but often unreliable on the individ-
ual basis. Further, most of the current deep learning models have no reasoning and
explaining capabilities, making them vulnerable to disastrous failures or attacks with-
out the ability to foresee and thus to prevent them. Moreover, the current NLP models
have not taken into account the need for developing and executing goals and plans
for decision-making via ultimate NLP systems. A more specific limitation of current
NLP methods based on deep learning is their poor abilities for understanding and
reasoning inter-sentential relationships, although huge progresses have been made
for interwords and phrases within sentences.
As discussed earlier, the success of deep learning in NLP has largely come from a
simple strategy thus far—given an NLP task, apply standard sequence models based
on (bidirectional) LSTMs, add attention mechanisms if information required in the
task needs to flow from another source, and then train the full models in an end-to-
end manner. However, while sequence modeling is naturally appropriate for speech,
human understanding of natural language (in text form) requires more complex
structure than sequence. That is, current sequence-based deep learning systems for
NLP can be further advanced by exploiting modularity, structured memories, and
recursive, tree-like representations for sentences and larger text (Manning 2016).
To overcome the challenges outlined above and to achieve the ultimate success
of NLP as a core artificial intelligence field, both fundamental and applied research
are needed. The next new wave of NLP and artificial intelligence will not come until
researchers create new paradigmatic, algorithmic, and computation (including hard-
ware) breakthroughs. Here, we outline several high-level directions toward potential
breakthroughs.

1.6 Future Directions of NLP

1.6.1 Neural-Symbolic Integration

A potential breakthrough is in developing advanced deep learning models and meth-


ods that are more effective than current methods in building, accessing, and exploit-
ing memories and knowledge, including, in particular, common-sense knowledge.
It is not clear how to best integrate the current deep learning methods, centered
on distributed representations (of everything), with explicit, easily interpretable, and
1 A Joint Introduction to Natural Language Processing and to Deep Learning 13

localist-represented knowledge about natural language and the world and with related
reasoning mechanisms.
One path to this goal is to seamlessly combine neural networks and symbolic
language systems. These NLP and artificial intelligence systems will aim to discover
by themselves the underlying causes or logical rules that shape their prediction and
decision-making processes interpretable to human users in symbolic natural language
forms. Recently, very preliminary work in this direction made use of an integrated
neural-symbolic representation called tensor-product neural memory cells, capable
of decoding back to symbolic forms. This structured neural representation is provably
lossless in the coded information after extensive learning within the neural-tensor
domain (Palangi et al. 2017; Smolensky et al. 2016; Lee et al. 2016). Extensions
of such tensor-product representations, when applied to NLP tasks such as machine
reading and question answering, are aimed to learn to process and understand mas-
sive natural language documents. After learning, the systems will be able not only to
answer questions sensibly but also to truly understand what it reads to the extent that
it can convey such understanding to human users in providing clues as to what steps
have been taken to reach the answer. These steps may be in the form of logical reason-
ing expressed in natural language which is thus naturally understood by the human
users of this type of machine reading and comprehension systems. In our view, natu-
ral language understanding is not just to accurately predict an answer from a question
with relevant passages or data graphs as its contextual knowledge in a supervised
way after seeing many examples of matched questions–passages–answers. Rather,
the desired NLP system equipped with real understanding should resemble human
cognitive capabilities. As an example of such capabilities (Nguyen et al. 2017)—
after an understanding system is trained well, say, in a question answering task
(using supervised learning or otherwise), it should master all essential aspects of the
observed text material provided to solve the question answering tasks. What such
mastering entails is that the learned system can subsequently perform well on other
NLP tasks, e.g., translation, summarization, recommendation, etc., without seeing
additional paired data such as raw text data with its summary, or parallel English and
Chinese texts, etc.
One way to examine the nature of such powerful neural-symbolic systems is
to regard them as ones incorporating the strength of the “rationalist” approaches
marked by expert reasoning and structure richness popular during the first wave of
NLP discussed in Sect. 1.2. Interestingly, prior to the rising of deep learning (third)
wave of NLP, (Church 2007) argued that the pendulum from rationalist to empiri-
cist approaches has swung too far at almost the peak of the second NLP wave, and
predicted that the new rationalist wave would arrive. However, rather than swinging
back to a renewed rationalist era of NLP, deep learning era arrived in full force in just
a short period from the time of writing by Church (2007). Instead of adding the ratio-
nalist flavor, deep learning has been pushing empiricism of NLP to its pinnacle with
big data and big compute, and with conceptually revolutionary ways of representing
a sweeping range of linguistic entities by massive parallelism and distributedness,
thus drastically enhancing the generalization capability of new-generation NLP mod-
els. Only after the sweeping successes of current deep learning methods for NLP
14 L. Deng and Y. Liu

(Sect. 1.4) and subsequent analyses of a series of their limitations, do researchers


look into the next wave of NLP—not swinging back to rationalism while abandon-
ing empiricism but developing more advanced deep learning paradigms that would
organically integrate the missing essence of rationalism into the structured neural
methods that are aimed to approach human cognitive functions for language.

1.6.2 Structure, Memory, and Knowledge

As discussed earlier in this chapter as well as in the current NLP literature (Man-
ning and Socher 2017), NLP researchers at present still have very primitive deep
learning methods for exploiting structure and for building and accessing memories
or knowledge. While LSTM (with attention) has been pervasively applied to NLP
tasks to beat many NLP benchmarks, LSTM is far from a good memory model
for human cognition. In particular, LSTM lacks adequate structure for simulating
episodic memory, and one key component of human cognitive ability is to retrieve
and re-experience aspects of a past novel event or thought. This ability gives rise
to one-shot learning skills and can be crucial in reading comprehension of natural
language text or speech understanding, as well as reasoning over events described by
natural language. Many recent studies have been devoted to better memory model-
ing, including external memory architectures with supervised learning (Vinyals et al.
2016; Kaiser et al. 2017) and augmented memory architectures with reinforcement
learning (Graves et al. 2016; Oh et al. 2016). However, they have not shown general
effectiveness, but have suffered from a number of of limitations including notably
scalability (arising from the use of attention which has to access every stored element
in the memory). Much work remains in the direction of better modeling of memory
and exploitation of knowledge for text understanding and reasoning.

1.6.3 Unsupervised and Generative Deep Learning

Another potential breakthrough in deep learning for NLP is in new algorithms for
unsupervised deep learning, which makes use of ideally no direct teaching signals
paired with inputs (token by token) to guide the learning. Word embedding discussed
in Sect. 1.4 can be viewed as a weak form of unsupervised learning, making use of
adjacent words as “cost-free” surrogate teaching signals, but for real-world NLP pre-
diction tasks, such as translation, understanding, summarization, etc., such embed-
ding obtained in an “unsupervised manner” has to be fed into another supervised
architecture which requires costly teaching signals. In truly unsupervised learning
which requires no expensive teaching signals, new types of objective functions and
new optimization algorithms are needed, e.g., the objective function for unsupervised
learning should not require explicit target label data aligned with the input data as
in cross entropy that is most popular for supervised learning. Development of unsu-
1 A Joint Introduction to Natural Language Processing and to Deep Learning 15

pervised deep learning algorithms has been significantly behind that of supervised
and reinforcement deep learning where backpropagation and Q-learning algorithms
have been reasonably mature.
The most recent preliminary development in unsupervised learning takes the
approach of exploiting sequential output structure and advanced optimization meth-
ods to alleviate the need for using labels in training prediction systems (Russell and
Stefano 2017; Liu et al. 2017). Future advances in unsupervised learning are promis-
ing by exploiting new sources of learning signals including the structure of input data
and the mapping relationships from input to output and vice versa. Exploiting the rela-
tionship from output to input is closely connected to building conditional generative
models. To this end, the recent popular topic in deep learning—generative adversar-
ial networks (Goodfellow et al. 2014)—is a highly promising direction where the
long-standing concept of analysis-by-synthesis in pattern recognition and machine
learning is likely to return to spotlight in the near future in solving NLP tasks in new
ways.
Generative adversarial networks have been formulated as neural nets, with dense
connectivity among nodes and with no probabilistic setting. On the other hand,
probabilistic and Bayesian reasoning, which often takes computational advantage
of sparse connections among “nodes” as random variables, has been one of the
principal theoretical pillars to machine learning and has been responsible for many
NLP methods developed during the empiricist wave of NLP discussed in Sect. 1.3.
What is the right interface between deep learning and probabilistic modeling? Can
probabilistic thinking help understand deep learning techniques better and motivate
new deep learning methods for NLP tasks? How about the other way around? These
issues are widely open for future research.

1.6.4 Multimodal and Multitask Deep Learning

Multimodal and multitask deep learning are related learning paradigms, both con-
cerning the exploitation of latent representations in the deep networks pooled from
different modalities (e.g., audio, speech, video, images, text, source codes, etc.) or
from multiple cross-domain tasks (e.g., point and structured prediction, ranking,
recommendation, time-series forecasting, clustering, etc.). Before the deep learning
wave, multimodal and multitask learning had been very difficult to be made effective,
due to the lack of intermediate representations that share across modalities or tasks.
See a most striking example of this contrast for multitask learning—multilingual
speech recognition during the empiricist wave (Lin et al. 2008) and during the deep
learning wave (Huang et al. 2013a).
Multimodal information can be exploited as low-cost supervision. For instance,
standard speech recognition, image recognition, and text classification methods make
use of supervision labels within each of the speech, image, and text modalities sepa-
rately. This, however, is far from how children learn to recognize speech, image, and
to classify text. For example, children often get the distant “supervision” signal for
Another random document with
no related content on Scribd:
to see some people at the Duc d’Harcourt’s, if well enough.
He will not be going away before Wednesday. His address is
55, Welbeck Street. I think you would do well to send your
letter to him by hand, sealed and addressed ‘À Monsieur
Seul,’ enclosed in an outer envelope with his ordinary
address: ‘Son Altesse Royale Monsieur, frère du Roi.’ Send
me a line to tell me what you have done. Adieu.”
It was not till after a long delay that Lady Atkyns at last succeeded
in meeting the Prince at an inn, only to meet with a point-blank
refusal. But she was not to be discouraged. The very next day she
wrote again to the Comte asking for an audience. This time it was
another member of his suite, the Bishop of Saint-Pol de Léon, who
replied to her communication—
“The moment I saw M[onsieur] yesterday, my dear lady, he
told me about your letter, which gave him great pleasure,
though it is a matter for great regret to him that he is quite
unable to do as you wish, and as he himself would wish.
Since his recent attack he has been unable to dress or go out;
he has not been able to receive any ladies, anxious though he
is to welcome those who are here and who were attached to
the Princess. He could not receive one without its being
known, and then he would be expected to receive a number
of others. You know how things get about and what a close
watch is kept on Princes, and how careful our Prince must be
to do nothing that would lay him open to criticism or even to
suspicion. If his stay here were prolonged, and he found he
could see other ladies also, the thing might be managed; but
there would be difficulties even then, in view of your secret
being perhaps of a compromising nature. I am but expressing
to you the Prince’s own views. I hope to see you to-morrow
between midday and three o’clock.”
The great of this world are never at a loss for pretexts for refusing
requests. Monsieur was particularly anxious to evade an interview
which he felt to be undesirable, and therefore confined himself to
sending her these amiable phrases.
About the same time M. de Thauvenay, one of the King’s most
devoted courtiers, who happened to be in London, seems to have
promised to use his good offices with his master on her behalf,
telling him of her record and perhaps of her hopes.
Having exhausted all the means at her disposal in England, Lady
Atkyns saw that she must manage her journey as best she could
from her own resources, and resolved to make yet another sacrifice
to this end. She had obtained a considerable loan already once upon
a mortgage on her beautiful estate of Ketteringham. As this was her
only source of revenue, there was no alternative to raising a further
mortgage on it, and this she managed, though with greater difficulty
than before. (The property was not, in fact, her own at this time,
being entailed on her son Edward.)
She seems to have raised in all about £3000 in this way in 1799
and the three following years.
Some weeks before the “18th Brumaire” and Bonaparte’s coup
d’état, she set out for the Continent. What exactly was her purpose?
What use was she going to make of her money? It is impossible to
say. To clothe her errand in the greater mystery, she decided to land
in France under an assumed name, and to veil her personality under
the designation of the “Little Sailor” (le petit matelot).
“I feel I must again send my good wishes for a pleasant
journey to the charming ‘Little Sailor,’” some unidentified
friend writes to her on September 7, “and I cannot too often
beg him to bear in mind that he leaves behind him in England
friends who take a deep interest in his welfare, and who will
learn with pleasure that he has arrived safely at his
destination, and, above all, that after fulfilling his mission he
has escaped all the unpleasantness and dangers to which his
truly admirable devotion and zeal will expose him. I hope one
day to prove to the ‘Little Sailor’ how he has long filled me
with the most genuine sentiments—sentiments which I have
refrained from expressing for reasons of which the ‘Little
Sailor’ will approve. I cannot say too often to the amiable
‘Little Sailor’ what pleasure I shall have in repeating to him in
France—and in France preferably to elsewhere—the
assurance of eternal and tender attachment that I have vowed
him for ever and ever.”
It is difficult to know what the “Petit Matelot” did on arriving in
Paris. It was a moment of crisis, for the Consulate was being
established. Most of those who had been mixed up in the Temple
affair were inaccessible, and yet it was important to get into touch
with them if anything was to be ascertained about the Dauphin. It
would not have done, however, to provoke suspicion, or Fouché
would have been on her track.
Certain only it is that for several months she seems to have
disappeared from sight. At last she was run to earth and hunted out
by Fouché’s agents, and was obliged to make away to the Loire,
where she had devoted friends.
The Verrière family lived in the country six miles from Saumur, in
Anjou, where many nobles, fleeing from the storm, had found a safe
refuge. The vicinity of the forest of Fontevrault enabled them to gain
the Vendée, and thus escape the fury of the Revolutionists. Mme.
Verrière had met Lady Atkyns in Paris years before, perhaps during
the golden days of Versailles. Recalling their former friendliness,
Lady Atkyns went to them in her trouble. The welcome they
extended to her justified her hopes, and she dwelt with them for
some time, until the police had lost all trace of her.
About this period, vague reports began to be spread about with
reference to the imprisonment of the children of Louis XVI. in the
Temple. The obscurity which had cloaked the last hours of the
Dauphin was still keeping certain brains at work. And a book which
was published in 1800 helped to reawaken public curiosity. In Le
Cimetière de la Madeleine, a romance written by an author until then
little known, Regnault-Warin deliberately questioned the alleged
death of the Dauphin, and, in fact, based a story of adventure upon
the supposition of his being still alive. Written in the fashion of the
time, full of surprising episodes, and bristling with more or less
untrustworthy anecdotes touching on the captivity of the Royal
Family in the Temple, this novel had an immense success. If it came
before Lady Atkyns it must have served to stimulate her anxiety to
solve the problem she had so much at heart.
In the summer of 1801 Lady Atkyns appears to have addressed
herself to Louis XVIII., unwarned by her failure with Monsieur. In this
case also failure was to be her portion.
“Your letter,” ran the reply, signed by M. de Thauvenay
(whom she had met some years before), and addressed, as a
precaution, to Monsieur James Brown, dated October 2,
“would have been enigmatic to me had I not placed it before
my master, who, by a curious series of accidents, had
received only a few days before the communication you sent
him on the 12th of July. In requesting me to reply to you,
monsieur, he charges me to express to you his recognition of
your constant interest and indefatigable zeal for his welfare,
and his regret that he is prevented by his present position
from learning the particulars of the speculation that your heart
has formed, and that he cannot have any share in it.”
Six weeks later Lady Atkyns received a second letter, despatched,
like the first, from Varsovie, reinforcing the above:—
“I wish,” writes M. de Thauvenay, “that I could convey to
you the deep and tender feeling with which my dear and
venerated master has read these new and touching
testimonies of your interest and friendship, and his deep
regret at being unable to enjoy the consolations that your
sympathetic and generous nature has proffered him! No,
monsieur, I swear to you, no other house has offered him any
kind of interest in the speculation you have proposed to him. I
should add that there is no one with whom he would rather
have shared the chances than with you; but his position is
such that, for the moment at least, he can only display
passive courage in the face of misfortune. I need not remind
you, monsieur, that the most appreciative and most generous
of hearts has eternal claim upon a heart such as yours.
Never, I feel convinced, will your noble and moving sentiment
be modified by time or place. This conviction is sweet to me,
and it is with the utmost sincerity that I render you once again
my tender (if I may use the word) and admiring respect.”
It is not easy at first to understand what M. de Thauvenay means
by this “speculation,” in which the King refuses to take part. On
reflection it seems probable that Lady Atkyns’s proposal, thus
described, had reference to the affair of the Temple, for it seems
impossible that she should have flattered herself that she could see
a way to the return of the exiled King.
However that may be, these two letters convinced her that it would
be useless to prolong her stay in France, and she returned to
Ketteringham, after an absence of three years, without having
effected her purpose.

Two tragic events occurred in the year 1804 to startle the French
who were still taking refuge in England. The first was the arrest and
shooting of the Duc d’Enghien at Vincennes. If Bonaparte had
punished one of the many schemers who had plotted against him on
English soil, his action would have found defenders. But this
execution of a Prince, who was absolutely innocent and who had
held apart from all political intrigues, aroused the same kind of horror
that had been evoked eleven years earlier by the death of Louis XVI.
and Marie-Antoinette.
The Prince de Condé, the Duc d’Enghien’s grandfather, was
staying in England at this time, like Monsieur, the King’s brother, and
their residences were naturally the centre of the excitement over this
event. The Baron de Suzannet describes the state of things in their
entourage in a letter to Lady Atkyns:—
“It would seem, madame,” he writes to her on April 14, “that
the murder of the Duc d’Enghien has horrified not merely all
true Frenchmen, as was to be expected, but also Englishmen
of every class, the perfidy as well as the cruelty of it is so
revolting to all in whom the sentiment of justice and honour is
not extinct. I shall not speak of the courageous and heroic
death of this ill-fated Prince, but of the condition of his
unhappy relatives. Since the day when Monsieur carried him
the terrible news, the Prince de Condé (save for two journeys
to London necessitated by his anxiety) has not left his room or
been down to dinner. Plunged in grief, he sees no one, and it
is much feared that his death may follow that of the Duc
d’Enghien. He loved the Duc as his grandson and his pupil,
and perhaps even more as one qualified by Providence to
add still further to the glory of his illustrious name. The sorrow
of the Duc de Bourbon is not less deep and intense.”
At the same moment, the news of the arrest of Cadoudal in Paris,
the discovery of his plot, the sensational trial of his twelve
accomplices, together with a number of insurgents—forty-seven
prisoners in all—and finally the execution of the famous brigand on
the morning of June 25, came to intensify the agitation of the French
in England.
Of these events Lady Atkyns heard particulars from the Comte de
Frotté, father of her friend the General. Throughout five years the
venerable Comte had followed with joy or anguish the career of his
son as a leader of the insurrection in Normandy. Repeatedly he had
come to his aid with money and encouragement. Suddenly the fatal
bullet had ended everything. Henceforth the unhappy father had
followed eagerly everything that could bring back the memory of the
Chevalier, and he had been drawn to Lady Atkyns by his knowledge
of the long-standing friendship that had existed between her and
him.
To add to his sorrow, Charles de Frotté, half-brother to Louis, had
been arrested and imprisoned by Napoleon’s police soon after the
execution at Verneuil. He was kept at the Temple for two years,
without apparent reason, then sent to the Fort of Toux, in Jura.
“I learnt yesterday,” writes the Comte de Frotté to Lady
Atkyns, “that my unhappy son has been transferred from the
Temple to a château in Franche Comté. How cruel is this
persecution! How terrible this imprisonment, not only for
himself, but for us who are bound to him by ties of kinship and
for his friends.”
Thus Lady Atkyns, though in the seclusion of the country, far from
London and the Continent, remained bound in thought to her life of
earlier days. She had no one now to love except her son, who was
an officer in the first regiment of Royal Dragoons. Owing to the
delicacy of his health, young Edward Atkyns had been obliged to go
on leave for a time, and his mother invited the son of Baron de
Suzannet to Ketteringham to keep him company. But the visit did not
come off, and two months later the young soldier died of the malady
from which he had been suffering for some years.
Two years earlier a somewhat strange incident had occurred in
France. On February 17, 1802, the police court at Vitry-le-Françoi
had to sit in judgment upon a young man named Jean-Marie
Hervagault, charged with swindling, passing under a false name, and
vagabondage. This individual, arrested and imprisoned for the first
time in 1799, claimed to be the Dauphin, escaped miraculously from
the Temple. The son of a tailor of Saint-Lo, Hervagault, in the course
of his wanderings, had managed to convince a certain number of
people that he really was the Prince. Public curiosity was aroused.
Many people went to visit the youth in prison. To put a stop to this
movement, the Vitry Tribunal condemned the adventurer to four
years’ imprisonment. His trial disclosed the fact that amongst his
dupes were many persons of distinction, including M. Lafont de
Savines. Some weeks later the Vitry sentence was ratified at Rouen,
and Hervagault was incarcerated in the prison of Bicêtre in Paris. But
the feelings of sympathy and pity that had been called forth,
Hervagault’s assertions and his circumstantial accounts of the way in
which he had been carried off from the Temple—all these things
attracted the more attention by reason of the appearance a short
time before of Beauchamp’s work, Le faux Dauphin actuellement en
France.
Lady Atkyns was quick to secure details as to the story of the
prisoner at Bicêtre. There were many contradictions in it that must
have come home to her. And Hervagault mentions the name of the
General Louis de Frotté as that of one of his liberators, whereas, in
his letters to her, the Chevalier had made it quite clear that this could
not be so. However, it seemed worth her while to write to the old
Comte de Frotté on the subject.
“I have just received your letter,” he replies, August 16,
1804, “and I hasten to send you a line. I have spent a whole
week rummaging among papers. I can assure you that what
is stated in the book you have sent me is all fiction. Louis and
Duchale are mentioned in 1802 because they were both
dead. I am almost certain that in 1795 (in the month in
question) Louis was fighting in Normandy, and that he did not
leave his companions once all that year. But we shall go into
all this on your return, and no doubt will be left in your mind. If
you arrive towards the end of the month, tell me at once. I
shall call on you and tell you all I can, and make you see why
I am convinced that this fellow is a puppet in some one else’s
hands.”
Lady Atkyns was reluctant to give up the faint hope that there
might be something in this Hervagault narrative, but after some
conversations with the Comte de Frotté, and after comparing the
pretender’s statement with documents left by the Chevalier, she was
at last convinced that the whole thing was a fraud.
We hear of her again in October, 1809, taking a prominent part in
the celebrations being held in her neighbourhood in honour of the
jubilee of George III. Then we lose sight of her until 1814, and the
triumphant return to Paris of Louis XVIII. Lady Atkyns hastens now to
secure the good offices of the Duc de Bourbon, with a view to
drawing attention to all her sacrifices and the sums of money she
has expended.
She is delayed, however, over her contemplated journey to France
for this purpose, and Napoleon’s escape postpones for two years
more all hope of accomplishing her return to Paris.
When at last the monarchy is restored once more, she finds that
her aspirations are destined to be disappointed, despite all the kind
words with which she was soothed in England, and we find her
uttering the word “ingratitude,” which is henceforth to be so often on
her tongue. There were so many who held themselves entitled to
gratitude and recognition at the hands of Louis XVIII.—émigrés
returned to France after twenty years of sorrow and indignities, and
now counting upon the recovery of their possessions or on being
reimbursed in some way by the act of the Sovereign. What an
awakening they met with when the time came to formulate their
applications and they found themselves obliged to condescend to
the drafting of innumerable documents, and to put up with
interminable delays!
On September 27, 1816, Lady Atkyns writes to her friend Mme. de
Verrière an account of her disappointing experiences. She had been
well received at Court, but that was all.
“The kind of ingratitude I have been meeting with is not very
consoling. They give me plenty of kind words, but nothing
more. I have written a long letter to the man of business,
begging of him to get the employer to reimburse me a little for
the moment, but I have received nothing yet, and this puts me
out greatly. Perhaps something will turn up between now and
the end of the week. If not I must go and see my poor mother
and beg to get my affairs into order.”
The state of her affairs, for long precarious, was now giving the
poor lady very serious anxiety. By recourse to various expedients
she had managed to hold out until the return of the Bourbons, and to
stave off her creditors. But she was now at her last gasp. If the King
refused to help her, to “reimburse” her, she was ruined. The Comte
de la Châtre had assured her that her application was under
favourable consideration, that the King regarded it approvingly, that
the Comte de Pradel, the head of the King’s household, had it in
hand; but, in spite of this, there was a series of delays.
At last, worn out with waiting, she writes in her naïve style to the
Comte on October 10, 1816—
“I beg of you to be good enough to get the King to decide
this matter as soon as ever possible. I must get away to
England in three weeks to see my mother, who is ill, and I
can’t possibly do this until I know the King’s decision in regard
to me. I know his Majesty is too good to injure one who has
given so many proofs of boundless devotion to the Royalist
cause and to the entire Royal Family. Although I have a
splendid estate in England, I am now in great difficulties by
reason of this devotion. I tell you all this, Monsieur le Comte,
so that, like the good Frenchman you are, you may do me this
kindness of getting the King to give you his orders. I have run
every conceivable kind of evils during these twenty-four
years. I beg of you to excuse all the trouble I am giving you,
and I have the honour to be, Monsieur le Comte,
“Your very obedient servant,

“Charlotte Atkyns.”
This appeal seems to have been no more successful than the
preceding ones, for three months later we find Lady Atkyns still
awaiting the promised audience. To distract her thoughts from the
subject, she goes about Paris—a new city now to her—is present at
sittings of the Chamber of Deputies, and hears the speech from the
Throne. On All Souls’ Day she joins in the solemn pilgrimage to the
Conciergerie. Who could have been more in place on such an
occasion? But with the sad thoughts evoked by the sight of the
Queen’s prison were mingled regrets that the sanctuary had not
been left as it was. The place had been enlarged, and a massive,
heavy-looking tomb stood now where the bed had been.
“I knelt before this tomb,” she writes to Mme. de Verrière,
“but I should have preferred to have seen the prison room
unaltered, and the tomb placed where the Queen used to
kneel down to pray. The place has been made to look too
nice, and a simple elegance has been imparted to it which
takes away all idea of the misfortunes of that time. I would
have left the bed, the table, and the chair. There is a portrait
of the Queen seated on the bed, her eyes raised
heavenwards with the resignation of a martyr. This portrait is
very like, especially the eyes, with that look of angelic
sweetness which she had. There is another tomb with a
crucifix on it, as on hers, upon which are inscribed the words:
‘Que mon fils n’oublie jamais les derniers mot de son père,
que je lui répète expressément; qu’il ne cherche jamais à
venger notre mort.’ You go in by the chapel, and behind the
altar, to get to where the Queen used to be.... I repeated on
the tomb what I vowed to the Queen—never to abandon the
cause of her children. It is true that only Madame remains
now, but she one day will be Queen of France, and if she has
need of a faithful friend she will find one in me.”
These last lines seem a strange avowal. Lady Atkyns seems to be
renouncing her faith. What is the explanation? It is simple enough.
She has realized that as long as she puts forward her inopportune
plea regarding the child in the Temple she must expect to find
nothing but closed doors. Yet she has by her proofs of what she
alleges, and she is prepared in substantiation of her memorials to
hand over a selection of the precious letters from her friends which
she has received in the course of her enterprise. These, doubtless,
would be accepted, but would never be given back. What, then, is
she to do? Threatened on the one side by the distress which is at
her heels; a prey, on the other hand, to her inalterable conviction, the
luckless lady comes for a moment to have doubts about her entire
past. However, this disavowal, as it seems, is but momentary; a
calmer mood supervenes, and she returns to her former point of
view, unable ever to free her mind from doubts as to the real fate of
the Dauphin.
The King’s generosity in this year, 1816, does not appear to have
given her much satisfaction.
“At last I have received a little money,” she writes to Mme. de
Verrière just before Christmas, when preparing to return to England,
“but so little that it is really shameful.”
The following spring she is back again in France, still carrying on
her campaign. From 1817 to 1821 her letters pour in upon the
Ministry of the Royal Household. Did they contain indiscreet
allusions to the affair of the Temple? Perhaps. In any case, with a
single exception, all these letters have disappeared.
We find a curious reference to Lady Atkyns in a letter dated
January 11, 1818, preserved in the archives of the Comte de Lair—
“She is still in Paris,” says the writer. “For the last two
months she has been going every week! She declares now
she will start without fail on Tuesday morning, but the Lord
knows whether she will keep her word.... She is still taken up
with the affair in question, and passes all her time in the
company of those who are mixed up in it. I assure you I don’t
know what to make of it all myself, but it is certain that a
number of people believe it.”
The “affair in question” was the detention at Bicêtre of an
individual about whom the most sensational stories were current. A
maker of sabots, come over—no one quite knew how—from
America, Mathurin Bruneau, playing anew the Hervagault comedy,
had been passing himself off as the Dauphin. Arrested and
imprisoned on January 21, by order of Decazes, the Minister of
Police, Bruneau had for two years been leading a very extraordinary
life for a prisoner.
He was by way of being in solitary confinement, but there was in
reality a never-ending succession of visitors to him in his prison. A
certain Branzon, formerly a customs-house officer at Rouen, who
had been condemned to five years’ imprisonment with hard labour,
had become his inseparable companion. With the support of a
woman named Sacques and a lady named Dumont, Branzon got
together a species of little court round the adventurer, issuing
proclamations, carrying on a regular correspondence with friends
outside, and playing cards until three o’clock in the morning—finally
composing, with the help of large slices out of Le Cimetière de la
Madeleine, a work entitled Mémoires du Prince. Some unknown
painter executed a portrait of the prisoner as “a lieutenant-colonel or
colonel-general of dragoons,” and a mysterious baron, come from
Rouen to set eyes upon his Sovereign, took the oath of fidelity to him
on the Holy Scriptures in the jailer’s own room! On April 29, 1817,
the walls of Maromme, Darnétal, and Boudeville, near Rouen, were
covered with placards calling upon France to proclaim its legitimate
King. And all this happened under the nose of Libois, the Governor
of Bicêtre.
There seems, in fact, to be no room for doubt that, as has been
well said, “in this prison, in which there has been a constant
procession of comtes and abbés, and a whole pack of women, there
has been enacted in the years 1816-1818 a farce of which his
Excellence Decazes is the author.” The object of this mystification
was simply to baffle the Duchesse d’Angoulême in the first instance,
and to prevent public opinion from being led astray in another
direction. Bruneau did not stand alone. Six months earlier another
pretendant, Nauendorff, a clockmaker at Spandau, had written to the
Duchesse d’Angoulême to solicit an interview. It was all important to
put a stop to this dangerous movement. Therefore when on February
9, 1818, the proceedings were opened at Rouen, no pains had been
spared to give the affair the appearance of a frivolous vaudeville. On
February 19 Mathurin was condemned to five years’ imprisonment.
The court was crowded with all kinds of loafers and queer
characters, many of them from Paris, drawn by the rumours so
industriously spread about.
Lady Atkyns would seem to have given some attention to this new
alleged Dauphin without being carried off her feet. She lost no
opportunity of endeavouring to get at the truth, it is clear, and this, as
we learn from a police report, involved a number of visits to the
house in which Gaillon was imprisoned, and to which Bruneau was
transferred after his condemnation. It was even stated that she had
offered sums of money to enable Bruneau to escape. She soon had
her eyes opened, however, to this new fraud.
The accession of Monsieur to the throne, in 1824, does not seem
to have had any favourable result for Lady Atkyns, for we find her at
last reduced in this year to taking a step, long contemplated but
dreaded—the handing over of Ketteringham to her sister-in-law,
Mary Atkyns, in consideration of a life annuity.
She continues, however, to make her way every year to France,
buoyed up by the assurances of interest in her which she has
received from officials of the Royal Household. At first she stays with
friends, the Comte and Comtesse de Loban. Then in 1826, when her
mother dies, aged eighty-six, she establishes herself definitively in
Paris, taking up her abode in a house in the Rue de Lille, No. 65,
where she rents a small appartement on the first floor. Here she gets
together the few souvenirs she has saved from Ketteringham—some
mahogany furniture covered with blue cloth, a sofa covered with light
blue silk, and portraits on the walls of the Dauphin, his father, his
uncle, and the Duc de Berry.
It was while residing here that Lady Atkyns lived through the
revolution of Italy, after witnessing in turn the reign of Louis XVI., the
Terror, the Empire, the Restoration, and the reign of Charles X. What
an eventful progress from the careless, happy days when she played
her part in the dizzying gaieties of Versailles!
Some weeks before the fall of Charles X., Lady Atkyns drew up yet
another petition for presentation to the chief of the King’s household.
She did not mince her words in this document.
“I little thought that lack of funds would be advanced as a
reason for delaying the execution of the King’s orders. I will
not enlarge upon the strangeness of such an avowal,
especially as a reimbursement of so sacred a character is in
question, sanctioned by the Royal will. I would merely point
out to you, Monsieur le Marquis, that I have contrived to find
considerable sums (thereby incurring great losses) when it
was to the interest of France, and of her King, and of her
august family. Failing a sufficiency of money to liquidate this
debt, I have the honour to propose to your Excellency that
you should make out an order for the payment, and I shall find
means of getting it discounted. In your capacity as a Minister
to the King, your Excellency will be able, without delay, to
obtain the amount necessary, minus a discount, from the
Court bankers. Will you not deign, monseigneur, to ask them
to do this, and I shall willingly forego the discount that may be
stipulated for.... Finally, monseigneur, I beg of you to tell me
immediately the day and the hour when I may present myself
at the Ministry to terminate this matter. I must venture to
remind you that the least delay will involve my ruin, and
therefore I cannot consent to it.”
Lady Atkyns’s persistence and the King’s procrastination seem
intelligible enough when one learns that the sums expended by her,
from the time when Louis XVI.’s reign was projected down to the last
year of the Consulate, amounted to more than £80,000. The
Englishwoman might well speak of the sacrifices she had made and
the loss of her fortune at the dictates of her heart.
One other letter we find amongst Lady Atkyns’s papers—a letter
notable for its fine, regular penmanship. It evidently reached her
about this date. The writer was yet another soi-disant Dauphin, the
third serious pretendant. The Baron de Richemont—his real name
was Hebert—had published in 1831 his Mémoires du Duc de
Normandie, fils de Louis XVI. écrits et publiés par lui-même, and he
was not long in convincing a number of people as to his identity. He
probably owed most of his particulars as to his alleged escape from
the Temple to the wife of Simon, whom he had visited at the Hospital
for Incurables in the Rue de Sèvres. Possibly it was through her also
that he heard of Lady Atkyns. At all events, he thought it worth while
to approach her.
“Revered lady,” he writes to her, “I am touched by your kind
remembrance.... The idea that I have found again in you the
friend who was so devoted to my unhappy family consoles
me, and enables me the better to bear up under the ills that
Providence has sent me. I shall never forget your good
deeds; ever present to my memory, they make me cherish an
existence which I owe to you. I cannot tell what the future may
have in store for me, but whatever my fate you may count
upon all my gratitude. May the Lord be with you and send
prosperity to all your enterprises! He will surely do so, for to
whatever country you may take your steps, you will set an
example of all the virtues.
“We shall see you, I hope, in a better world. Then and in the
company of the august and ill-fated author of my sad days,
you will be in enjoyment of all the good you have done, and
will receive your due recompense from the Sovereign
Dispenser of all things.
“There being no other end to look for, I beg of God, most
noble lady, to take you under His protection.
“Louis Charles.”
Richemont shows some aptness and cleverness in the way he
touches the note of sensibility, and attains to the diapason
appropriate to the rôle he is playing. Had his letter the effect
desired? It is hardly likely, but it is the last item in Lady Atkyns’s
correspondence, and we have no means of finding an answer to the
question.

In the night of February 2, 1836, Lady Atkyns died. By her bedside


one person watched—her devoted servant, Victoire Ilh, whose
conduct, according to her mistress’s own statement, “had at all times
been beyond praise.”
The few friends who could attend gathered together in due course
to pay the last honours to the dead. Her remains were conveyed to
England for burial at Ketteringham, in accordance with the wish she
herself had expressed.
Time passed inexorably over her memory, and twenty years later
there was nothing to recall the life of love and devotion of this loyal
and unselfish Englishwoman.

FOOTNOTES:
[76] This is far below the actual figure.
EPILOGUE
Not more than two or three generations separate us from the
period through which we have seen come and go the various actors
involved in the enterprise of which the prisoners in the Temple were
the stake. The rôle that was played by Lady Atkyns and her
confidants, forming as it did a minor episode in the changeful story of
the emigration and of the fortunes of the Royalists during the
Revolution, deserved to be set forth. But the interest attached to
such narratives becomes greatly intensified the more completely the
records of those mentioned in it can be traced to the end. It seems
well, therefore, to see what became of the principal performers who
have passed before our eyes in this slight study.
Of Cormier’s two sons, Achille, the elder, disappears from the
scene completely, and all our efforts to trace him have been in vain.
In the case of his brother we have been more fortunate. Having
served as an officer in the army of Vendée, Patrice de Cormier, from
the moment of the Restoration, sought to return to active service.
Being on terms of intimacy with Prince de la Tremoïlle, Frotté’s old
friend, who was then presiding over the commission inquiring into
the claims of Royalist officers with a view to according recompense
to them, Cormier petitioned for employment in the company of light
horse. His loyalty was not in doubt, for a memorandum supporting
his application recorded that “when the allies entered Paris, he
secured a drum of the National Guard by purchase, and beating it in
front of a white flag, made his way through the streets of Paris.” The
“Hundred Days” interfered with his ambitions, and he was obliged to
betake himself to England, whence he made his way back to France
in July, 1815. The warmth with which Prince de la Tremoïlle
recommended him to the favour of Louis XVIII. showed what value
he attached to his friendship.
“He has accompanied me constantly as my aide-de-camp,”
said the Prince, “on my mission, at the time of my arrest and
during my escape, and he has never failed to give me new
proofs daily of his intelligence and zeal, of his boundless
devotion to the King, and of his capacity. If ever services
rendered could justify me in recommending any one to the
favour of the King, this estimable officer would be the first I
should venture to recommend.”
These words produced their effect. On February 24, 1816,
Cormier was appointed “chef de bataillon” in the first regiment of
infantry of the Royal Guard. Three years later he became a
lieutenant-colonel, and as such he took part in the Spanish
expedition of 1823, under the command of the Duc d’Angoulême.
Charged with the carrying out of an order to the Royal Spanish
troops before Figuières, he fell into an ambuscade of thirty
Constitutional soldiers, and received their volley at a distance of a
few yards. By a miracle he escaped with a wound on the hip, and
succeeded in fulfilling his mission.
Promoted to the rank of colonel, November 1, 1823, Cormier was
stationed at the Garrison at Rochefort at the outbreak of the
Revolution of July. Refusing to serve under the new régime, he sent
in his resignation to the Minister for War, August 5, 1830. This is the
last we hear of him. He died in a suburb of Paris.
His uncle, de Butler, after living for some time in Hamburg, where
he doubtless was regarded by the other émigrés with suspicion by
reason of his intimate relations with Bourrienne, Minister to the
Emperor, returned to England, where he is lost sight of. He died at
Gothenburg in Sweden in 1815.
Bereft of his two sons, Comte Henri de Frotté remained in England
entirely alone for a time, but returned to France on the restoration of
the Bourbons, obtained the rank of maréchal de camp, and died in
Paris, February 28, 1823. An enthusiastic Royalist, active and keen,
the Comte de Frotté had never ceased to interest himself in the
welfare of the émigrés in England, and came to be regarded as their
benefactor.
The career of Jean-Gabriel Peltier was of a more singular
description. This energetic pamphleteer had been editing in London
ever since 1802 a journal entitled L’Ambigu, in which he unceasingly
vented his spleen against the “Premier Consul.” His violence went to
such lengths that in 1803, on the reiterated demand of the incensed
Napoleon, he was brought before an English tribunal. He was
defended by the famous counsel, James Mackintosh, and received
only a mild sentence, with the result that he left the court in triumph,
and attained wide celebrity throughout Europe over the affair. In
addition to his newspaper work, Peltier was interested in a number of
publishing enterprises, which helped to make a livelihood for him.
Some years later the Fates made him chargé d’affaires to the
Emperor of Haïti. The amusement produced by this strange
appointment may be imagined. What made the thing still funnier was
the fact that His Haïtian Majesty paid his representative not in money
but in kind, transmitting to him cargoes of sugar and coffee. Peltier
negotiated these from time to time for the benefit of his creditors.
On the return of the Bourbons, Peltier returned to France, hoping,
like every one else, for his share of recompense, but only to be
disillusioned. He consoled himself with the reflection that if his King
treated him like a nigger, at least his nigger (the Emperor of Haïti!)
treated him like a king—

“Mon roi me traite comme un nègre


Mais mon nègre a son tour me traite comme un roi.”

Unfortunately, the “nègre” soon had enough of his epigrams, and


abandoned him, and this brought about his ruin. Haying lost all his
means of subsistence, he went back once more to Paris to implore
pity from the King, but in vain, and on March 29 he died miserably in
an attic in the Rue Montmartre, aged sixty-five.
APPENDIX
We print here a certain number of letters and documents found for
the most part among the unpublished papers of Lady Atkyns, not
used in the body of this book, yet too interesting to be entirely
omitted. The letters of the Princess de Tarente, in particular, seem to
deserve inclusion in their entirety.
Letter from Jean-Gabriel Peltier to Lady Atkyns.
“London, January 1, 1803.
“I have the honour of sending you, Madame, a letter which I
received yesterday from my friend.[77] The ferment Paris is
now in makes me fear that he may have been obliged to
leave the night of December 27-28, and it must have been
very stormy.
“I have at last managed to get at Mr. Burke in the House of
Commons. He has promised me an interview at as early a
date as possible. I introduced M. Goguelat to him, and he
seemed very glad to make his acquaintance. He had been
driving the evening before with M. de Choiseul, Mr. Pitt, Lord
Grenville, and Lord Loughborough, at Lord Hawkesbury’s. We
had time only for a word.
“I cannot close my letter, January 1, without sending you à
la Française, my good wishes for the New Year. I know well
what the object is that you yourself would wish most to see
achieved.”
Letter from Louis de Frotté to Lady Atkyns.
“London, December 10, 1794.
“Lovers and Ministers who don’t realize their opportunities
often regret them afterwards when they are gone, never to be
found again. This is what I fear is happening to us ... for your

You might also like