Professional Documents
Culture Documents
111
Introduction
NLP Applications
http://www.geekwire.com/2013/ibm-takes-watson-cloud/
NLP Applications
• Search engines
– Google, Bing, DuckDuckGo, Baidu, Yandex
• Question answering
– IBM Watson
• Natural language assistants
– Siri, Alexa, Google, Cortana
• Machine translation
– Google Translate
• Email assistants
– Autocomplete
What is Natural Language Processing
• Natural Language Processing (NLP)
– The study of the computational treatment of natural (human) language.
– In other words, building computers that understand (and generate) language.
• Computational Linguistics (CL)
– The use of computers to study language
– Applications to social science, literature, finance
Notes
• Computers are designed to deal with human language
– Specific techniques are needed
• NLP draws on research in many fields
– Linguistics, Theoretical Computer Science, Mathematics,
Statistics, Artificial Intelligence, Psychology, Databases, etc.
Language and Communication
• Speaker
– Intention (goals, shared knowledge and beliefs)
– Generation (tactical)
– Synthesis (text or speech)
• Listener
– Perception
– Interpretation (syntactic, semantic, pragmatic)
– Incorporation (internalization, understanding)
• Both
– Context (grounding)
Basic NLP Pipeline
(U) (G)
Language Computer Language
Natural Language Understanding
root
ate
nsubj dobj nmod
• Inference
∀z,t0,t1,y,e: Hungry (z,t0) EatingEvent(e) Eater (e,z) Eaten (e,y) Time (e,t0)
Food (y) Precedes (t0,t1) Hungry (z,t1)
• Conclusion
Hungry (g1,now)
Machine Learning
Distributed representations
[Mikolov 2013]
Demos
• https://playground.tensorflow.org/
• https://p.migdal.pl/interactive-machine-learning-list/
• https://demo.allennlp.org/reading-comprehension
• https://allenai.org/demos/
• http://text-processing.com/demo/
• https://explosion.ai/demos/
• https://huggingface.co/hmtl/
• https://corenlp.run/
• http://nlp.stanford.edu:8080/corenlp/
• https://talktotransformer.com/
• https://cs.stanford.edu/people/karpathy/convnetjs/
Sequence to Sequence
112
Examples of Text
CPSC 477/577
• Instructor:
– Dragomir Radev
– dragomir.radev@yale.edu
• Class times:
– TTh 1-2:15
– Location: 211 Mason Lab
• Teaching staff:
– Irene Li (TF)
– Aarohi Srivastava (ULA), Sally Ma (ULA), others TBA
• Office hours
– Drago F 2-4 in 17 HH (Room 319)
Course Readings
• Speech and Language Processing
– Daniel Jurafsky and James Martin
– Third edition, 2019
– http://web.stanford.edu/~jurafsky/slp3/
• Introduction to Natural Language Processing
– Jacob Eisenstein
– First edition, 2019
– https://github.com/jacobeisenstein/gt-nlp-
class/blob/master/notes/eisenstein-nlp-notes.pdf
• Additional readings:
– Natural Language Processing using NLTK (Bird et al.)
http://www.nltk.org
– AAN http://aan.how
Course Dates
• Jan 14 16 21 23 28 30
• Feb 4 6 11 13 18 20 25 27
• Mar 3 5 24 26 31
• Apr 2 7 9 14 16 21 23
• Collocations
– Strong beer but *powerful beer
– Big sister but *large sister
– Stocks rise but ?stocks ascend
Class Logistics
Grading
• Assignments (50%)
– HW0+HW1 = 2+8=10%
– HW2 = 10%
– HW3 = 10%
– HW4 = 10%
– HW5 = 10%
• Exams (45%)
– midterm = 20%
– final exam = 25%
• Class participation (5%)
– In-class participation, asking questions on Piazza, answering questions, office hours
Sample Programming Assignments
• Language Modeling and Part of Speech Tagging
• Dependency Parsing
• Vector Semantics and Word Sense Disambiguation
• Question Answering
• Deep Learning
• Machine Translation
• Sentiment Analysis
• Natural Language Interface to a Database
• Semantic Parsing
How to get the most out of the class?
• Attend the lectures and study the slides
– Course syllabus + slides = road map
– Some material may not be found in any of the readings
• Hands on experience
– Implement what you’ve learned
• Ask questions in and after class
Questions?
112
Examples of Text
What NLP is not about
http://www.upassoc.org/upa_publications/jus/2009february/smelcer5.html
Literary Texts
• Project Gutenberg (http://www.gutenberg.org/browse/scores/top)
• A team of horses passed from Finglas with toiling plodding tread, dragging through the funereal silence a
creaking waggon on which lay a granite block. The waggoner marching at their head saluted.
– Ulysses - http://www.gutenberg.org/files/4300/4300-h/4300-h.htm
• There was no possibility of taking a walk that day. We had been wandering, indeed, in the leafless shrubbery an
hour in the morning; but since dinner (Mrs. Reed, when there was no company, dined early) the cold winter wind
had brought with it clouds so sombre, and a rain so penetrating, that further out-door exercise was now out of the
question.
– Jane Eyre - http://www.gutenberg.org/files/1260/1260-h/1260-h.htm
• Dorothy lived in the midst of the great Kansas prairies, with Uncle Henry, who was a farmer, and Aunt Em, who
was the farmer's wife. Their house was small, for the lumber to build it had to be carried by wagon many miles.
There were four walls, a floor and a roof, which made one room; and this room contained a rusty looking
cookstove, a cupboard for the dishes, a table, three or four chairs, and the beds. Uncle Henry and Aunt Em had
a big bed in one corner, and Dorothy a little bed in another corner. There was no garret at all, and no cellar--
except a small hole dug in the ground, called a cyclone cellar, where the family could go in case one of those
great whirlwinds arose, mighty enough to crush any building in its path. It was reached by a trap door in the
middle of the floor, from which a ladder led down into the small, dark hole.
– The Wizard of Oz - http://www.gutenberg.org/files/55/55-h/55-h.htm
A Really Long Literary Sentence
• Try parsing this
– “Bloat is one of the co-tenants of the place, a maisonette erected last century, not far
from the Chelsea Embankment, by Corydon Throsp, an acquaintance of the Rossettis'
who wore hair smocks and liked to cultivate pharmaceutical plants up on the roof (a
tradition young Osbie Feel has lately revived), a few of them hardy enough to survive
fogs and frosts, but most returning, as fragments of peculiar alkaloids, to rooftop earth,
along with manure from a trio of prize Wessex Saddleback sows quartered there by
Throsp's successor, and dead leaves off many decorative trees transplanted to the roof
by later tenants, and the odd unstomachable meal thrown or vomited there by this or
that sensitive epicurean-all got scumbled together, eventually, by the knives of the
seasons, to an impasto, feet thick, of unbelievable black topsoil in which anything could
grow, not the least being bananas.“
• “Gravity’s Rainbow” (by Thomas Pynchon), known for its use of very
arcane words and complicated sentence (and plot) structure.
• Another such work is “Finnegans Wake” by James Joyce.
• Poetry is even more difficult.
Introduction to NLP
• http://lily.yale.edu
• http://aan.how
• http://www.nacloweb.org
LILY Projects
• Text summarization
• Crosslingual information retrieval
• Analysis of electronic health records
• Survey generation
• Text to SQL
• Dialogue systems
• Multilingual computing
• Text generation
113
Brief History of NLP
The Turing Test
• Alan Turing: the Turing test
– language as test for intelligence
• Three participants
– a computer and two humans (one is an interrogator)
• Interrogator’s goal
– to tell the machine and human apart
• Machine’s goal
– to fool the interrogator into believing that a person is responding
• Other human’s goal
– to help the interrogator reach his goal
NACLO
NACLO and IOL
• The North American Computational Linguistics Open Competition
– Competition held since 2007 in the USA and Canada
– http://www.nacloweb.org
• Best individual US performers so far:
– Adam Hesterberg (2007)
– Rebecca Jacobs (2007-2009) – 3 team golds + 2 individual medals
– Ben Sklaroff (2010)
– Morris Alper (2011)
– Alex Wade (2012, 2013) – 2 team golds + 2 individual golds + 1 individual silver
– Tom McCoy (2013) – Yale grad
– Darryl Wu (2012, 2014)
– James Wedgwood (2015, 2016) – Yale senior
– Brian Xiao (2017), Swapnil Garg (2018)
• Other strong countries:
– Russia, UK, Netherlands, Poland, Bulgaria, South Korea, Canada, China, Sweden, Slovenia, Brazil, Taiwan
• IOL – the International contest
– Since 2003
– 2014 in China, 2015 in Bulgaria, 2016 in India, 2017 in Ireland, 2018 in Czechia, 2019 in Korea, 2020 in Latvia
– http://www.ioling.org
[A Donkey in Every House, by Todor Tchervenkov, NACLO 2007]
[Tenji Karaoke, by Patrick Littell, NACLO 2009]
[aw-TOM-uh-tuh, by Patrick Littell, NACLO 2008]
On her visit to Armenia, Millie has gotten lost in Yerevan, the
nation’s capital. She is now at the Metropoliten (subway) station
named Shengavit, but her friends are waiting for her at the station
named Barekamutyun. Can you help Millie meet up with her
friends?
Note that all names of stations listed below appear on the map.
a. Gortsaranayin
b. Zoravar Andranik
c. Charbakh
d. Garegin Njdehi Hraparak
e. none of the above
http://www.nacloweb.org/resources.php
NLP
Introduction to NLP
114
Why is language processing hard?
Silly Sentences
• Children make delicious snacks
• Stolen painting found by tree
• I saw the Rockies flying to San Francisco
• Court to try shooting defendant
• Ban on nude dancing on Governor’s desk
• Red tape holds up new bridges
• Government head seeks arms
• Cameron wins on budget, more lies ahead
• Local high school dropouts cut in half
• Hospitals are sued by seven foot doctors
• Dead expected to rise
• Miners refuse to work after death
• Patient at death's door - doctors pull him through
• In America a woman has a baby every 15 minutes. How does she do that?
The Winograd Schema Challenge
The Winograd Schema Challenge
More Classic Examples
• Beverly Hills
• Beverly Sills
• The box is in the pen
• The pen is in the box
• Mary and Sue are mothers
• Mary and Sue are sisters
• Every American has a mother
• Every American has a president
Ambiguous Words
• ball, board, plant
– meaning
• fly, rent, tape
– part of speech
• address, resent, entrance, number, unionized
– pronunciation – give it a try
Syntax vs. Semantics
[Chomsky 1957]
Syntactic Amgbiguity
• Phonetic:
– Joe’s finger got number.
• Part of speech:
– Joe won the first round.
• Syntactic:
– Call Joe a taxi.
• Sense:
– Joe took the bar exam.
Other Sources of Difficulty
• Subjectivity:
– Joe believes that stocks will rise.
• Cc attachment:
– Joe likes ripe apples and pears.
• Negation:
– Joe likes his pizza with no cheese and tomatoes.
• Referential:
– Joe yelled at Mike. He had broken the bike.
– Joe yelled at Mike. He was angry at him.
• Reflexive:
– John bought him a present.
– John bought himself a present.
• Ellipsis and parallelism:
– Joe gave Mike a beer and Jeremy a glass of wine.
• Metonymy:
– Boston called and left a message for Joe.
Other Sources of Difficulties
• Non-standard, slang, and novel words and usages
– A360, 7342.67, +1-646-555-2223
– “spam” or “friend” as verbs
– yolo, selfie, chillax – recently recognized as dictionary words
– www.urbandictionary.com – (Parental Warning!)
• Inconsistencies
– junior college, college junior
– pet spray, pet llama
• Typoes and gramattical erorz
– reciept, John Hopkins, should of
• Parsing problems
– Selbständigkeit (self-reliance)
– cup holder
– Federal Reserve Board Chairman
Other Sources of Difficulties
• Complex sentences
• Counterfactual sentences
• Humor and sarcasm
• Implicature/inference/world knowledge:
– I was late because my car broke down.
– Implies I have a car, I use the car to get to places, the car has wheels, etc.
– What is not explicitly mentioned, what is world knowledge?
• Lots of data
• Linguistic intuition
• Appropriate representation
• Robust algorithms
• World knowledge
• Grounding
[Example from Dan Klein]
NLP