Professional Documents
Culture Documents
Introduction to NLP
Diyi Yang
Some slides borrowed from Yulia Tsvetkov at CMU and Noah Smith at UW 1
Welcome!
2
Course Website
https://www.cc.gatech.edu/classes/AY2020/cs7650_spring
3
Communication With Machines
5
6
Question Answering
8
Natural Language Processing
10
Phonetics, Phonology
¡ Pronunciation Modeling
11
Words
¡ Language Modeling
¡ Tokenization
¡ Spelling correction
12
Morphology
¡ Morphology analysis
¡ Tokenization
¡ Lemmatization
13
Part of Speech
14
Syntax
¡ Syntactic parsing
15
Semantics
16
Discourse
17
Where Are We Now?
18
Where Are We Now?
VS
19
Where Are We Now?
20
Why NLP is Hard?
1. Ambiguity
2. Scale
3. Sparsity
4. Variation
5. Expressivity
6. Unmodeled Variables
7. Unknown representations
21
Why NLP is Hard?
1. Ambiguity
2. Scale
3. Sparsity
4. Variation
5. Expressivity
6. Unmodeled Variables
7. Unknown representations
22
Ambiguity
23
24
Ambiguity
and
Scale
25
The Challenges of “Words”
26
Part of Speech Tagging
27
Part of Speech Tagging
28
Part of Speech Tagging
29
Syntax
30
Morphology + Syntax
A ship-shipping
ship, shipping
shipping-ships
31
Semantics
¡ Every fifteen minutes a woman in this country gives birth.. Our job is to find
this woman, and stop her!
– Groucho Marx
32
Semantics
¡ Every fifteen minutes a woman in this country gives birth. Our job is to find
this woman, and stop her!
– Groucho Marx
33
Syntax + Semantics
34
Syntax + Semantics
35
Dealing with Ambiguity
36
Corpora
38
Why NLP is Hard?
1. Ambiguity
2. Scale
3. Sparsity
4. Variation
5. Expressivity
6. Unmodeled Variables
7. Unknown representations
39
Sparsity
40
Sparsity
41
Sparsity
42
Why NLP is Hard?
1. Ambiguity
2. Scale
3. Sparsity
4. Variation
5. Expressivity
6. Unmodeled Variables
7. Unknown representations
43
Variation
¡ Suppose we train a part of speech tagger or a parser on the Wall Street Journal
¡ What will happen if we try to use this tagger/parser for social media?
¡ “ikr smh he asked fir yo last name so he can add u on fb lololol”
44
Variation
45
Why NLP is Hard?
1. Ambiguity
2. Scale
3. Sparsity
4. Variation
5. Expressivity
6. Unmodeled Variables
7. Unknown representations
46
Expressivity
¡ Not only can one form have different meanings (ambiguity) but the same meaning
can be expressed with different forms:
¡ She gave the book to Tom vs. She gave Tom the book
¡ Some kids popped by vs. A few children visited
¡ Is that window still open? vs. Please close the window
47
Unmodeled Variables
World knowledge
I dropped the glass on the floor and it broke
I dropped the hammer on the glass and it broke 48
Unmodeled Representation
Very difficult to capture what is !, since we don’t even know how to represent the
knowledge a human has/needs:
¡ What is the “meaning” of a word or sentence?
¡ How to model context?
¡ Other general knowledge?
49
Desiderate for NLP Models
50
Symbolic and Probabilistic NLP
51
Probabilistic and Connectionist NLP
52
NLP vs. Machine Learning
53
NLP vs. Linguistics
54
Fields with Connections to NLP
¡ Machine learning
¡ Linguistics (including psycho-, socio-, descriptive, and theoretical)
¡ Cognitive science
¡ Information theory
¡ Logic
¡ Data science
¡ Political science
¡ Psychology
¡ Economics
¡ Education
55
Today’s Applications
¡ Conversational agents
¡ Information extraction and question answering
¡ Machine translation
¡ Opinion and sentiment analysis
¡ Social media analysis
¡ Visual understanding
¡ Essay evaluation
¡ Mining legal, medical, or scholarly literature
57
Factors Changing NLP Landscape
58
Logistics
59
What is this Class?
¡ Linguistic Issues
¡ What are the range of language phenomena?
¡ What are the knowledge sources that let us disambiguate?
¡ What representations are appropriate?
¡ How do you know what to model and what not to model?
¡ Statistical Modeling Methods
¡ Increasingly complex model structures
¡ Learning and parameter estimation
¡ Efficient inference: dynamic programming, search
¡ Deep neural networks for NLP: LSTM, CNN, Seq2seq 60
Outline of Topics
¡ Books:
¡ Primary text: Jurafsky and Martin, Speech and Language Processing, 2nd or 3 rd Edition
¡ https://web.stanford.edu/~jurafsky/slp3/
¡ Also: Eisenstein, Natural Language Processing
¡ https://github.com/jacobeisenstein/gt-nlp-class/blob/master/notes/eisenstein-nlp-notes.pdf
62
Course Website & Piazza
www.cc.gatech.edu/classes/AY2020/cs7650_spring/
piazza.com/gatech/spring2020/cs7650cs4650
63
Your Instructors
¡ Instructor:
¡ Diyi Yang
¡ Assistant professor
¡ Research interests: NLP, Computational Social Science
¡ TAs:
¡ Ian Stewart: PhD, Computational Sociolinguistics
¡ Jiaao Chen: PhD, NLP/ML
¡ Nihal Singh: MSCS, NLP
64
TA Office Hours
65
Grading
¡ 1 Midterm (15%)
66
Late Polices
¡ Late Policy
¡ 4 late days to use over the duration of the semester for homework assignments only.
There are no restrictions on how the late days can be used (e.g., all 4 can be used on one homework). Using
late days will not affect your grade. But homework submitted late after all 4 late days have been used will
receive no credit.
¡ No make-up exam
¡ Unless under emergency situation
67
Course Project
¡ Course Contacts:
¡ Webpage: materials and announcements
¡ Piazza: discussion forum
¡ Homework questions: Piazza, TAs’ office hours
¡ Computing Resources:
¡ Experiments can take up to hours, even with efficient computation
¡ Recommendation: start assignments early
69
What’s Next?
¡ Text Classification
70