Announcements
Course Description
This course is designed to introduce students to the fundamental
concepts and ideas in natural language processing (NLP), and to
get them up to speed with current research in the area. It
develops an in-depth understanding of both the algorithms
available for the processing of linguistic information and the
underlying computational properties of natural
languages. Word-level, syntactic, and semantic processing from
both a linguistic and an algorithmic perspective are
considered. The focus is on modern quantitative techniques in
NLP: using large corpora, statistical models for acquisition,
disambiguation, and parsing. Also, it examines and constructs
representative systems.
Prerequisites
-
Adequate experience with programming and formal structures
(e.g., CS106 and CS103X).
-
Programming projects will be written in Java, so knowledge of
Java (or a willingness to learn on your own) is required.
-
Knowledge of standard concepts in artificial intelligence
and/or computational linguistics (e.g., CS121/221 or Ling
138/238).
-
Basic familiarity with logic, vector spaces, and probability.
Intended Audience
Graduate students and advanced undergraduates specializing in
computer science, linguistics, or symbolic systems.
Textbook and Readings
The most used book will be:
-
Christopher Manning and Hinrich Schütze,
Foundations of Statistical Natural Language Processing.
MIT Press, 1999.
Buy
at Amazon ($67 new)!
Read the text
online!
We will distribute the most vital parts. It's referred to
as M&S below. Please see http://nlp.stanford.edu/fsnlp/
for supplementary information about the text, including errata,
and pointers to online resources.
Other useful reference texts for NLP are:
-
James Allen. 1995.
Natural Language Understanding.
Benjamin/Cummings, 2ed.
-
Gerald Gazdar and Chris Mellish. 1989.
Natural Language Processing in X.
Addison-Wesley.
-
Dan Jurafsky and James Martin. 2000.
Speech and Language Processing.
Prentice Hall.
Papers will occasionally be distributed and
discussed during the course of the class.
Copies of in-class hand-outs, such as readings and homework
assignments, will be posted on the syllabus, and hard copies will also be
available outside Gates 158 (in front of Prof. Manning's
office) while supplies last.
Homework and Grading
There will be four homeworks centered around substantial
programming assignments, each exploring a core NLP task.
In addition, there will be a final programming project on a
topic of your own choosing. A short, ungraded project proposal
will be due on Monday 5/9/05. Final project write-ups will be
due on the last day of class, Wednesday 6/1/05. Students will
give short project presentations during the time slot allocated
for the final exam, on Tuesday 6/7/05. You may find it helpful to
look at final
projects from previous years.
Course grades will be based 2/3 on homeworks (1/6 each) and 1/3
on the final project.
Be sure to read the policies on late
days and collaboration.
Section
Sections will be held most weeks to go over background
material, or to address issues related to the programming
assignments. Sections are optional, but students are
encouraged to attend for a better understanding of
background material and the assignments.
Syllabus
Wed 3/30/05 |
Introduction
[slides:
ppt,
pdf]
Overview of NLP. Statistical machine translation. Language models
and their role in speech processing.
Course introduction and administration.
Good background reading: M&S 1.0-1.4, 4.1-4.2,
Homework Collaboration Policy
Optional reading: Ken Church's tutorial
Unix for Poets
[ps, pdf]
(If your knowledge of probability theory is limited, also
read M&S 2.0-2.1.7. If that's too condensed, read the
probability chapter of an intro statistics textbook,
e.g. Rice, Mathematical Statistics and Data
Analysis, ch. 1.)
Homework 1 distributed today
|
Mon 4/4/05 |
N-gram Language Models and Information Theory
[slides: ps
MegaHal]
n-gram models. Entropy, relative entropy, cross
entropy, mutual information, perplexity. Statistical
estimation and smoothing for language models.
Assigned reading: M&S 2.2
Optional reading: Joshua Goodman (2001),
A Bit of Progress in Language Modeling, Extended Version
[pdf,
ps]
Optional reading: Stanley Chen and Joshua Goodman (1998),
An empirical study of smoothing techniques for language modeling
[pdf,
ps]
|
Wed 4/6/05 |
Statistical Machine Translation (MT), Alignment Models
[slides: pdf
ps]
Assigned reading: Kevin Knight,
A Statistical MT Tutorial Workbook
[rtf].
MS., August 1999.
Further reading: M&S 13
|
Fri 4/8/05 |
Section 1
[notes:
xls,
pdf]
Smoothing: absolute discounting, proving you have a proper
probability distribution, Good-Turing implementation.
Information theory examples and intuitions. Java
implementation issues.
|
Mon 4/11/05 |
Statistical Alignment Models and Expectation Maximization
(EM)
[slides:
pdf,
spreadsheet:
xls]
EM and its use in statistical MT alignment models.
Reference reading: Geoffrey J. McLachlan and
Thriyambakam Krishnan. 1997. The EM Algorithm and
Extensions. Wiley
Homework 2 distributed today
Homework 1 due today
|
Wed 4/13/05 |
Putting together a complete statistical MT system.
[slides:
pdf]
Decoding and A* Search. Recent work in statistical MT.
Further reading: Brown, Della Pietra, Della Pietra, and Mercer,
The Mathematics of Statistical Machine Translation: Parameter Estimation
[pdf,
pdf].
Computational Linguistics.
Ulrich Germann, Michael Jahr, Kevin Knight, Daniel Marcu, and Kenji
Yamada. 2001.
Fast
Decoding and Optimal Decoding for Machine Translation. ACL.
K. Yamada and K. Knight. 2002.
A Decoder for Syntax-Based Statistical MT. ACL.
|
Fri 4/15/05 |
Section 2
[notes:
xls]
The EM algorithm.
|
Mon 4/18/05 |
Word Sense Disambiguation (WSD) and Naïve Bayes (NB) Models
[slides: pdf]
Information sources, performance bounds, dictionary methods,
supervised machine learning methods, Naïve Bayes
classifiers.
Assigned Reading: M&S Ch. 7.
Reference:
Computational
Linguistics 24(1),
1998. Special issue on Word Sense Disambiguation.
Proceedings
of Senseval-3: The Third
International Workshop on the Evaluation of Systems for the
Semantic Analysis of Text
|
Wed 4/20/05 |
Maximum Entropy Classifiers
[slides: pdf]
Assigned Reading: class slides.
Other references: Adwait Ratnaparkhi.
A
Simple Introduction to Maximum Entropy Models for Natural Language
Processing.
Technical Report 97-08, Institute for Research in Cognitive Science,
University of Pennsylvania.
M&S section 16.2
|
Fri 4/22/05 |
Section 3
[notes: txt]
Corpora and other resources.
Homework 2 due today
|
Mon 4/25/05 |
Maximum Entropy Classifiers, Part II
[slides: pdf]
Assigned Reading: class slides.
Other references: Adwait Ratnaparkhi.
A
Simple Introduction to Maximum Entropy Models for Natural Language
Processing.
Technical Report 97-08, Institute for Research in Cognitive Science,
University of Pennsylvania.
M&S section 16.2
Adam Berger,
A Brief Maxent Tutorial
Homework 3 distributed today
|
Wed 4/27/05 |
Part of Speech Tagging and Sequence Inference
[slides: pdf]
Parts of speech and the tagging problem: sources of
evidence; easy and difficult cases.
Probabilistic sequence inference: Hidden Markov Models
(HMMs), Conditional Markov Models (CMMs), and the Viterbi
algorithm.
Assigned reading: M&S Ch. 10, pp. 341-356.
Further reading on HMMs: M&S Ch. 9.
HMM POS tagger: Thorsten Brants, TnT - A
Statistical Part-of-Speech Tagger, ANLP 2000.
CMM POS tagger: Kristina Toutanova and Christopher
D. Manning. 2000.
Enriching the
Knowledge Sources Used in a Maximum Entropy Part-of-Speech
Tagger. EMNLP 2000.
|
Fri 4/29/05 |
Section 4
Maximum entropy models, HMMs
|
Mon 5/2/05 |
Named Entity Recognition (NER) and Information Extraction (IE)
[slides: pdf]
Evaluation reading: M&S 8.1
HMMs for IE reading:
Dayne Freitag and Andrew McCallum (2000),
Information Extraction with HMM Structures Learned by
Stochastic Optimization, AAAI-2000
Maxent NER reading: Jenny
Finkel et al., 2005. Exploring the Boundaries: Gene and Protein Identification
in Biomedical Text
Background IE reading:
Ion Muslea (1999),
Extraction Patterns for Information
Extraction Tasks: A Survey
[pdf,
ps],
AAAI-99 Workshop on Machine Learning for Information Extraction.
Background IE reading:
Douglas E. Appelt. 1999. Introduction
to Information Extraction Technology
|
Wed 5/4/05 |
Parsing for Context-Free Grammars (CFGs)
[slides: pdf]
Top-down parsing, bottom-up parsing, empty constituents, left
recursion.
Background reading: M&S 3 (if you haven't done
any linguistics courses) or J&M ch. 9
Optional reading: J&M ch. 10
Final project guide
distributed today
|
Fri 5/6/05 |
no section today
Homework 3 due today
|
Mon 5/9/05 |
Dynamic Programming for Parsing
[handout: pdf]
Dynamic programming methods, chart parsing, the CKY algorithm.
Optional reading: J&M ch. 10
Homework 4 distributed today
Final project proposals due today
|
Wed 5/11/05 |
Probabilistic Context-Free Grammars (PCFGs)
[slides:
pdf (probparse),
pdf (search),
pdf
(unlexicalized)]
PCFGs, finding the most likely parse, refining PCFGs
Other questions for PCFGs: the inside-outside algorithm,
and learning PCFGs.
Assigned reading: M&S Ch. 11
|
Fri 5/13/05 |
Section 5
Parsing, PCFGs
|
Mon 5/16/05 |
Modern Statistical Parsers [slides: see last time]
Parsing for disambiguation, weakening independence assumptions,
lexicalization, search methods, Charniak's parser, probabilistic
left corner grammars, parser evaluation.
Assigned reading: M&S 8.3, 12
Optional readings:
Eugene Charniak (2000),
A Maximum-Entropy-Inspired Parser, Proceedings of NAACL-2000.
Eugene Charniak (1997),
Statistical techniques for natural language parsing,
AI Magazine.
Eugene Charniak (1997),
Statistical parsing with a context-free grammar and word
statistics, Proceedings of the Fourteenth National
Conference on Artificial Intelligence. AAAI Press/MIT
Press, Menlo Park (1997).
|
Wed 5/18/05 |
Question Answering (QA)
[handout: pdf]
TREC-style robust QA, natural language database interfaces
Assigned reading:
Marius Pasca, Sanda M. Harabagiu.
High Performance Question/Answering. SIGIR 2001: 366-374.
|
Fri 5/20/05 |
no section today
Homework 4 due today
|
Mon 5/23/05 |
Compositional Semantics
Semantic representations, lambda calculus, compositionality,
syntax/semantics interfaces, logical reasoning.
Assigned reading:
An Informal but Respectable Approach to Computational Semantics
[pdf,
ps]
|
Wed 5/25/05 |
Compositional Semantics
Assigned reading:
I. Androutsopoulos et al.,
Language Interfaces to Databases
|
Mon 5/30/05 |
Memorial Day
no class
|
Wed 6/1/05 |
Dialog & Discourse Systems
Rhetorical structure, planning and requests.
Assigned reading: handout
Optional reading: Gazdar & Mellish ch. 10
Final projects due today
|
Tue 6/7/05 |
Final Project Presentations
|
|
|
Course Information
Electronic Communications
Web:
http://cs224n.stanford.edu
Newsgroup:
su.class.cs224n
Questions mailing list:
cs224n-spr0405-staff@lists.stanford.edu
Send your questions here!
Announcements mailing list:
cs224n-spr0405-students@lists.stanford.edu
Enrolled students are automatically subscribed. Others wishing to
receive announcements should send an email to majordomo@lists.stanford.edu
with message body "subscribe cs224n-spr0405-guests".
Homeworks
Homework 1 (due 4/11/05)
Homework 2 (due 4/20/05)
Homework 3 (due 5/4/05)
Homework 4 (due 5/18/05)
Final project
Late Day Policy
Regrading Policy
Homework Collaboration Policy
Handouts
Lecture slides: intro
[ppt,
pdf]
Lecture slides: n-grams
[ps]
M&S Chapters 1 & 4
[ps]
M&S Chapter 3
[pdf]
M&S Chapter 6
[ps]
M&S Chapter 10
[pdf]
M&S Chapters 11 & 12
[pdf]
Section notes: smoothing
[xls,
pdf]
Section notes: EM
[xls]
Lecture slides: WSD
[pdf]
Lecture slides: MaxEnt
[pdf]
Section notes: corpora etc.
[txt]
Staff
Professor: Chris Manning
Office: Gates 158
Office Hours: M 4-5, W 2-3
Phone: 650-723-7683
Fax: 650-725-2588
Email:
manning@cs.stanford.edu
TA: Bill MacCartney
Office: Gates 114
Office Hours: Tu Th 11-12, W 1-2
Phone: 650-723-3796
Email:
TA: Guy Isely
Office: Gates B24A
Office Hours: M F 10-11
Email:
guyi@stanford.edu
Admin: Colleen Scott-Fields
Office: Gates 150
Phone: 650-723-0748
Email:
colleen8@cs.stanford.edu
Links
The Stanford NLP Group
Linguistic Corpora at Stanford
Statistical NLP links
Probabilistic parser links
Java 1.5 Overview
Java 1.5 New Features
|