CS224n: Natural Language Processing with Deep Learning
A distilled compilation of my notes for Stanford's CS224n: Natural Language Processing with Deep Learning.
Natural language processing (NLP) is a crucial part of artificial intelligence (AI), modeling how people share information. In recent years, deep learning approaches have obtained very high performance on many NLP tasks. In this course, students gain a thorough introduction to cutting-edge neural networks for NLP.
Each section is divided by topic to help you get bite-sized information that is easy to understand!
Natural language processing (NLP) is a crucial part of artificial intelligence (AI), modeling how people share information. In recent years, deep learning approaches have obtained very high performance on many NLP tasks. In this course, students gain a thorough introduction to cutting-edge neural networks for NLP.
Each section is divided by topic to help you get bite-sized information that is easy to understand!
📌 Table of Contents
Word Vectors
NLP overview; historical context; word vectors; Word2Vec (CBOW + Skip-gram)
NLP Tasks
named entity recognition; dependency parsing
Neural Networks
RNNs; LSTMs; CNNs; regularization; dropout
Language Models
n-gram language model; neural language model; ELMO; GPT-3
Machine Translation
seq2seq; neural machine translation
Attention
the bottleneck problem; seq2seq with attention; self attention; multi-headed attention
Knowledge Graphs
knowledge graph overview; entity linking; ERNIE; KGLM
Transformer
encoder; decoder; BERT; T5 language model
Tokenizer
process of tokenization; sub-word tokenization; Byte Pair Encoding (BPE); WordPiece; SentencePiece
Course Info
Course description:
- Natural language processing (NLP) or computational linguistics is one of the most important technologies of the information age. Applications of NLP are everywhere because people communicate almost everything in language: web search, advertising, emails, customer service, language translation, virtual agents, medical reports, etc.
- In recent years, deep learning (or neural network) approaches have obtained very high performance across many different NLP tasks, using single end-to-end neural models that do not require traditional, task-specific feature engineering.
- In this course, students will gain a thorough introduction to cutting-edge research in Deep Learning for NLP. Through lectures, assignments and a final project, students will learn the necessary skills to design, implement, and understand their own neural network models. As piloted last year, CS224n will be taught using PyTorch this year.
📸 Credits
The in-line diagrams are taken from the link's below:
- Stanford's CS224n: Natural Language Processing with Deep Learning. - Jay Alammar's Blog.
- Stanford's CS224n: Natural Language Processing with Deep Learning. - Jay Alammar's Blog.
Citation
If you found our work useful, please cite it as:
@misc{Chadha2020DistilledNotesCS231n, author = {Jain, Vinija and Chadha, Aman}, title = {Distilled Notes for Stanford CS224n: Natural Language Processing with Deep Learning}, howpublished = {\url{https://www.aman.ai}}, year = {2020}, note = {Accessed: 2020-07-01}, url = {www.aman.ai} }
A. Chadha, Distilled Notes for Stanford CS224n: Natural Language Processing with Deep Learning, https://www.aman.ai, 2020, Accessed: July 1 2020.