CS: 533 Intructor: Karl Stratos, Rutgers University
The topics covered are:
- Language modeling: n-gram models, log-linear models, neural models
- Deep learning in NLP: RNN, LSTM, Attention models, Transformers, BERT (transfer knowledge models)
- Structured Prediction in NLP: Tagging and Parsing (Constituency and Dependency)
- Unsupervised Learning in NLP: Latent-Variable Generative Models, EM Algorithm, Autoencoders, VAEs
- Information Extraction
- Large-Scale Transfer Learning
All the assignments are implemented in Python using PyTorch.
- N-gram models: Relative Frequency Lemma, Maximum Likelihood Estimation (MLE) of the Trigram Language Model
- Preliminary probability and statistics, Linear Algebra, Optimization
- Log-linear language models
- NLTK tokenizer
- Feedforward Neural Language Model
- Backpropogation in Neural Networks
- Self-attenion and LSTM models
- Transformers (encoder/decoder models)
- BLUE (bilingual evaluation understudy) algorithm
- Structured Prediction in NLP: Tagging
- Hidden Markov Models (HMMs)
- Conditional Random Fields (CRFs)
- Structured Prediction in NLP: Constituency and Dependency Parsing
- Probabilistic Context-Free Grammars (PCFGs)
- Expectation Maximization (EM) Algorithm
- Variational Autoencoders (VAE)