C S 574
Download as PDF
Transformer Models for Natural Language Processing
Computer Science
College of Physical and Mathematical Sciences
Course Description
Understand and learn how to use state of the art transformer language models for many Natural Language Processing and Understanding tasks, such as text classification, text generation, question answering, machine translation, and many other tasks.
When Taught
Fall
Fixed
3
Fixed
3
Fixed
0
Title
Transformer Model
Learning Outcome
Students will study and demonstrate understanding of the transformer model and multi-headed attention mechanisms.
Title
Transformers for NLP
Learning Outcome
Students will demonstrate understanding and develop code that uses transformer models for many Natural Language Processing tasks such as text classification, named entity recognition, question answering, and summarization.
Title
Generative Models
Learning Outcome
Students will develop generative transformer models to generate appropriate text for a project of their choice.
Title
Zero Shot and Few Shot Learning
Learning Outcome
Students will demonstrate understanding of zero shot and few shot learning and prompt engineering for generative models.
Title
Scaling
Learning Outcome
Students will demonstrate understanding of scaling methods for large language models such as distillation, quantization, as well as methods for training models with small amounts of training data.