Solutions of the course assignments.
- Describe fundamental tasks in natural language processing.
- Describe modern approaches to solve these tasks.
- Identify weaknesses and strengths of the proposed approaches.
- Implementing the base
Tensor
class. - Implementing differentiable operations for the
Computational Graph
. - Implementing
Momentum Training
feature for the optimizer. - Classification head and training loop provided in the skeleton code.
The main implementations is related to the Transformer
architecture.
- Implementing the
self-attention
head. - Implementing the
normalization
layer. - Implementing the
multiheaded
attentionforward
layer.
This part is a replication of a scientific paper and will be provided in later work.
- Carnegie Mellon University (CMU), Language Technologies Institute (LTI)
- Course website
- Instructors
- Graham Neubig (gneubig@cs.cmu.edu)
- Robert Frederking (ref@cs.cmu.edu)