About This Course

A 12-week NLP course for undergraduate students, bridging statistical language models and modern deep learning.

Course Structure

Part 1: Foundations (1-3)

N-grams, Word2Vec, RNN/LSTM

Part 2: Core (4-6)

Seq2Seq, Transformers, BERT

Part 3: Advanced (7-9)

Scaling, Tokenization, Decoding

Part 4: Applications (10-12)

Fine-tuning, Efficiency, Ethics

Learning Objectives

  • Implement n-gram models with smoothing techniques
  • Build word embeddings (Word2Vec, GloVe)
  • Design RNNs and LSTMs for sequence modeling
  • Implement attention and transformer architectures
  • Fine-tune BERT/GPT for downstream tasks
  • Apply LoRA, quantization, and distillation
  • Evaluate models and detect biases

Technology Stack

Python 3.8+ PyTorch Hugging Face Jupyter NumPy Matplotlib

Prerequisites

  • Python programming proficiency
  • Linear algebra basics (vectors, matrices)
  • Probability and statistics fundamentals
  • Intro ML concepts (gradient descent)

Resources

Attribution

Developed by Digital AI Finance with QuantLet. All materials open-source.