Natural Language Processing

A BSc-level course bridging statistical language models with modern transformer architectures.

Word Embeddings 3D Word Embeddings
BERT Architecture BERT Architecture
Beam Search Beam Search

Course Topics

N-grams & Language Models

Statistical foundations of NLP: probability chains, Markov assumptions, and perplexity evaluation.

Beginner 2 hours

Word2Vec & Embeddings

Dense word representations, Skip-gram architecture, and semantic geometry in vector space.

Beginner 2.5 hours

RNN & LSTM Networks

Sequential processing, vanishing gradients, and memory mechanisms with gated architectures.

Intermediate 3 hours

Seq2Seq & Attention

Encoder-decoder models, attention mechanisms, and alignment in neural machine translation.

Intermediate 2.5 hours

Transformers

Self-attention, multi-head attention, positional encoding, and parallel processing.

Intermediate 3.5 hours

BERT & Pre-training

Pre-trained language models, masked language modeling, and transfer learning pipelines.

Intermediate 3 hours

Scaling Laws

Compute-optimal training, Chinchilla scaling, and emergent capabilities in large models.

Advanced 2 hours

BPE & Tokenization

Subword tokenization, Byte-Pair Encoding, and vocabulary optimization strategies.

Intermediate 2 hours

Text Generation & Decoding

Greedy, beam search, sampling strategies, temperature, and nucleus sampling.

Advanced 3 hours

Fine-tuning & LoRA

Parameter-efficient fine-tuning, adapters, LoRA, and prompt engineering techniques.

Advanced 3 hours

Efficiency & Quantization

Model compression, quantization, distillation, pruning, and deployment optimization.

Advanced 2.5 hours

Ethics & Bias

Bias detection, fairness metrics, WEAT tests, and responsible AI development.

Beginner 2 hours

Resources