Negative sampling in the word2vec notebook. What is a language model? N-gram models (unigrams, bigrams, trigrams), together with their probability modeling and issues. Chain rule and n-gram estimation. Static vs. contextualized embeddings. Introduction to Recurrent Neural Networks.
Home Page and Blog of the Multilingual NLP course @ Sapienza University of Rome
Friday, March 22, 2024
Lecture 6 (15/03/2024, 4h): Negative sampling, homework 1 assignment
Negative sampling: the skipgram case; changes in the loss function. Homework 1 assignment.
Thursday, March 14, 2024
Lecture 4 (08/03/2024, 4h): first hands-on with PyTorch with language detection
Recap of the Supervised Learning framework, hands on practice with PyTorch on the Language Detection Model: tensors, gradient tracking, the Dataset and DataLoader class, the Module class, the backward step, the training loop, evaluating a model.
Lecture 3 (07/03/2024, 2h): Supervised vs. unsupervised vs. reinforcement learning. PyTorch
Thursday, March 7, 2024
Lecture 2 (01/03/2024, 4h): Machine Learning for NLP and Logistic Regression
Basics of Machine Learning for NLP. Probabilistic classification. Logistic Regression and its use
for classification. Explicit vs. implicit features. The cross-entropy
loss function.
Lecture 1 (29/2/2024, 2h): Introduction
Introduction to the course. Introduction to Natural Language Processing: understanding and generation. What is NLP? The Turing Test, criticisms and alternatives. Tasks in NLP and its importance (with examples). Key areas and publication venues.