The EM algorithm for word alignment in SMT. Beam search for decoding. Introduction to neural machine translation: the encoder-decoder neural architecture; back translation; byte pair encoding. The BLEU evaluation score. Performances and recent improvements. End of the course!
Home Page and Blog of the Multilingual NLP course @ Sapienza University of Rome
Monday, June 4, 2018
Lecture 25 (31/05/2018): bilingual/multilingual embeddings; semantic parsing
Bilingual and multilingual embeddings. Offline vs. online embeddings. Semantic parsing: definition, comparison to Semantic Role Labeling, approaches, a recent approach in detail. The Abstract Meaning Representation formalism. Introduction to machine translation (MT) and history of MT. Overview of statistical MT.
Friday, May 25, 2018
Wednesday, May 23, 2018
Sunday, May 20, 2018
Lecture 22 (18/05/2018): unsupervised and knowledge-based WSD
Unsupervised Word Sense Disambiguation: Word Sense Induction. Context-based clustering. Co-occurrence graphs: curvature clustering, HyperLex. Knowledge-based Word Sense Disambiguation. The Lesk and Extended Lesk algorithm. Structural approaches: similarity measures and graph algorithms. Conceptual density. Structural Semantic Interconnections. Evaluation: precision, recall, F1, accuracy. Baselines. Entity Linking.
Friday, May 18, 2018
Lecture 21 (17/05/2018): supervised Word Sense Disambiguation
Two important dimensions: supervision and knowledge. Supervised Word Sense Disambiguation: pros and cons. Vector representation of context. Main supervised disambiguation paradigms: decision trees, neural networks, instance-based learning, Support Vector Machines, IMS with embeddings, neural approaches to WSD.
Friday, May 11, 2018
Thursday, May 10, 2018
Lecture 19 (10/05/2018): semantic vector representations (2)
Semantic vector representations: importance of their multilinguality; linkage to BabelNet; latent vs. explicit representations; monolingual vs. multilingual representations. The NASARI lexical, unified and embedded representations.
Lecture 18 (04/05/2018): BabelNet; semantic vector representations (1)
Introduction to BabelNet (http://babelnet.org): multilingual synsets, resources integrated, accuracy, applications. Semantic vector representations: SensEmbed.
Friday, May 4, 2018
Lecture 17 (03/05/2018): computational lexicons; WordNet; introduction to WSD; homework 2
Encoding word senses: paper dictionaries, thesauri, machine-readable dictionary, computational lexicons. WordNet. Introduction to Word Sense Disambiguation (WSD). Homework 2: supervised and knowledge-based Word Sense Disambiguation.
Saturday, April 28, 2018
Lecture 16 (27/04/2018): introduction to computational and lexical semantics
Introduction to computational semantics. Syntax-driven semantic analysis. Semantic attachments. First-Order Logic. Lambda notation and lambda calculus for semantic representation. Lexicon, lemmas and word forms. Word senses: monosemy vs. polysemy. Special kinds of polysemy. Computational sense representations: enumeration vs. generation. Graded word sense assignment.
Lecture 15 (26/04/2018): transition-based syntactic parser; introduction to semantics
Transition-based dependency parsing with buffer and stack; main transition rules and their implementation with BiLSTMs. Introduction to semantics.
Saturday, April 21, 2018
Lecture 13 (19/4/2018): syntactic parsing (1/2)
Introduction to syntax. Context-free grammars and languages. Treebanks. Normal forms. Dependency grammars. Syntactic parsing: top-down and bottom-up. Structural ambiguity. Backtracking vs. dynamic programming for parsing. The CKY algorithm.
Friday, April 13, 2018
Friday, April 6, 2018
Friday, March 23, 2018
Friday, March 16, 2018
Lecture 5 (15/03/2018): language modeling
We introduced N-gram models (unigrams, bigrams, trigrams), together with their probability modeling and issues. We discussed perplexity and its close relationship with entropy, we introduced smoothing
Friday, March 9, 2018
Friday, March 2, 2018
Lecture 2 (02/03/2018): Introduction to NLP (2)
We continued our introduction to NLP, with a focus on the Turing Test as a tool to understand whether "machines can think". We also discussed the pitfalls of the test, including Searle's Chinese Room argument.
Wednesday, January 17, 2018
Ready, steady, go!
Welcome to the Sapienza NLP course blog! This year there will be important changes:
IMPORTANT: The 2018 class hour schedule will be on Thursday 16.30-19 and Fridays 14.00pm-16.30pm, Aula 2 - Aule L ingegneria.
Please sign up to the NLP class!
- You will write a paper
- The course will be much more deep learning oriented
- For attending students, there will be only three homeworks (and no additional duty), one of which will be done with delivery by the end of September and will replace the project. Non-attending students, instead, will have to work on a full-fledged project.
IMPORTANT: The 2018 class hour schedule will be on Thursday 16.30-19 and Fridays 14.00pm-16.30pm, Aula 2 - Aule L ingegneria.
Please sign up to the NLP class!
Subscribe to:
Posts (Atom)