Multilingual Natural Language Processing

From language modeling to generation: deep learning for multilingual NLP

Basic information

The Multilingual Natural Language Processing course introduces the fundamentals of AI for automatically processing, understanding, and generating human language. Taught in English, this course introduces Natural Language Processing from foundational concepts to modern Large Language Models. Students begin with words, tokens, tokenization (including BPE), word embeddings and language modeling, and how to evaluate NLP systems’ outputs.

The course progresses to neural networks for NLP, attention, and the Transformer architecture, followed by Large Language Models, their training phases, scaling laws, PEFT, and hands-on applications (we will also cover the making of our Minerva LLM and its current evolution!). Advanced topics include machine translation, generative model evaluation, reinforcement learning for LLMs, Retrieval-Augmented Generation, advanced architectures (MoE, MAMBA), semantics, coreference resolution, and narrative understanding.

It is part of the curricula for the Master’s in AI and Robotics, Engineering in Computer Science and Artificial Intelligence, and Data Science.

Semester

Spring 2026

When and where

February 27 – May 29, 2026

  • Wednesday (10:15 – 12:00)
  • Friday (8:30 – 10:45)

Venue: S. Pietro in Vincoli, via delle Sette Sale, 29 (room/aula 41)

Classroom

All the class material can be found in the dedicated classroom: https://classroom.google.com/c/ODQyOTcxMjM1NTQx?cjc=2zvchzzx

Course Syllabus

Foundations of NLP: words, tokens and language models

  • Introduction to the course
  • Words & tokens
  • Tokenization techniques
  • Language models
  • Evaluation basics
Hands-on
Language modeling

Machine Learning for NLP

  • Machine Learning & classification
  • Logistic regression
  • Cross-entropy (CE)
  • Gradient descent
  • Evaluation: Accuracy, Precision, Recall, F-measure
  • Train/dev/test sets & statistical significance
Hands-on
Logistic regression

Word Representations & Semantics

  • Count-based semantics
  • Cosine similarity
  • Word2vec
  • Embedding properties
  • Visualization & bias
Hands-on
Word embeddings

Neural Networks & Transformer Encoders

  • Neural networks & deep learning
  • The attention mechanism
  • Transformer: The Encoder architecture
  • From BERT to mmBERT
  • Sentence embeddings
Hands-on
Transformer encoder models
Homework 1

Decoders: Large Language Models

  • Transformer: The Decoder architecture
  • Training phases (pretraining, post-training)
  • Scaling laws
  • PEFT
  • Reinforcement learning for LLMs
  • Retrieval-Augmented Generation
Hands-on
LLMs (including Minerva!) and RAG
Homework 2

Advanced Topics & Applications

  • Machine Translation
  • Evaluation of generative models
  • BabelNet and Word Sense Disambiguation
  • Semantic Role Labeling and Semantic Parsing
  • Coreference Resolution
  • Narrative Understanding
  • Advanced architectures (MoE, MAMBA)
Hands-on
Surprise!

Teaching Staff

Course Instructor

Prof. Roberto Navigli

Office: room B119, via Ariosto, 25

Email: surname chiocciola diag plus uniroma1 plus it (if you are a human being, please replace plus with . and chiocciola with @)

Website: https://www.diag.uniroma1.it/navigli/

Teaching Assistant
Marina Aur
Leonardo Colosi
Alberte Fernández
Bruno Gatti
Luca Gioffré
Elena Marafatto
Luca Moroni
Francesco Ortame