Portrait of Marcos Treviso

Marcos Treviso

Assistant Professor at Instituto Superior Técnico, University of Lisbon

I work at the intersection of efficiency, long-context modeling, and interpretability in machine learning. I'm a co-PI at the SARDINE Lab 🐟 and a member of the ELLIS Unit Lisbon 🏛️.


News

Feb 2026

Our paper, Long-context Generalization with Sparse Attention was accepted at ICLR 2026. See you in Rio!

Sep 2025

New blog post: SLURM in the Wild: A Practical Guide for Academic Labs — 50 min read on scaling research compute.

Aug 2025

I started as Tenure-Track Assistant Professor at the Computer Engineering Department, IST — University of Lisbon.

May 2025

AdaSplash: Adaptive Sparse Flash Attention was presented at ICML 2025 as a Spotlight (top 1%) ⚡.

May 2025

LaTIM: Measuring Latent Token-to-Token Interactions in Mamba Models received an Outstanding Paper Award 🏆 at ACL 2025.


Recent Publications

Full list on Google Scholar →

2026

arXiv · Preprint

Sparse Attention as Compact Kernel Regression

Saul Santos, Nuno Gonçalves, Daniel C. McNamee, Marcos Treviso, André F. T. Martins

ICLR 2026

Long-Context Generalization with Sparse Attention

Pavlo Vasylenko, Hugo Pitorro, André F. T. Martins, Marcos V. Treviso

2025

ICML 2025 ⚡ Spotlight & Oral · Top 1%

AdaSplash: Adaptive Sparse Flash Attention

Nuno Gonçalves, Marcos V. Treviso, André F. T. Martins

ACL 2025 🏆 Outstanding Paper

LaTIM: Measuring Latent Token-to-Token Interactions in Mamba Models

Hugo Pitorro, Marcos V. Treviso

TAFFC 2025 · Journal

AMUSED: A Multi-Modal Dataset for Usability Smell Identification

Flavia Santos, Marcos Treviso, Kamila Rodrigues, Renata Fortes, Sandra Gama


Students

Current

PV
Pavlo Vasylenko PhD
HP
Hugo Pitorro PhD
AB
Afonso Baldo PhD
NG
Nuno Gonçalves PhD
GD
Gonçalo Duarte MSc
JG
Julio Galdino Visiting PhD
RL
Rodrigo Lima Visiting BSc
WL
Weiming Li Visiting PhD
EJ
Erik Jhones Researcher

Previous

TN
Titus Naber MSc
HP
Hugo Pitorro MSc
NG
Nuno Gonçalves MSc

I have open projects at MSc and PhD level — if you're interested in working on topics related to efficiency, long-context modeling, or interpretability, feel free to send me an email.


Tools, Tutorials, and Puzzles


Teaching

Invited Assistant Professor

  • Machine Learning — MEEC, IST · 120 students · Lab component

Teaching Assistant

Service

Reviewer

  • ACL, EMNLP, EACL, NAACL
  • ICML, ICLR, NeurIPS
  • PROPOR, STIL, EAMT
  • 🏅 Outstanding Reviewer — EMNLP, NeurIPS, ICLR (multiple years)

Area Chair & Senior Area Chair

  • ACL 2024 & 2025, EMNLP 2025, ACL 2026

ACL Tech Team


Research Projects

SMURF4EU

Co-PI EuroHPC · 2026–2027

A Suite of Multimodal Reasoning Foundation Models for Europe

Developing and releasing a suite of fully open, high-performance multimodal reasoning foundation models spanning text, code, speech, vision, and video — with support for all 24 official EU languages. Targets multiple model sizes and long multimodal contexts up to 1M tokens using efficient attention and memory-compression techniques.

AMALIA

Team member Portugal · 2025–2026

European-Portuguese Large Language Model

Project page →

An open LLM developed specifically for Portuguese as used in Portugal — preserving culturally grounded language use and supporting data sovereignty for Public Administration use cases. Built by a national consortium (NOVA, IST, Coimbra, Porto, Minho, FCT/Arquivo.PT), pre-trained on ~4 trillion words, with a roadmap toward multimodality and a targeted release around June 2026.