InCLow
InCLow
Home
People
Reading Group
Projects
Publications
Paper-Conference
Non Verbis, Sed Rebus: Large Language Models Are Weak Solvers of Italian Rebuses
Gabriele Sarti
,
Tommaso Caselli
,
Malvina Nissim
,
Arianna Bisazza
Cite
URL
BabyLM Challenge: Exploring the effect of variation sets on language model training efficiency
Akari Haga
,
Akiyo Fukatsu
,
Miyu Oba
,
Arianna Bisazza
,
Yohei Oseki
Cite
URL
Model Internals-based Answer Attribution for Trustworthy Retrieval-Augmented Generation
Ensuring the verifiability of model answers is a fundamental challenge for retrieval-augmented generation (RAG) in the question …
Jirui Qi
,
Gabriele Sarti
,
Raquel Fernández
,
Arianna Bisazza
Cite
DOI
URL
Multi-property Steering of Large Language Models with Dynamic Activation Composition
Daniel Scalena
,
Gabriele Sarti
,
Malvina Nissim
Cite
DOI
URL
NeLLCom-X: A Comprehensive Neural-Agent Framework to Simulate Language Learning and Group Communication
Yuchen Lian
,
Tessa Verhoef
,
Arianna Bisazza
Cite
URL
The SIFo Benchmark: Investigating the Sequential Instruction Following Ability of Large Language Models
Following multiple instructions is a crucial ability for large language models (LLMs). Evaluating this ability comes with significant …
Xinyi Chen
,
Baohao Liao
,
Jirui Qi
,
Panagiotis Eustratiadis
,
Christof Monz
,
Arianna Bisazza
,
Maarten De Rijke
Cite
DOI
URL
Democratizing Advanced Attribution Analyses of Generative Language Models with the Inseq Toolkit
Gabriele Sarti
,
Nils Feldhus
,
Jirui Qi
,
Malvina Nissim
,
Arianna Bisazza
PDF
Cite
DecoderLens: Layerwise Interpretation of Encoder-Decoder Transformers
In recent years, several interpretability methods have been proposed to interpret the inner workings of Transformer models at different …
Anna Langedijk
,
Hosein Mohebbi
,
Gabriele Sarti
,
Willem Zuidema
,
Jaap Jumelet
Cite
URL
Encoding of lexical tone in self-supervised models of spoken language
Gaofei Shen
,
Michaela Watkins
,
Afra Alishahi
,
Arianna Bisazza
,
Grzegorz Chrupała
Cite
DOI
URL
IT5: Text-to-text Pretraining for Italian Language Understanding and Generation
We introduce IT5, the first family of encoder-decoder transformer models pretrained specifically on Italian. We document and perform a …
Gabriele Sarti
,
Malvina Nissim
Cite
URL
«
»
Cite
×