Categories
Meetings

LLM Beliefs Are in Their Heads

CCC Seminar by Alessandro Corona Mendozza, predoc researcher at the center for language Technology (Copenhagen University) and visiting at the University of Turin.

Abstract:
We investigate belief-like representations in decoder-only autoregressive LLMs using linear controlled probes on residual stream activations and single attention heads. Following Herrmann and Levinstein’s (2025) criteria (Accuracy, Use, Coherence, and Uniformity) we find that large models exhibit strong truth sensitivity (Accuracy), and steering activations along probe directions reliably changes downstream behavior (Use). Coherence, measured via calibrated probes and cross-dataset probing, is moderate across models, while training on diverse data yields domain-consistent truth directions (Uniformity). The results are particularly encouraging at the head level and align with some standard philosophical accounts of belief, e.g., minimal functionalism, supporting the view that LLMs can maintain propositional attitudes under such theoretical frameworks.

Short bio:
Alessandro Corona Mendozza is a predoc researcher working at the intersection of LLM interpretability, AI epistemology and philosophy of mind/language. He is currently assisting in research for an eye-tracking project at the center for language technology (Copenhagen University) and University of Turin (visiting). 

When: 18/02/2026, h 14:00
Where: Sala Conferenze, 3rd Floor

Categories
Meetings

Evaluation Under Variation: References, Annotators, and Languages

CCC Seminar by Silvia Casola, postdoc researcher at the MaiNLP group of the Ludwig Maximilian University of Munich and Munich Center for Machine Learning.

Abstract:
Automatic evaluation in NLP often assumes a single ground truth, such as a reference or a gold label. However, language is inherently variable: multiple outputs can be valid, annotators frequently disagree, and metric behaviours can differ across languages. In this talk, I will present three case studies showing how evaluation can fail and how it can be improved under such variation. Focusing on NLG, I will show that metrics can be highly sensitive to the choice of reference, leading to large changes in system rankings. I will then examine classification evaluation under annotator disagreement and present an approach for accounting for systematic disagreement. Finally, I will discuss recent work on steering multilingual neural metrics to improve their correlation with humans.
Starting from these failure modes, the talk shows how studying and modeling variation in references, annotations, and languages can improve the stability and reliability of automatic evaluation.

Short Bio:
I am a postdoc researcher at the MaiNLP group of the Ludwig Maximilian University of Munich and Munich Center for Machine Learning, supervised by Barbara Plank. Previously, I was a postdoc at the University of Turin, where I worked on perspective-aware NLP. I completed my PhD at the University of Padua and Fondazione Bruno Kessler, working on NLG. During my Ph.D., I visited UPF and interned with Spotify and Huawei Research. My research interests lie in NLP, focusing on Natural Language Generation and evaluation.

When: 31/03/2026, afternoon
Where: Sala Conferenze, 3rd Floor

Categories
Meetings

More with Less – Sustainable AI Approaches for Natural Language Processing  and Introduction to the  Brazilian National Institute for Responsible AI (INCT  TILDIAR)

CCC Seminary by Marcos Gonçalves, Full Professor of Computer Science at the Federal University of Minas Gerais (UFMG), who will present his work INCT TILDIAR: Data-Centric and Sustainable Paths Beyond the ‘Law of More’ in NLP.

Abstract:
This talk introduces the INCT TILDIAR, a national Brazilian research network dedicated to responsible and sustainable Artificial Intelligence, and presents the main research directions developed within the institute. We illustrate how our work challenges the prevailing “Law of More” in NLP by emphasizing data-centric and efficiency-driven approachesto AI.

As a concrete example, the talk briefly summarizes research on instance selection and data engineering, showing that substantial reductions in training data and energy consumption can be achieved while maintaining model effectiveness. The overarching message is that sustainable NLP is possible by rethinking how data is selected and used, rather than relying solely on ever-larger models.

Short bio:
Marcos André Gonçalves is a Full Professor of Computer Science at the Federal University of Minas Gerais (UFMG). He holds a PhD in Computer Science from Virginia Tech, with prior degrees from UFC and UNICAMP, and has completed postdoctoral research at UFMGand Politecnico di Torino. His research focuses on Information Retrieval, Machine Learning, and Natural Language Processing, with over 400 peer-reviewed publications, an h-index of 61, and more than 15,000 citations.

He has received multiple national awards, including CAPES Thesis Awards (2024 – Advisor; 2020 – Honorable Mention), and several Best Paper awards. Prof. Gonçalves has served as General Chair of ACM/IEEE JCDL 2018, is a Senior Program Committee member of leading conferences (SIGIR, ACL, CIKM, WSDM, RecSys, ECIR), and serves on the editorial boards of TACL and the Journal of the Brazilian Computer Society. He is also the Coordinator of the Brazilian National Institute for Responsible AI in NLP (INCT TILDIAR).


When: 02/02/2026, h 15:00
Where: Sala Riunioni, 1st Floor

Categories
Meetings

Don’t Classify, Rank: Retrieval, Fusion, and Label Semantics for XMTC and MCTC

CCC Seminary by Celso França, Ph.D. student at UFMG, who will present his work, xCoRetriev: A Retrieval-Centric Paradigm for Extreme and Multi-Class Text Classification.

Abstract:
We address Extreme Multi-Label Text Classification (XMTC) and Multi-Class Text Classification (MCTC) under a unified paradigm that reframes classification as a ranking and retrieval problem over large, noisy, and skewed label spaces. In this talk, we synthesize our recent SIGIR 2025 paper and our best paper of SBBD 2025 to demonstrate how retrieval-based formulations can jointly improve scalability, effectiveness, and label semantics across both XMTC and MCTC settings. Our core proposal is xCoRetriev, a dynamic two-stage retrieval and fusion pipeline designed to tackle the main challenges of label space volume, extreme skewness, and label quality by effectively combining dense and sparse representations. We further discuss recent attempts to enhance xCoRetriev’s effectiveness through Dimension Importance Estimation (DIMES) strategies and learned sparse representations trained via masked language modeling (MLM). While these approaches show promise in emphasizing discriminative signals and improving tail-label sensitivity, our analysis highlights their current limitations. Across multiple large-scale datasets, our results demonstrate consistent gains in propensity-scored metrics, improved robustness to noisy and weakly supervised label spaces through RAG-enhanced labels, and strong scalability at both training and inference time. Overall, this work advocates for a retrieval-centric view of large-scale text classification, bridging XMTC and MCTC through ranking, fusion, and importance-aware representations.

When: 02/02/2026, h 14:00
Where: Sala Riunioni, 1st Floor

Categories
Meetings

NLP for morphologically rich languages


CCC Seminar by Prof Gülşen Eryiğit, faculty of the Artificial Intelligence and Data Engineering Department at Istanbul Technical University, coordinator of ITU Natural Language Processing Group, and the director of ITU TÖMER (Turkish Teaching Application and Research Center).


Abstract
Natural language processing (NLP) has made remarkable progress in recent years, yet many of its most celebrated models and methods remain shaped by the structure and resources of Indo-European languages. This seminar examines the unique linguistic and computational challenges presented by agglutinative and morphologically rich languages such as Turkish, where high morphological variability, extensive affixation, and productive derivation complicate core NLP tasks including tokenization, morphology analysis, syntactic parsing, semantic representation, and downstream applications like information extraction and text understanding. The talk also offers a brief overview of the broader research agenda and key publications in this area.

Short Bio
Gülşen Eryiğit is a faculty of the Artificial Intelligence and Data Engineering Department at Istanbul Technical University, the coordinator of ITU Natural Language Processing Group, and the director of ITU TÖMER (Turkish Teaching Application and Research Center). She is the top-cited researcher in Turkey in NLP and listed in Stanford/Elsevier’s Top 2% Scientist Rankings career-long impact category. She received her master’s and PhD degrees from ITU Computer Engineering Department in 2002 and 2007. She has worked as a referee and author in many prestigious journals and conferences on NLP. She has worked as a coordinator or researcher in many scientific projects funded by EU, Tubitak, and the Ministry of Industry and Technology, and as a consultant in several industrial R&D projects funded by EU and Tubitak TEYDEB. She also acts as a project evaluator and observer for these funding agencies. She owns two issued patents. She is the person who realized the first software export from the ITU Technology transfer office. She is actively a management committee member in the European Cost Action UNIDIVE where she co-chairs the work package on “Multilingual and Cross-Lingual Language Technology”. She prioritizes the development of local technologies and has been actively involved in R&D projects on language technologies with various institutional firms in Turkey through university-industry collaboration.

When:  20/02/26 , h 11.00

Where:  Sala Conferenze – 3rd floor

Categories
Meetings

Cross-platform analysis and modeling of social media harmful content targeting children and adolescents

Kênia Carolina Gonçalves will present her doctoral research on the cross-platform analysis and modeling of harmful social media content targeting children and adolescents. The talk addresses the growing exposure of young users to potentially harmful online content and the specific vulnerabilities they face due to ongoing emotional, social, and cognitive development. It will provide a comprehensive characterization of content directed at children across major social media platforms, with a focus on identifying and understanding potential harms. The presentation will introduce a proposed taxonomy for classifying harmful content aimed at children and adolescents, grounded in empirical analyses of cross-platform datasets. Finally, it will explore the use of unsupervised machine learning techniques based on dataset similarity to support the automatic detection of harmful content, with the goal of enabling early alerts for immediate risks and identifying more subtle and insidious forms of harm beyond explicit manifestations.

Where: Sala Conferenze (3rd Floor)
When: 16/01/26 11:30

Categories
Meetings

HurtNet: a Multilingual Dictionary of Hurtful Words

Adel M. Wizani will present one of his recent collaborative works: HurtNet, a multilingual dictionary of hurtful terms, their senses, and examples of hurtful usage. The resource spans five languages: Arabic, French, Bulgarian, Greek, and Italian. The talk will outline the creation process for each language and the methods used for validation, focusing on the challenges of modeling hurtful language in Arabic (the presenter’s native language), particularly given its diglossic nature and semantic richness. It will also present results from a knowledge-injection zero-shot classification experiment, which demonstrate improved recall across languages, along with a qualitative analysis of model-generated explanations that reveal cross-linguistic patterns in the expression and detection of hurtful language.

Where: Sala Riunioni (1st floor)
When: 21/11/25 11:30

Categories
Meetings

Less than Meets the Eye: Representing Compounds in Large Language Models 

CCC Seminar Prof. Aline Villavicencio, Director of the Institute of Data Science and Artificial Intelligence, University of Exter, UK

Abstract

Large language models have been successfully used for capturing distinct (and very specific) word usages, and therefore could provide an attractive alternative for accurately determining meaning in language. However, these models still face a serious challenge when dealing with non-literal language, like that involved in Multiword Expressions (MWEs) such as idioms (make ends meet), light verb constructions (give a sigh), verb particle constructions (shake up) and noun compounds (loan shark). MWEs are an integral part of the mental lexicon of native speakers often used to express complex ideas in a simple and conventionalised way accepted by a given linguistic community. Although they may display a wealth of idiosyncrasies, from lexical, syntactic and semantic to statistical, that represents a real challenge for current LLMs, their accurate integration has the potential for improving the precision, naturalness and fluency of many tasks. In this talk, I will present an overview of how advances in LLMs have made an impact for the identification and modelling of idiomaticity and MWEs. I will concentrate on what models seem to incorporate of idiomaticity, as idiomatic interpretation may require knowledge that goes beyond what can be gathered from the individual words of an expression (e.g. “dark horse” as an unknown candidate who unexpectedly succeeds). I will also present an initiative to construct a multilingual idiomatic dataset.

Short Bio

Aline Villavicencio is the Director of the Institute of Data Science and Artificial Intelligence, University of Exeter, affiliated to the  Department of Computer Science, University of Sheffield (UK), is a member of ELLIS and has a Fellowship at the Alan Turing Institute. Before these, she held academic positions in the Institute of Informatics, Federal University of Rio Grande do Sul, Brazil (between 2005 and 2021) and in the School of Computer Science and Electronic Engineering, University of Essex, UK. She received her PhD from the University of Cambridge (UK) in 2001, and held postdoc positions at the University of Cambridge and University of Essex (UK). She was a Visiting Scholar at the Massachusetts Institute of Technology (USA, 2011-2012 and 2014-2015), at the École Normale Supé­rieure (France, 2014), an Erasmus-Mundus Visting Scholar at Saarland University (Germany in 2012/2013) and at the University of Bath (UK, 2006-2009).  She held a Research Fellowship from the  Brazilian National Council for Scientific and Technological Development (Brazil, 2009-2017). She is a member of the editorial board of Computational Linguistics, TACL and of JNLE. She is the General Chair of EACL 2026 and was the PC Co-Chair of ACL 2022, CoNLL 2019, Senior Area Chair for EMNLP 2025, ACL 2020 and ACL 2019 among others and General co-chair for the 2018 International Conference on Computational Processing of Portuguese. She was also a member of the NAACL board, SIGLEX board and of the program committees of various *ACL and AI conferences, and has co-chaired  several *ACL workshops on Cognitive Aspects of Computational Language Acquisition and on Multiword Expressions. Her research interests include lexical semantics, multilinguality, multiword expressions and cognitively motivated NLP, and has co-edited special issues and books dedicated to these topics.

When:  13/11/25 , h 11.00

Where:  Sala Conferenze – 3rd floor

Categories
Meetings

Bootstrapping UMRs from UD for Scalable Multilingual Annotation

CCC Seminar by Federica Gamba – PhD student in Computational Linguistics, Charles University, Czech Republic

Abstract

Uniform Meaning Representation (UMR) offers a cross-linguistically applicable framework for capturing sentence- and document-level semantics, but producing UMR annotations from scratch is a time-intensive process. In this talk, I will present an approach for bootstrapping UMR graphs by leveraging Universal Dependencies (UD), a richly annotated multilingual syntactic resource covering a wide range of language families. I will describe how structural correspondences between UD and UMR can be exploited to automatically derive partial UMR graphs from UD trees, providing annotators with an initial representation to refine rather than create from scratch. While UD is not inherently semantic, it encodes syntactic information that maps well onto UMR structures, allowing us to extract meaningful correspondences that simplify annotation. This method not only reduces annotation effort but also facilitates scalable UMR creation across typologically diverse languages, aligning with UMR’s cross-linguistic design goals.

When:  12/11/25 , h 9.00

Where:  Sala Conferenze – 3rd floor

Categories
Meetings

How Irrelevant Contextual Information Can Systematically Influence the Outputs of LLMs


Samuele D’Avenia will talk about one of his recent works, the abstract of which is reported below.

Several recent works have examined the generations produced by large language models (LLMs) on subjective topics such as political opinions and attitudinal questionnaires, with a particular interest in controlling these outputs to align with specific users or perspectives.

This work investigates how irrelevant contextual information can systematically influence the outputs of Large Language Models (LLMs) when generating opinions on subjective topics. Using the Political Compass Test as a case study, we analyze LLM-generated responses in an open-generation setting and conduct a detailed statistical analysis to quantify shifts in opinions caused by unrelated contextual cues. Our findings reveal that some of these elements can predictably bias model responses, further highlighting challenges in ensuring the robustness and reliability of LLMs when generating opinions on subjective topics

When:  4/07/2025, h 11.30

Where:  Sala conferenze – 3rd floor