Explainable Machine Learning

We are interested in studying method to explain relationships between inputs and outputs of black-box machine learning models, particularly in the context of challenging NLU tasks such as fact checking.

We are researching methods for explainable stance detection in the context of a DFF Sapere Aude Research Leader project, and explainable fact checking as part of an ERC Starting Grant project.

Moreover, we are investigating fair and accountable Natural Language Processing methods to understand what influences the employer images that organisations project in job ads, as part of a Carlsberg-funded project.

Publications

Modern Question Answering (QA) and Reasoning approaches based on Large Language Models (LLMs) commonly use prompting techniques, such …

Knowledge-intensive language understanding tasks require Language Models (LMs) to integrate relevant context, mitigating their inherent …

Uncovering latent values and opinions in large language models (LLMs) can help identify biases and mitigate potential harm. Recently, …

The large and ever-increasing amount of data available on the Internet coupled with the laborious task of manual claim and fact …

The emergence of tools based on large language models (LLMs), like OpenAI’s ChatGPT and Google’s Gemini, has garnered immense public …

How much meaning influences gender assignment across languages is an active area of research in modern linguistics and cognitive …

Explaining the decision-making process of machine learning models is crucial for ensuring their reliability and fairness. One popular …

We are exposed to much information trying to influence us, such as teaser messages, debates, politically framed news, and propaganda - …

Language Models (LMs) acquire parametric knowledge from their training process, embedding it within their weights. The increasing …

What can large language models learn? By definition, language models (LM) are distributions over strings. Therefore, an intuitive way …

Explainable AI methods facilitate the understanding of model behaviour, yet, small, imperceptible perturbations to inputs can vastly …

Human values play a vital role as an analytical tool in social sciences, enabling the study of diverse dimensions within society as a …

Recent studies of the emergent capabilities of transformer-based Natural Language Understanding (NLU) models have indicated that they …

NLP models are used in a variety of critical social computing tasks, such as detecting sexist, racist, or otherwise hateful content. …

Reasoning over spans of tokens from different parts of the input is essential for natural language understanding (NLU) tasks such as …

Answering complex queries on incomplete knowledge graphs is a challenging task where a model needs to answer complex logical queries in …

Explanations of neural models aim to reveal a model’s decision-making process for its predictions. However, recent work shows …

Language embeds information about social, cultural, and political values people hold. Prior work has explored social and potentially …

The success of pre-trained contextualized representations has prompted researchers to analyze them for the presence of linguistic …

Fact-checking systems have become important tools to verify fake and misguiding news. These systems become more trustworthy when …

There have been many efforts to try to understand what grammatical knowledge (e.g., ability to understand the part of speech of a …

Two of the most fundamental challenges in Natural Language Understanding (NLU) at present are: (a) how to establish whether deep …

With the substantial rise in the amount of mis- and disinformation online, fact checking has become an important task to automate. This …

Counterfactually Augmented Data (CAD) aims to improve out-of-domain generalizability, an indicator of model robustness. The improvement …

The success of multilingual pre-trained models is underpinned by their ability to learn representations shared by multiple languages …

Automating the fact checking (FC) process relies on information obtained from external sources. In this work, we posit that it is …

Explanations shed light on a machine learning model’s rationales and can aid in identifying deficiencies in its reasoning …

Medical artificial intelligence (AI) systems have been remarkably successful, even outperforming human performance at certain tasks. …

As NLP models are increasingly deployed in socially situated settings such as online abusive content detection, ensuring these models …

Sparse attention has been claimed to increase model interpretability under the assumption that it highlights influential inputs. Yet …

The past decade has seen a substantial rise in the amount of mis- and disinformation online, from targeted disinformation campaigns to …

Recent developments in machine learning have introduced models that approach human performance at the cost of increased architectural …

Adversarial attacks reveal important vulnerabilities and flaws of trained models. One potent type of attack are universal adversarial …

While state-of-the-art NLP explainability (XAI) methods focus on supervised, per-instance end or diagnostic probing task evaluation[4, …

This paper provides the first study of how fact checking explanations can be generated automatically based on available claim context, …

Talks

PhD fellowship on Interpretable Machine Learning available. The successfull candidate will be supervised by Pepa Atanasova and Isabelle …

Starting in September 2024, Pepa is taking on a new role as Tenure-Track Assistant Professor in the NLP Section at the University of …

We are recruiting professional fact checkers to take part in an interview and/or a survey about their experiences of fact checking and …

A PhD and two postdoc positions on natural language understanding are available. The positions are funded by the Pioneer Centre for AI.

A PhD position on explainable natural language understanding is available in CopeNLU. The positions is funded by the ERC Starting Grant …

On 1 September 2023, the ERC Starting Grant project ExplainYourself on ‘Explainable and Robust Automatic Fact Checking’ is …

PhD and postdoctoral fellowships on explainable fact checking are available in CopeNLU. The positions are funded by the ERC Starting …

On 1 September 2021, the DFF Sapere Aude project EXPANSE on ‘Learning to Explain Attitudes on Social Media’ is kicking off, …