Workshop Schedule

Monday, 1st of July

08:00 Registration & Badge Pickup

09:00 Welcome Address

09:30 Keynote Talk: Prof. Bettina Berendt, TU Berlin, Weizenbaum Institute, and KU Leuven

De-biased, diverse, divisive - On ethical perspectives regarding the de-biasing of GenAI and their actionability

AI tech companies cannot seem to get it right. After years of evidence-based criticism of biases in AI, in particular decision models, LLMs and other generative AI, after years of research and toolbox provision for de-biasing, many companies have implemented such safeguards into their services. However, ridicule and protests have recently erupted when users discovered generated images that were “(overly?) diversified” with respect to gender and ethnicity and answers to ethical questions that were “(overly?) balanced” with regard to moral stances. Is this seeming contradiction just a backlash, or does it point to deeper issues? In this talk, I will analyse instances of recent discourse on too little or too much “diversification” of (Gen)AI and relate this to methodological criticism of “de-biasing”. A second aim is to contribute to the broadening and deepening of answers that computer science and engineering can and should give to enhance fairness and justice.

10:30 Coffee & Posters

11:00 Lightning Round 1

Room: Linke Aula

Algorithmic Fairness Over Time: Advances & Prospects

Miriam Rateike

Can generative AI-based data balancing mitigate unfairness issues in Machine Learning?

Benoît Ronval, Siegfried Nijssen and Ludwig Bothmann

An analysis of Facebook's privacy practices through the lens of compliance, integrity, and excellence

Marco Tulio Daza Ramirez

Governing online platforms after the Digital Services Act: an analysis of the Commission decision on initiating proceedings against X

Matteo Fabbri

Algorithmic Fairness in Geo-intelligence Workflows through Causality

Brian Masinde, Caroline Gevaert, Michael Nagenborg, Marc van den Homberg and Jaap Zevenbergen

Threshold Recourse for Dynamic Allocations 

Meirav Segal, Anne-Marie George and Christos Dimitrakakis

Categorizing algorithmic recommendations: a matter of system, agent, patient

Matteo Fabbri and Jesus Salgado

Exploring Fusion Techniques in Multimodal AI-Based Recruitment: Insights from FairCVdb

Swati Swati, Arjun Roy and Eirini Ntoutsi

Social Assessment and Cultural Resistance: The Public Distribution System in Tamil Nadu, India

Sumathi Rajesh

12:00 Lunch

13:00 Keynote Talk: Prof. Dr. Virginia Dignum, Umeå University

Beyond the AI hype: Balancing Innovation and Social Responsibility

AI can extend human capabilities but requires addressing challenges in education, jobs, and biases. Taking a responsible approach involves understanding AI's nature, design choices, societal role, and ethical considerations. Recent AI developments, including foundational models, transformer models, generative models, and large language models (LLMs), raise questions about whether they are changing the paradigm of AI, and about the responsibility of those that are developing and deploying AI systems. In all these developments, is vital to understand that AI is not an autonomous entity but rather dependent on human responsibility and decision-making. 

In this talk, I will further discuss the need for a responsible approach to AI that emphasize trust, cooperation, and the common good. Taking responsibility involves regulation, governance, and awareness. Ethics and dilemmas are ongoing considerations, but require understanding that trade-offs must be made and that decision processes are always contextual. Taking responsibility requires designing AI systems with values in mind, implementing regulations, governance, monitoring, agreements, and norms. Rather than viewing regulation as a constraint, it should be seen as a stepping stone for innovation, ensuring public acceptance, driving transformation, and promoting business differentiation. Responsible Artificial Intelligence (AI) is not an option but the only possible way to go forward in AI.

Virginia Dignum is Professor of Responsible Artificial Intelligence at Umeå University, Sweden where she leads the AI Policy Lab. She is also senior advisor on AI policy to the Wallenberg Foundations. She has a PHD in Artificial Intelligence from Utrecht University in 2004, is member of the Royal Swedish Academy of Engineering Sciences (IVA), and Fellow of the European Artificial Intelligence Association (EURAI). She is a member of the United Nations Advisory Body on AI, the Global Partnership on AI (GPAI), UNESCO’s expert group on the implementation of AI recommendations, OECD’s Expert group on AI, founder of ALLAI, the Dutch AI Alliance, and co-chair of the WEF’s Global Future Council on AI. She was a member of EU’s High Level Expert Group on Artificial Intelligence and leader of UNICEF's guidance for AI and children. Her new book “The AI Paradox” is planned for publication in late 2024.

14:00 Coffee & Posters

14:30 Lightning Round 2

Room: Linke Aula

Implications of the AI Act for Non-Discrimination Law and Algorithmic Fairness 

Luca Deck, Jan-Laurin Müller, Conradin Braun, Domenique Zipperling and Kühl Niklas

Policy Advice and Best Practices on Bias and Fairness in Artificial Intelligence 

Jose M. Alvarez, Alejandra Bringas-Colmenarejo, Alaa Elobaid, Simone Fabrizzi, Miriam Fahimi, Antonio Ferrara, Siamak Ghodsi, Carlos Mougan, Ioanna Papageorgiou, Paula Reyero, Russo Mayra, Kristen M. Scott, Laura State, Xuan Zhao and Salvatore Ruggieri

Watching the Watchers: A Comparative Fairness Audit of Cloud-based Content Moderation Services 

David Hartmann, Amin Oueslati and Dimitri Staufer

Enhancing Fairness through Time-Aware Recourse: A Pathway to Realistic Algorithmic Recommendations 

Isacco Beretta, Martina Cinquini and Isabel Valera

Beyond Silos: An Interdisciplinary Analysis of Intersectional Discrimination from an EU Perspective 

Stephan Wolters

Risk Scores as Statistical Fatalism  

Sebastian Zezulka and Konstantin Genin

Eliciting Discrimination Risks in Algorithmic Systems: Taxonomies and Recommendations

Marta Marchiori Manerba

Risky Complaints: Unpacking Recent Trends in Risk Assessment Across Global Supply Chains 

Gabriel Grill

On Prediction-Modelers and Decision-Makers: Why Fairness Requires More Than a Fair Prediction Model 

Teresa Scantamburlo, Joachim Baumann and Christoph Heitz 

15:30 Coffee & Posters

16:30 In-Depth Session 1

Room: Linke Aula

Mapping Policymakers' and Laypeople's Perceptions of genAI and FPT in the Context of the EU AI Act 

Chiara Ullstein, Michel Hohendanner and Jens Grossklags 

Fair Balancing? Evaluating LLM-based Privacy Policy Ethics Assessments 

Vincent Freiberger and Erik Buchmann 

Room: Atrium Maximum

Unfairness in AI Anti-Corruption Tools: Main Drivers and Consequences 

Fernanda Odilla 

Mapping the Potential of Explainable Artificial Intelligence (XAI) for Fairness Along the AI Lifecycle 

Luca Deck, Astrid Schoemäcker, Timo Speith, Jakob Schöffer, Lena Kästner and Niklas Kühl 

17:30 Poster Session

Room: Atrium Maximum

Papers from Lightning Rounds 1&2, Papers from In-Depth Session 1

Various Authors

19:00 Social Event @Salon 3SEIN

Salon 3Sein, Große Bleiche 60-62, 55116 Mainz

Tuesday, 2nd of July

08:00 Registration & Badge Pickup

09:30 Keynote Talk: Prof. Seth Lazar, Australian National University (online)

10:30 Coffee & Posters

11:00 Lightning Round 3

Room: Linke Aula

A Fair Selective Classifier to Put Humans in the Loop

Daphne Lenders, Andrea Pugnana, Roberto Pellungrini, Toon Calders, Fosca Giannotti and Dino Pedreschi

Fairness Beyond Binary Decisions: A Case Study on German Credit

Deborah Dormah Kanubala, Isabel Valera and Kavya Gupta

Challenging "Subgroup Fairness": Towards Intersectional Algorithmic Fairness Based on Personas

Marie Decker, Laila Wegner and Carmen Leicht-Scholten

Deciding the Future of Refugees: Rolling the Dice or Algorithmic Location Assignment?

Clara Strasser Ceballos and Christoph Kern

How to be fair? A discussion and future perspectives

Marco Favier and Toon Calders

Enabling users’ control on recommender systems for short videos: a design proposal for the implementation of the requirements of the Digital Services Act

Matteo Fabbri, Jingyi Jia and Wolfgang Woerndl

Algorithmic Fairness in Clinical Natural Language Processing: Challenges and Opportunities

Daniel Anadria, Anastasia Giachanou, Jacqueline Kernahan, Roel Dobbe and Daniel Oberski

Trust in Fair Algorithms: Pilot Experiment

Mattia Cerrato, Marius Köppel, Kiara Stempel, Alesia Vallenas Coronel and Niklas Witzig

Identifying Gender Stereotypes and Biases in Automated Translation from English to Italian using Similarity Networks

Fatemeh Mohammadi, Marta Anamaria Tamborini, Paolo Ceravolo, Costanza Nardocci and Samira Maghool

12:00 Lunch

13:00 Interactive Sessions 1

Room: Linke Aula

Building Bridges from and Beyond the EU Artificial Intelligence Act. Regulating AI-based Discrimination in the European Scenario

Marilisa D'Amico, Ernesto Damiani, Costanza Nardocci, Paolo Ceravolo, Samira Maghool, Marta Annamaria Tamborini, Paolo Gambatesa and Fatemeh Mohammadi

The panel aims to discuss the implications following the approval of the EU Artificial Intelligence Act also in light of the additional initiatives ongoing in the European and global scenario (e.g. Council of Europe, UNESCO, United Nations) from the perspective of their adequacy to ensure the fairness of algorithms and to tackle discriminations resulting from the massive resort to AI technologies. By bringing together constitutional law and computer science expertise, the discussion has a twofold aim. On the one hand, the panel aims to illustrate the criticisms that underlie the risks of AI systems from a constitutional and human rights perspective, supported by the analysis of computer science; on the other hand, it aims to explore innovative strategies to promote an inclusive use of AI technologies by designers and implementers in order to enhance their positive potential.

Room: Atrium Maximum

Moral Exercises for Human Oversight of AI Systems 

Teresa Scantamburlo and Silvia Crafa

The interactive session addresses the challenge of ethical reflection in AI research and practice. Human judgement is crucial in balancing accuracy and fairness trade-offs and overseeing AI system behaviour. To support ethical reflection and judgement we propose the experience of moral exercises: structured activities aimed at engaging AI actors in realistic ethical problems involving the development or use of an AI system. Participants undergo guided exercises involving scenario analysis, individual and group work, highlighting consensus and divergences on the problem at stake. The initiative will promote the development of moral exercises and help refine the methodology for AI ethics education.

14:30 Coffee & Posters

15:30 Interactive Sessions 2

Room: Linke Aula

How the Digital Services Act can enable researcher access to data of the largest online platforms and search engines - interactive session with the European Commission 

EC Joint Research Centre

The Digital Services Act (DSA) entered into full force in February, and aims to create a safer and more trustworthy digital space, where the fundamental rights of all users are protected.  As part of the DSA’s transparency obligations, Article 40 of the DSA establishes the obligation of Very

Large Online Platforms (VLOPs) and Very Large Online Search Engines (VLOSEs) to provide researchers with access to data for the purposes of conducting research that

contributes to the detection, understanding and mitigation of systemic risks in the European Union, such as discrimination and the spread of disinformation.    

In particular, Article 40(12) of the DSA obliges providers of VLOPs and VLOSEs to give researchers access to data that is publicly available in their interfaces. In addition, Article 40(4) of the DSA establishes a data access mechanism through which researchers who undergo a vetting procedure can obtain access to non-public data for the study of systemic risks in the European Union.  

In this workshop, the Commission will present the data access mechanism for vetted researchers and participants will have the opportunity to provide feedback on a detailed proposal for its procedural, technical and operational elements, which is currently being prepared in the form of a delegated act. Participants will get the chance to explore how data access could benefit their research, which challenges they foresee and how these may be overcome. Participants will also hear about how the DSA protects researcher access to publicly available data, and will be able to provide feedback on the data access tools and procedures made available by VLOPs and VLOSEs to this end so far. 

Room: Atrium Maximum

Fairness, or not Fairness, That is the Question. Rethinking Virtual Assistants' Responses From an Ethical Perspective 

Giulia Teverini, Joy Ciliani and Alessia Nicoletta Marino

Abstract: The workshop aims to delve into the intricate aspects of ethics applied to human-computer interaction. Participants will engage in practical activities in which they will analyze conversations with virtual assistants, distinguishing between appropriate and inappropriate interactions. In this context, the ethical principles proposed by the European Commission will be used as guidelines to ensure favorable conditions for the development of trustworthy AI-based systems. Immediately after participants will be asked to propose scenario-based solutions to tackle ethical issues which have emerged in the previous stage. At the end of the workshop, each group will show the result to the audience. The collaborative nature of the workshop aims at fostering critical thinking, practical understanding, and collective reasoning.

17:00 Poster Session

Room: Atrium Maximum

Papers from Lightning Round 3

Various Authors

What's Distributive Justice Got to Do with It? Rethinking Algorithmic Fairness from the Perspective of Approximate Justice

Corinna Hertweck, Christoph Heitz and Michele Loi

Reranking individuals: the effect of fair classification within-groups

Sofie Goethals and Toon Calders

More than Minorities and Majorities: Understanding Multilateral Bias in Language Generation

Jiaxu Zhao, Zijing Shi, Yitong Li, Yulong Pei, Ling Chen, Meng Fang and Mykola Pechenizkiy

An epistemic-based decision-making framework for modeling fairness in the spread of technologies for sustainability

Camilla Quaresmini, Eugenia Villa, Valentina Breschi, Viola Schiaffonati and Mara Tanelli

Scaling Up Causal Algorithmic Recourse with CARMA

Ayan Majumdar and Isabel Valera

What Is Fairness? On the Role of Protected Attributes and Fictitious Worlds
Ludwig Bothmann, Kristina Peters and Bernd Bischl

Can generative AI-based data balancing mitigate unfairness issues in Machine Learning?
Benoît Ronval, Siegfried Nijssen and Ludwig Bothmann

A Neuro-Symbolic Approach to Counterfactual Fairness
Xenia Heilmann, Chiara Manganini, Mattia Cerrato and Vaishak Belle

Towards Fair Co-clustering
Federico Peiretti and Ruggero G. Pensa

Integrating Global Justice into AI Fairness Criteria: A Novel Approach to Environmentally Sustainable and Equitable Fleet Management
Ahmet Bilal Aytekin, Batuhan Çınar, Ahu Ece Hartavi Karci

19:00 Social Event @Schlossbiergarten

Wednesday, 3rd of July

08:00 Registration & Badge Pickup

09:00 Keynote Talk: Prof. Isabel Valera, University of Saarland

"Society-centered AI: An Integrative Perspective on Algorithmic Fairness"

Abstract:  In this talk, I will share my never-ending learning journey on algorithmic fairness. I will give an overview of fairness in algorithmic decision making, reviewing the progress and wrong assumptions made along the way, which have led to new and fascinating research questions. Most of these questions remain open to this day, and become even more challenging in the era of generative AI. Thus, this talk will provide only few answers but many open challenges to motivate the need for a paradigm shift from owner-centered to society-centered AI. With society-centered AI, I aim to bring the values, goals, and needs of all relevant stakeholders into AI development as first-class citizens to ensure that these new technologies are at the service of society.

10:00 Coffee Break

10:30 Interactive Sessions 3

Room: Linke Aula

Artificial Intelligence for Assessment in the Context of Asylum Procedures – Lessons Learned and Reflections Upon Fairness-Related Challenges via Participatory Methods and Science-Policy Dialogue 


Room: Atrium Maximum

Striving for Equity: Navigating Algorithmic Fairness for AI in the Workplace 

Thea Radüntz, Martin Brenzke and Dominik Köhler 

12:00 Lunch

13:00 Lightning Round 4

Room: Linke Aula

Building job seekers’ profiles: can LLM’s level the playing field?

Susana Lavado and Leid Zejnilovic

FAIRRET: A flexible PyTorch library for fairness

Marybeth Defrance, Maarten Buyl and Tijl De Bie

Unveiling the Blindspots: Examining Availability and Usage of Protected Attributes in Fairness Datasets

Jan Simson, Alessandro Fabris and Christoph Kern

Beyond Distributions: A Systematic Review on Relational Algorithmic Justice

Laila Wegner, Marie Decker and Carmen Leicht-Scholten

Pricing Risk: Analysis of Irish Car Insurance Premiums

Adrian Byrne

20% Increase in Fairness for Black Applicants: A Critical Examination of Fairness Measurements Offered by Startups

Corinna Hertweck and Maya Guido

Integrating Fairness in AI Development: Technical Insights from the fAIr by design Framework

Mira Reisinger and Rania Wazir

Algorithmic Collective Action in Recommender Systems: Promoting Songs by Reordering Playlists

Joachim Baumann and Celestine Mendler-Dünner

Fairness in Graph-Theoretical Optimization Problems

Christopher Hojny, Frits Spieksma and Sten Wessel

It is not about Bias but Discrimination

Chaewon Yun, Claudia Wagner and Jan-Christoph Heilinger

Exploration of potential new benchmark for fairness evaluation in Europe

Magali Legast, Lisa Koutsoviti Koumeri, Yasaman Yousefi and Axel Legay

14:15 Coffee & Posters

15:00 In-Depth Session 2

Room: Linke Aula

Measurement Modeling of Predictors and Outcomes in Algorithmic Fairness 

Elisabeth Kraus and Christoph Kern 

Proxy Fairness under the GDPR and the AI Act: A Perspective of Sensitivity and Necessity 

Ioanna Papageorgiou 

Room: Atrium Maximum

When the Ideal Does Not Compute: Nonideal Theory and Fairness in Machine Learning 

Otto Sahlgren

Inherent Limitations of AI Fairness 

Maarten Buyl and Tijl De Bie 

16:00 Poster Session

Room: Atrium Maximum

Papers from Lightning Round 4, Papers from In-Depth Session 2

Various Authors

Exploration of potential new benchmark for fairness evaluation in Europe

Magali Legast, Lisa Koutsoviti Koumeri, Yasaman Yousefi and Axel Legay

Benchmarking Audio DeepFake Detection

Maria-Isavella Manolaki-Sempagios and Mykola Pechenizkiy

Input-debias: A Post-Processing Technique for Bias Mitigation in Contextualized Models

Nick Wils, Ewoenam Kwaku Tokpo and Toon Calders

FairFlow: An Automated Approach to Model-based Counterfactual Data Augmentation For NLP

Ewoenam Kwaku Tokpo and Toon Calders

Enhancing Model Fairness and Accuracy with Similarity Networks: A Methodological Approach

Samira Maghool and Paolo Ceravolo

17:00 Closing