Periodic Reporting for period 1 - NoBIAS (Artificial Intelligence without Bias)
Período documentado: 2020-01-01 hasta 2021-12-31
At each step in this process biases may arise, which need to be accounted for and countered in order to produce business benefits while addressing related legal and ethical concerns. In particular, the three core challenges are: (C1) Data can be biased; (C2) Algorithms can be biased; (C3) Results can be biased.
Research, development and training of early-stage researchers (ESRs) in NoBIAS is organized around the AI-based decision making pipeline and the identified C1-C3 core challenges and the corresponding O1-O3 objectives (as listed next) to ensure that crucial skills for AI-based decision making in industry and society are acquired and are well-aligned with business value creation.
Objective 1 [O1]: Understanding bias in data. The quality of the data provided as input to AI-decision making processes has a strong influence on the results. Understanding why and how bias is manifested in data is of paramount importance.
Objective 2 [O2]: Mitigating bias in algorithms. To account for bias in AI, we can improve the bias-related quality of the data or we can introduce extra constraints/costs in the utility measure of the model to “enforce” fairness. The former approach is independent of the algorithm whereas the latter depends on the algorithm per se. In the context of NoBIAS, we aim to tackle both model-independent and model-dependent challenges as well as connecting them with legal issues and contexts.
Objective 3 [O3]: Accounting for bias in results. Results of AI-decision making systems might be biased, even if the data has been corrected for bias and even if the algorithms have been modified to account for bias. Moreover, new sources of biases are introduced by the interpretation of the results and application context, when continuous model outputs are converted into binary decisions or when concept-drift arises over time. Therefore, an important challenge is to understand the models extracted from data and how certain decisions are taken and evolve over time.
Each ESR started out exploring the problem from a specific direction and each acquired a much deeper understanding of the nature of the manifestation of bias in data within their individual area. The important but common questions that O1 ESRs have been working on are:
(i) What is bias?
(ii) How is bias created?
(iii) How can we detect bias?
To answer these questions, long ethnographic fieldwork has been conducted by ESR Miriam Fahimi, suitable technical frameworks have been identified/ developed, methodological innovations and testing are currently underway for various use cases by ESRs Kristen Marie Scott, Jose Manuel Alvarez, Simone Fabbrizzi and Mayra Russo.
[O2]: Mitigating bias in algorithms.
O2 ESRs focus on the following multi-disciplinary research directions.
(i) Development of model-independent approaches for mitigating bias at the data level
(ii) Development of model-dependent approaches for mitigating bias at the model/algorithm and at the output level
(iii) Reconciling bias mitigation approaches with legal norms and legal theory
Legal exploration into the lawfulness of processing personal data for pre-processing debiasing purposes is about to be completed by ESR Ioanna Papageorgiou. Prior works on bias mitigation in different contexts have been surveyed by ESR Alaa Elobaid, and new techniques for NoBIAS’s ranking and classification use cases are currently being tested by ESRs Antonio Ferrara and Paula Reyero.
[O3]: Accounting for bias in results.
All ESRs have conducted extensive overviews of the literature in a rather large spectrum of multidisciplinary topics. Specifically the ESRs have looked into the following aspects;
(i) an ethical and legal perspective on accountability,
(ii) the field of eXplainable AI (XAI),
(iii) the issue of monitoring time-evolving AI models and their biases.
While a legal framework which encompasses the rights to information and an explanation is currently under development by ESR Alejandra Bringas Colmenarejo, the technical aspects of O3 are being explored through various ongoing use cases and experimental analyses on explainability and temporal bias mitigation by ESRs Xuan Zhao, Laura State, Ioannis Koulierakis, Carlos Mougan Navarro, and Seyed Siamak Ghodsi.
Other achievements of the network as of Dec’21 are listed next. The network has successfully organized various training programs including the onboarding week, a summer school, and the European AI regulation week, NoBIAS monthly colloquium talks. The ESRs have already published 7 research papers and attended 29 events including conferences, workshops, and panel discussions thereby contributing to dissemination and communication goals of NoBIAS. The network has garnered a lot of attention on its social media handle (#followers=267) and also on the NoBIAS website (#visitors=4974).
Related ETN projects include analysis of Big Data (ETN Longpop, ID: 676060), privacy and usability (ETN Privacy.us 675730), risk analysis (ETN BigDataFinance, 675044), Big Data management (ETN BigStorage, 642963), or question answering from Big Data (ETN WDAqua, 642795). However, the focus in these projects lies on generating value from big data by new methods, and with some lesser focus on privacy. The issues arising with regard to bias, such as fairness, legality, or discrimination are not addressed.