Description du projet
Atténuer les modèles biaisés fondés sur l’AA sur le marché du travail
Il ressort d’une étude récente de Sage que 24 % des entreprises utilisent le traitement du langage naturel (TLN) dans les applications de gestion des ressources humaines (GRH). Là où le bât blesse, c’est que le TLN se fonde sur des modèles biaisés. Cette utilisation du TLN dans le contexte du marché de l’emploi pourrait conduire à des décisions biaisées qui compromettent les objectifs du socle européen des droits sociaux et des objectifs de développement durable (ODD) des Nations unies (ONU). Le projet BIAS, financé par l’UE, étudiera et atténuera les biais de diversité de l’IA sur le marché du travail. Le projet développera une preuve de concept pour une technologie innovante reposant sur le TLN et le raisonnement à partir de cas (RàPC) destinée à être utilisée dans un cas de recrutement RH.
Objectif
Artificial Intelligence (AI) is increasingly used in the employment sector to manage and control individual workers. One type of AI is Natural Language Processing (NLP) based tools that can analyze text to make inferences or decisions. A recent Sage study found that 24% of companies used AI for hiring purposes. In an employment context, this can involve analyzing text created by an employee or recruitment candidate in order to assist management in deciding to invite a candidate for an interview, to training and employee engagement, or to monitor for infractions that could lead to disciplinary proceedings. However, the models that NLP-based systems are based on are biased. Additionally, it has been shown that bias in an underlying AI model is reproduced in applications based on that model). This can lead to biased decisions that run contrary to the goals of the European Pillar of Social Rights in relationship to work and employment, specifically Pillar 2 (Gender Equality), Pillar 3 (Equal Opportunity), Pillar 5 (Secure and Adaptable Employment) and the United Nations’ (UN) Sustainable Development Goals (SDGs), specifically SDG 5 (Gender Equality), SDG 8 (Decent Work and Economic Growth). It is therefore necessary to identify and mitigate biases that occur in applications used in a Human Resources Management (HRM) context. Addressing such concerns in an employment context is especially relevant, as most existing European studies on employment discrimination have indeed found that discrimination exists, both when considering individual diversity criteria and multiple criteria in intersectional analyses. In order to investigate and mitigate these biases, we apply this “BIAS”-project, for mitigating diversity biases of AI in the labor market. The chief technical objective of BIAS is the development of a proof-of-concept for an innovative technology based on Natural Language Processing (NLP) and Case Based Reasoning (CBR) for use in an HR recruitment use case.
Champ scientifique
- natural sciencescomputer and information sciencesartificial intelligence
- natural sciencescomputer and information sciencesdata sciencenatural language processing
- social sciencessociologysocial issuessocial inequalities
- social sciencessociologygender studiesgender equality
- social scienceseconomics and businessbusiness and managementemployment
Mots‑clés
Programme(s)
Régime de financement
HORIZON-RIA - HORIZON Research and Innovation ActionsCoordinateur
7491 Trondheim
Norvège