Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
@@ -1,22 +1,69 @@
|
|
1 |
---
|
2 |
title: >-
|
3 |
-
|
4 |
-
Descriptions.
|
5 |
emoji: 🌍
|
6 |
colorFrom: gray
|
7 |
colorTo: pink
|
8 |
sdk: static
|
9 |
pinned: false
|
10 |
license: apache-2.0
|
11 |
-
short_description:
|
12 |
---
|
13 |
|
14 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
-
Our hypothesis is that LLMs, due to their reasoning skills, can de-bias the HR ecosystem. We aim to validate this through three research questions:
|
17 |
|
18 |
-
|
19 |
-
|
20 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
|
22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
title: >-
|
3 |
+
Large Language Models for Detecting Bias in Job Descriptions.
|
|
|
4 |
emoji: 🌍
|
5 |
colorFrom: gray
|
6 |
colorTo: pink
|
7 |
sdk: static
|
8 |
pinned: false
|
9 |
license: apache-2.0
|
10 |
+
short_description: LLms for Detecting Bias in Job Descriptions.
|
11 |
---
|
12 |
|
13 |
+
Abstract—This study explores the application of large language
|
14 |
+
(LLM) models for detecting implicit bias in job descriptions,
|
15 |
+
an important concern in human resources that shapes applicant
|
16 |
+
pools and influences employer perception. We compare different
|
17 |
+
LLM architectures—encoder, encoder-decoder, and decoder mod-
|
18 |
+
els—focusing on seven specific bias types. The research questions
|
19 |
+
address the capability of foundation LLMs to detect implicit
|
20 |
+
bias and the effectiveness of domain adaptation via fine-tuning
|
21 |
+
versus prompt tuning. Results indicate that fine-tuned models
|
22 |
+
outperform non-fine-tuned models in detecting biases, with Flan-
|
23 |
+
T5-XL emerging as the top performer, surpassing the zero-shot
|
24 |
+
prompting of GPT-4o model. A labelled dataset comprising gold,
|
25 |
+
silver, and bronze-standard data was created for this purpose
|
26 |
+
and open-sourced1 to advance the field and serve as a valuable
|
27 |
+
resource for future studies.
|
28 |
|
|
|
29 |
|
30 |
+
Short Introduction:
|
31 |
+
In human resources, bias affects both employers and em-
|
32 |
+
ployees in explicit and implicit forms. Explicit bias is
|
33 |
+
conscious and controllable, but can be illegal in employment
|
34 |
+
contexts. Implicit bias is subtle, unconscious, and harder to
|
35 |
+
address. Implicit bias in job descriptions is a major
|
36 |
+
concern as it shapes the applicant pool and influences appli-
|
37 |
+
cants’ decisions. Bias in the language of job descriptions can
|
38 |
+
affect how attractive a role appears to different individuals and
|
39 |
+
can impact employer perception. The challenge is to efficiently
|
40 |
+
identify and mitigate these biases.
|
41 |
|
42 |
+
|
43 |
+
The application of large language models (LLMs) for de-
|
44 |
+
tecting bias in job descriptions is promising but underexplored.
|
45 |
+
This study examines the effectiveness of various LLM archi-
|
46 |
+
tectures (encoder, encoder-decoder, decoder) less than 10 billion parameters in detecting implicit
|
47 |
+
bias.
|
48 |
+
|
49 |
+
|
50 |
+
We conceptualise the task of identifying implicit bias in
|
51 |
+
job descriptions as a multi-label classification problem, where
|
52 |
+
each job description is assigned a subset of labels from a
|
53 |
+
set of eight categories—age, disability, feminine, masculine,
|
54 |
+
general exclusionary, racial, sexuality, and neutral. This study
|
55 |
+
investigates two primary research questions:
|
56 |
+
|
57 |
+
1) Can foundation LLMs accurately detect implicit bias in
|
58 |
+
job descriptions without specific task training? We evalu-
|
59 |
+
ate the performance of three topical decoder-only models
|
60 |
+
under four distinct prompt settings, assessing their ability
|
61 |
+
to extract relevant information from job descriptions and
|
62 |
+
identify implicit bias.
|
63 |
+
|
64 |
+
2) Does domain adaptation via fine-tuning foundational
|
65 |
+
LLMs outperform prompt tuning for detecting implicit
|
66 |
+
bias in job descriptions? We fine-tune models with vary-
|
67 |
+
ing architectures as text-classifiers on task-specific data
|
68 |
+
and compare their performance to that of prompt-tuned
|
69 |
+
models.
|