ep9io commited on
Commit
85f5d47
1 Parent(s): d45914b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +56 -9
README.md CHANGED
@@ -1,22 +1,69 @@
1
  ---
2
  title: >-
3
- Employing Large Language Models to Identify Biased Language in Job
4
- Descriptions.
5
  emoji: 🌍
6
  colorFrom: gray
7
  colorTo: pink
8
  sdk: static
9
  pinned: false
10
  license: apache-2.0
11
- short_description: Advancing NLP for bias detection in job ads.
12
  ---
13
 
14
- Our research aims to advance the application of Natural Language Processing (NLP) techniques in the detection of biases in job advertisements, which can affect applicant diversity by embedding prejudices within text such as age and gender.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
 
16
- Our hypothesis is that LLMs, due to their reasoning skills, can de-bias the HR ecosystem. We aim to validate this through three research questions:
17
 
18
- 1) Can foundation LLMs faithfully identify implicit and subtle bias in job descriptions? Ideally, we hope to study the implicit information extraction capability of LLMs without being trained explicitly for the task. We plan to test various LLMs and scales (<= n billion-parameter models) to establish our hypothesis.
19
- 2) Which prompting method (zero-shot, few-shot, or chain-of-thought) most efficiently identifies bias in text? We aim to carry out comprehensive experiments on different prompting strategies already proposed in the literature. We also aim to propose different prompting templates covering the three settings and document our observations across the LLMs.
20
- 3) Does domain adaptation via fine-tuning foundational LLMs improve on prompt tuning? We aim to perform domain adaptation by fine-tuning the selected models on our datasets. We will seek to investigate if fine-tuning improves or lowers their performance under the same prompt settings.
 
 
 
 
 
 
 
 
21
 
22
- NOTE: Work in progress. The benchmark dataset and models aren't complete.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  title: >-
3
+ Large Language Models for Detecting Bias in Job Descriptions.
 
4
  emoji: 🌍
5
  colorFrom: gray
6
  colorTo: pink
7
  sdk: static
8
  pinned: false
9
  license: apache-2.0
10
+ short_description: LLms for Detecting Bias in Job Descriptions.
11
  ---
12
 
13
+ Abstract—This study explores the application of large language
14
+ (LLM) models for detecting implicit bias in job descriptions,
15
+ an important concern in human resources that shapes applicant
16
+ pools and influences employer perception. We compare different
17
+ LLM architectures—encoder, encoder-decoder, and decoder mod-
18
+ els—focusing on seven specific bias types. The research questions
19
+ address the capability of foundation LLMs to detect implicit
20
+ bias and the effectiveness of domain adaptation via fine-tuning
21
+ versus prompt tuning. Results indicate that fine-tuned models
22
+ outperform non-fine-tuned models in detecting biases, with Flan-
23
+ T5-XL emerging as the top performer, surpassing the zero-shot
24
+ prompting of GPT-4o model. A labelled dataset comprising gold,
25
+ silver, and bronze-standard data was created for this purpose
26
+ and open-sourced1 to advance the field and serve as a valuable
27
+ resource for future studies.
28
 
 
29
 
30
+ Short Introduction:
31
+ In human resources, bias affects both employers and em-
32
+ ployees in explicit and implicit forms. Explicit bias is
33
+ conscious and controllable, but can be illegal in employment
34
+ contexts. Implicit bias is subtle, unconscious, and harder to
35
+ address. Implicit bias in job descriptions is a major
36
+ concern as it shapes the applicant pool and influences appli-
37
+ cants’ decisions. Bias in the language of job descriptions can
38
+ affect how attractive a role appears to different individuals and
39
+ can impact employer perception. The challenge is to efficiently
40
+ identify and mitigate these biases.
41
 
42
+
43
+ The application of large language models (LLMs) for de-
44
+ tecting bias in job descriptions is promising but underexplored.
45
+ This study examines the effectiveness of various LLM archi-
46
+ tectures (encoder, encoder-decoder, decoder) less than 10 billion parameters in detecting implicit
47
+ bias.
48
+
49
+
50
+ We conceptualise the task of identifying implicit bias in
51
+ job descriptions as a multi-label classification problem, where
52
+ each job description is assigned a subset of labels from a
53
+ set of eight categories—age, disability, feminine, masculine,
54
+ general exclusionary, racial, sexuality, and neutral. This study
55
+ investigates two primary research questions:
56
+
57
+ 1) Can foundation LLMs accurately detect implicit bias in
58
+ job descriptions without specific task training? We evalu-
59
+ ate the performance of three topical decoder-only models
60
+ under four distinct prompt settings, assessing their ability
61
+ to extract relevant information from job descriptions and
62
+ identify implicit bias.
63
+
64
+ 2) Does domain adaptation via fine-tuning foundational
65
+ LLMs outperform prompt tuning for detecting implicit
66
+ bias in job descriptions? We fine-tune models with vary-
67
+ ing architectures as text-classifiers on task-specific data
68
+ and compare their performance to that of prompt-tuned
69
+ models.