ep9io commited on
Commit
32a5375
1 Parent(s): cdc0fe9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -43
README.md CHANGED
@@ -196,29 +196,42 @@ model-index:
196
  The model is a multi-label classifier designed to detect various types of bias within job descriptions.
197
 
198
  - **Developed by:** Tristan Everitt and Paul Ryan
199
- - **Funded by [optional]:** [More Information Needed]
200
- - **Shared by [optional]:** [More Information Needed]
201
- - **Model type:** [More Information Needed]
202
  - **Language(s) (NLP):** en
203
  - **License:** apache-2.0
204
  - **Finetuned from model [optional]:** google/flan-t5-xl
205
 
206
- ### Model Sources [optional]
207
 
208
- <!-- Provide the basic links for the model. -->
209
-
210
- - **Repository:** https://gitlab.computing.dcu.ie/everitt2/2024-mcm-everitt-ryan
211
- - **Paper [optional]:** [More Information Needed]
212
- - **Demo [optional]:** [More Information Needed]
213
 
214
  ## Uses
215
 
216
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
 
 
217
 
218
- ### Direct Use
219
 
220
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
 
 
221
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
222
 
223
  ```python
224
  from transformers import pipeline
@@ -237,37 +250,6 @@ The model is a multi-label classifier designed to detect various types of bias w
237
  {'label': 'racial', 'score': 0.014618005603551865},
238
  {'label': 'sexuality', 'score': 0.005568435415625572}
239
  ]]
240
-
241
-
242
- ### Downstream Use [optional]
243
-
244
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
245
-
246
- [More Information Needed]
247
-
248
- ### Out-of-Scope Use
249
-
250
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
251
-
252
- [More Information Needed]
253
-
254
- ## Bias, Risks, and Limitations
255
-
256
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
257
-
258
- [More Information Needed]
259
-
260
- ### Recommendations
261
-
262
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
263
-
264
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
265
-
266
- ## How to Get Started with the Model
267
-
268
- Use the code below to get started with the model.
269
-
270
- [More Information Needed]
271
 
272
  ## Training Details
273
 
 
196
  The model is a multi-label classifier designed to detect various types of bias within job descriptions.
197
 
198
  - **Developed by:** Tristan Everitt and Paul Ryan
199
+ - **Model type:** Encoder-Decoder
 
 
200
  - **Language(s) (NLP):** en
201
  - **License:** apache-2.0
202
  - **Finetuned from model [optional]:** google/flan-t5-xl
203
 
204
+ ### Model Sources
205
 
206
+ - **Repository:** https://github.com/2024-mcm-everitt-ryan
207
+ - **Paper:** In Progress
 
 
 
208
 
209
  ## Uses
210
 
211
+ The primary target audience for these models are researchers dedicated to identifying biased language in job descriptions.
212
+
213
+ ### Out-of-Scope Use
214
 
215
+ Due to the limitations inherent in large-scale language models, they should not be utilised in applications requiring factual or accurate outputs. These models do not distinguish between fact and fiction, and implicit biases are inherently subjective.
216
 
217
+ Moreover, as language models mirror the biases present in their training data, they should not be deployed in systems that directly interact with humans unless the deployers have first conducted a thorough analysis of relevant biases for the specific use case.
218
+
219
+ ## Bias, Risks, and Limitations
220
 
221
+ It is imperative that all users, both direct and downstream, are aware of the risks, biases, and limitations associated with this model. Important considerations include:
222
+
223
+ - Bias in Training Data: The model may inherit and perpetuate biases from the data it was trained on.
224
+ - Subjectivity of Bias: Bias detection is inherently subjective, and perceptions of bias can differ across contexts and users.
225
+ - Accuracy Concerns: The model’s outputs are not guaranteed to be true or accurate, making it unsuitable for applications that require reliable information.
226
+ - Human Interaction Risks: When incorporated into systems that interact with humans, the model’s biases may affect interactions and decision-making, potentially leading to unintended consequences.
227
+
228
+ It is crucial for users to conduct comprehensive evaluations and consider these factors when applying the model in any context.
229
+
230
+
231
+
232
+ ## How to Get Started with the Model
233
+
234
+ Use the code below to get started with the model.
235
 
236
  ```python
237
  from transformers import pipeline
 
250
  {'label': 'racial', 'score': 0.014618005603551865},
251
  {'label': 'sexuality', 'score': 0.005568435415625572}
252
  ]]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
253
 
254
  ## Training Details
255