tsantosh7 commited on
Commit
0c513eb
1 Parent(s): c75f240

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -24
README.md CHANGED
@@ -1,41 +1,47 @@
 
 
 
 
1
 
 
2
 
3
- # mdeberta-wl-base-es
4
 
5
- This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on Bailii dataset. This dataset is not available publicly.
6
 
7
- ## Model description
8
 
9
- More information needed
10
 
11
- ## Intended uses & limitations
12
 
13
- More information needed
14
 
15
- ## Training and evaluation data
16
 
17
- More information needed
18
 
19
- ## Training procedure
 
20
 
21
- ### Training hyperparameters
22
 
23
- The following hyperparameters were used during training:
24
- - learning_rate: 5e-05
25
- - train_batch_size: 16
26
- - eval_batch_size: 16
27
- - seed: 42
28
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
29
- - lr_scheduler_type: linear
30
- - num_epochs: 3.0
31
 
32
- ### Training results
33
 
 
34
 
 
35
 
36
- ### Framework versions
 
37
 
38
- - Transformers 4.21.0.dev0
39
- - Pytorch 1.11.0+cu113
40
- - Datasets 2.3.3.dev0
41
- - Tokenizers 0.12.1
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+ # Pre-trained Language Model for England and Wales Court of Appeal (Criminal Division) Decisions
5
 
6
+ ## Introduction
7
 
8
+ The research for understanding the bias in criminal court decisions need the support of natural language processing tools.
9
 
10
+ The pre-trained language model has greatly improved the accuracy of text mining in general texts. At present, there is an urgent need for a pre-trained language model specifically for the automatic processing of court decision texts.
11
 
12
+ We used the text from the [Bailii website](https://www.bailii.org/ew/cases/EWCA/Crim/) as the training set. Based on the deep language model framework of RoBERTa, we constructed bailii-roberta pre-training language model by [transformers/run_mlm.py](https://github.com/huggingface/transformers/blob/main/examples/pytorch/language-modeling/run_mlm.py) and [transformers/mlm_wwm](https://github.com/huggingface/transformers/tree/main/examples/research_projects/mlm_wwm).
13
 
 
14
 
15
+ ## How to use
16
 
17
+ ### Huggingface Transformers
18
 
19
+ The `from_pretrained` method based on [Huggingface Transformers](https://github.com/huggingface/transformers) can directly obtain bailii-roberta model online.
20
 
 
21
 
22
+ ```python
23
+ from transformers import AutoTokenizer, AutoModel
24
 
25
+ tokenizer = AutoTokenizer.from_pretrained("tsantosh7/bailii-roberta")
26
 
27
+ model = AutoModel.from_pretrained("tsantosh7/bailii-roberta")
28
+ ```
 
 
 
 
 
 
29
 
30
+ ### Download Models
31
 
32
+ - The version of the model we provide is `PyTorch`.
33
 
34
+ ### From Huggingface
35
 
36
+ - Download directly through Huggingface's official website.
37
+ - [tsantosh7/bailii-roberta](https://huggingface.co/tsantosh7/Bailii-Roberta/)
38
 
39
+ ## Disclaimer
40
+
41
+ - The experimental results presented in the report only show the performance under a specific data set and hyperparameter combination, and cannot represent the essence of each model. The experimental results may change due to the random number of seeds and computing equipment.
42
+ - **Users can use the model arbitrarily within the scope of the license, but we are not responsible for the direct or indirect losses caused by using the content of the project.**
43
+
44
+
45
+ ## Acknowledgment
46
+
47
+ - bailii-roberta was trained based on [roberta-base](https://arxiv.org/abs/1907.11692)).