pronics2004 commited on
Commit
f12ea3b
1 Parent(s): 8f52763

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -10,7 +10,7 @@ This model is IBM's lightweight, 4-layer toxicity binary classifier for English.
10
 
11
 
12
  ## Feature
13
- This model offers very low inference latency and is capable of running on CPUs apart from GPUs and AIUs. It features 38 million parameters, reducing the number of hidden layers from 12 to 4, decreasing the hidden size from 768 to 576, and the intermediate size from 3072 to 768, compared to the original RoBERTa model architecture.
14
 
15
  ![Description of Image](38m_latency.png)
16
 
@@ -37,7 +37,7 @@ with torch.no_grad():
37
 
38
  ```
39
  ## Performance Comparison with Other Models
40
- The model delivers competitive performance with significantly lower inference latency. If a better F1 score is required, please refer to IBM's 12-layer model here.
41
 
42
  ![Description of Image](38m_comparison_a.png)
43
  ![Description of Image](38m_comparison_b.png)
 
10
 
11
 
12
  ## Feature
13
+ This model offers very low inference latency and is capable of running on CPUs apart from GPUs and AIUs. It features 38 million parameters, reducing the number of hidden layers from 12 to 4, decreasing the hidden size from 768 to 576, and the intermediate size from 3072 to 768, compared to the original RoBERTa model architecture. The latency on CPU vs accuracy numbers for this model in comparision to others is shown in the chart below.
14
 
15
  ![Description of Image](38m_latency.png)
16
 
 
37
 
38
  ```
39
  ## Performance Comparison with Other Models
40
+ The model outperforms most popular models with significantly lower inference latency. If a better F1 score is required, please refer to IBM's 12-layer model [here](https://huggingface.co/ibm-granite/granite-guardian-hap-125m).
41
 
42
  ![Description of Image](38m_comparison_a.png)
43
  ![Description of Image](38m_comparison_b.png)