ohwi commited on
Commit
ca2c22c
1 Parent(s): be6a47b

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +60 -0
README.md ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ja
4
+ tags:
5
+ - japanese-stablelm
6
+ - causal-lm
7
+ pipeline_tag: text-generation
8
+ base_model: stabilityai/japanese-stablelm-base-gamma-7b
9
+ datasets: argilla/ultrafeedback-binarized-preferences-cleaned
10
+ license: apache-2.0
11
+ extra_gated_fields:
12
+ Name: text
13
+ Email: text
14
+ Country: text
15
+ Organization or Affiliation: text
16
+ I allow Stability AI to contact me about information related to its models and research: checkbox
17
+ ---
18
+
19
+
20
+ # Reproduced Japanese Stable LM Instruct Gamma 7B
21
+
22
+ ## Model Description
23
+
24
+ This is a reproduction of 7B-parameter decoder-only Japanese language model fine-tuned on instruction-following datasets, built on top of the base model [Japanese Stable LM Base Gamma 7B](https://huggingface.co/stabilityai/japanese-stablelm-base-gamma-7b).
25
+
26
+ This model is trained with [notus](https://github.com/argilla-io/notus) code base.
27
+
28
+ *If you are in search of the official model, please check [Japanese Stable LM Instruct Gamma 7B](https://huggingface.co/stabilityai/japanese-stablelm-instruct-gamma-7b).*
29
+
30
+
31
+
32
+ ## Model Details
33
+
34
+ ### Training Datasets
35
+
36
+ - [Japanese translation of the Databricks Dolly-15k dataset](https://huggingface.co/datasets/kunishou/databricks-dolly-15k-ja)
37
+ - [Japanese translation of the subset of the Anthropic HH dataset](https://huggingface.co/datasets/fujiki/japanese_hh-rlhf-49k)
38
+ - [Wikinews](https://ja.wikinews.org/wi) [subset](https://huggingface.co/datasets/fujiki/llm-japanese-dataset_wikinews) of the [izumi-lab/llm-japanese-dataset](https://huggingface.co/datasets/izumi-lab/llm-japanese-dataset)
39
+
40
+
41
+ ### Benchmarks
42
+
43
+ The result is evaluated by [Nejumi-leaderboard Neo](https://github.com/wandb/llm-leaderboard/tree/b2723944d4955768cb93c18ffe162a8ff4e88955).
44
+
45
+ - llm-jp-eval:
46
+
47
+ |AVG |EL |FA |MC |MR |NLI |QA |RC |chabsa_set_f1|jamp_exact_match|janli_exact_match|jcommonsenseqa_exact_match|jemhopqa_char_f1|jnli_exact_match|jsem_exact_match|jsick_exact_match|jsquad_char_f1|niilc_char_f1|
48
+ |------|---|---|----|---|-----|------|------|-------------|----------------|-----------------|--------------------------|----------------|----------------|----------------|-----------------|--------------|-------------|
49
+ |0.1691|0.0|0.0|0.24|0.0|0.286|0.1688|0.4887|0.0 |0.3 |0.56 |0.24 |0.1334 |0.08 |0.28 |0.21 |0.4887 |0.2042 |
50
+
51
+
52
+ - Japanese Mt-Bench:
53
+
54
+ |coding|extraction|humanities|math|reasoning|roleplay|stem|writing|
55
+ |------|----------|----------|----|---------|--------|----|-------|
56
+ |1.3 |1.75 |2.35 |1.45|3.4 |5.8 |4.3 |3.1 |
57
+
58
+
59
+ - Overall Average: 0.266
60
+