jiachenli commited on
Commit
361f7b6
1 Parent(s): 2bb137f
Files changed (4) hide show
  1. .gitattributes +5 -0
  2. README.md +36 -0
  3. tokenizer.model +3 -0
  4. training_args.bin +3 -0
.gitattributes CHANGED
@@ -33,3 +33,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ model-00001-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
37
+ model-00002-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
38
+ model-00003-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
39
+ model-00004-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text
40
+ tokenizer.model filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,39 @@
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
4
+
5
+ # CuMo Model Card
6
+
7
+ ## Model details
8
+
9
+ **Model type:**
10
+ CuMo-7B is based on [Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
11
+
12
+ **Model date:**
13
+ LLaVA-v1.6-Mistral-7B was trained in December 2023.
14
+
15
+ **Paper or resources for more information:**
16
+ https://llava-vl.github.io/
17
+
18
+ ## License
19
+ [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) license.
20
+
21
+ **Where to send questions or comments about the model:**
22
+ https://github.com/haotian-liu/LLaVA/issues
23
+
24
+ ## Intended use
25
+ **Primary intended uses:**
26
+ The primary use of LLaVA is research on large multimodal models and chatbots.
27
+
28
+ **Primary intended users:**
29
+ The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
30
+
31
+ ## Training dataset
32
+ - 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.
33
+ - 158K GPT-generated multimodal instruction-following data.
34
+ - 500K academic-task-oriented VQA data mixture.
35
+ - 50K GPT-4V data mixture.
36
+ - 40K ShareGPT data.
37
+
38
+ ## Evaluation dataset
39
+ A collection of 12 benchmarks, including 5 academic VQA benchmarks and 7 recent benchmarks specifically proposed for instruction-following LMMs.
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e07781058fad2dc0cf6ae56705ae727b72e514536c68c3320062ed136159a485
3
+ size 7608