PerRing commited on
Commit
6a2a725
1 Parent(s): fb5a352

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -1
README.md CHANGED
@@ -26,5 +26,52 @@ print(processor.decode(output[0], skip_special_tokens=True))
26
  ```
27
  ## result
28
  ```output
 
 
29
 
30
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
26
  ```
27
  ## result
28
  ```output
29
+ A chat between a curious human and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the human's questions. USER:
30
+ explain about this image. ASSISTANT:
31
 
32
+ This image shows a clear glass vase filled with water, which is placed on a surface that appears to be a balcony or a patio. Inside the vase, there are several purple flowers with ruffled petals, which could be a type of iris or a similar flower. The flowers are in full bloom, and their vibrant purple color stands out against the green leaves and stems. The background is blurred but suggests an outdoor setting with greenery, indicating that the flowers are likely in a garden or a balcony garden. The overall atmosphere of the image is serene and natural, with a focus on the beauty of the flowers.
33
+ ```
34
+
35
+ <br>
36
+ <br>
37
+
38
+ # Original(liuhaotian/llava-v1.6-34b) README.md
39
+ <br>
40
+ # LLaVA Model Card
41
+
42
+ ## Model details
43
+
44
+ **Model type:**
45
+ LLaVA is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data.
46
+ It is an auto-regressive language model, based on the transformer architecture.
47
+ Base LLM: [lmsys/vicuna-13b-v1.5](https://huggingface.co/lmsys/vicuna-13b-v1.5)
48
+
49
+ **Model date:**
50
+ LLaVA-v1.6-Vicuna-13B was trained in December 2023.
51
+
52
+ **Paper or resources for more information:**
53
+ https://llava-vl.github.io/
54
+
55
+ ## License
56
+ Llama 2 is licensed under the LLAMA 2 Community License,
57
+ Copyright (c) Meta Platforms, Inc. All Rights Reserved.
58
+
59
+ **Where to send questions or comments about the model:**
60
+ https://github.com/haotian-liu/LLaVA/issues
61
+
62
+ ## Intended use
63
+ **Primary intended uses:**
64
+ The primary use of LLaVA is research on large multimodal models and chatbots.
65
+
66
+ **Primary intended users:**
67
+ The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.
68
+
69
+ ## Training dataset
70
+ - 558K filtered image-text pairs from LAION/CC/SBU, captioned by BLIP.
71
+ - 158K GPT-generated multimodal instruction-following data.
72
+ - 500K academic-task-oriented VQA data mixture.
73
+ - 50K GPT-4V data mixture.
74
+ - 40K ShareGPT data.
75
+
76
+ ## Evaluation dataset
77
+ A collection of 12 benchmarks, including 5 academic VQA benchmarks and 7 recent benchmarks specifically proposed for instruction-following LMMs.