Teja-Gollapudi commited on
Commit
602a29a
1 Parent(s): c0efbe8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -11,7 +11,7 @@ pipeline_tag: conversational
11
  # VMware/open-llama-0.3T-7B-open-instruct-v1.1
12
 
13
  ## License
14
- - Commercially viable
15
  - Instruction dataset, [VMware/open-instruct-v1.1-oasst-dolly-hhrlhf](https://huggingface.co/datasets/VMware/open-instruct-v1.1-oasst-dolly-hhrlhf) is under cc-by-sa-3.0
16
  - Language Model ([openlm-research/open_llama_7b_preview_300bt](https://huggingface.co/openlm-research/open_llama_7b_preview_300bt/tree/main/open_llama_7b_preview_300bt_transformers_weights)) is under apache-2.0 License
17
 
@@ -65,7 +65,11 @@ This way, the model can better understand the relationship between different par
65
 
66
  ## Drawbacks
67
 
68
- - The model was trained on a partially trained Open-LLaMA checkpoint. (300B tokens).
 
 
 
 
69
 
70
  ## Evaluation
71
 
 
11
  # VMware/open-llama-0.3T-7B-open-instruct-v1.1
12
 
13
  ## License
14
+ - <b>Commercially Viable </b>
15
  - Instruction dataset, [VMware/open-instruct-v1.1-oasst-dolly-hhrlhf](https://huggingface.co/datasets/VMware/open-instruct-v1.1-oasst-dolly-hhrlhf) is under cc-by-sa-3.0
16
  - Language Model ([openlm-research/open_llama_7b_preview_300bt](https://huggingface.co/openlm-research/open_llama_7b_preview_300bt/tree/main/open_llama_7b_preview_300bt_transformers_weights)) is under apache-2.0 License
17
 
 
65
 
66
  ## Drawbacks
67
 
68
+ - The model was trained on a partially trained Open-LLaMA checkpoint (300B tokens or 30% training life cycle), there is a huge potential for improvement when trained on fully trained Open-LLaMA checkpoints
69
+ - From what we have observed, the model strugles with few-shot prompting (We plan on addressing it with future iterations)
70
+ - When asked for code, it may or maynot output it with markdown format
71
+ - It doesn't indent python code
72
+
73
 
74
  ## Evaluation
75