renillhuang commited on
Commit
8fcce6d
β€’
1 Parent(s): b6095e9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -10
README.md CHANGED
@@ -8,10 +8,11 @@ widget:
8
  text: "Hello! How can I help you today?"
9
  pipeline_tag: text-generation
10
  ---
11
-
12
  <!-- markdownlint-disable first-line-h1 -->
13
  <!-- markdownlint-disable html -->
14
- ![](./assets/imgs/orion_start.PNG)
 
 
15
 
16
  <div align="center">
17
  <h1>
@@ -26,7 +27,7 @@ pipeline_tag: text-generation
26
  <p>
27
  <b>🌐English</b> |
28
  <a href="https://huggingface.co/OrionStarAI/Orion-14B-Base-Int4/blob/main/README_zh.md">πŸ‡¨πŸ‡³δΈ­ζ–‡</a><br><br>
29
- πŸ€— <a href="https://huggingface.co/OrionStarAI" target="_blank">HuggingFace Mainpage</a> | πŸ€– <a href="https://modelscope.cn/organization/OrionStarAI" target="_blank">ModelScope Mainpage</a><br>🎬 <a href="https://huggingface.co/spaces/OrionStarAI/Orion-14B-App-Demo" target="_blank">HuggingFace Demo</a> | 🎫 <a href="https://modelscope.cn/studios/OrionStarAI/Orion-14B-App-Demo/summary" target="_blank">ModelScope Demo</a>
30
  <p>
31
  </h4>
32
 
@@ -40,20 +41,34 @@ pipeline_tag: text-generation
40
  - [πŸ”— Model Download](#model-download)
41
  - [πŸ”– Model Benchmark](#model-benchmark)
42
  - [πŸ“Š Model Inference](#model-inference)
43
- - [πŸ₯‡ Company Introduction](#company-introduction)
44
  - [πŸ“œ Declarations & License](#declarations-license)
 
45
 
46
  # 1. Model Introduction
47
 
48
- - Orion-14B-Base-Int4 is quantized using awq from Orion-14B-Base while reducing model size by 70% and improving inference speed by 30%, with performance loss less than 1%.
49
- <div align="center">
50
- <img src="./assets/imgs/model_cap_en.png" alt="model_cap_en" width="50%" />
51
- </div>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
 
53
  - Orion-14B series models including:
54
  - **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
55
  - **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
56
- - **Orion-14B-LongChat:** This model is optimized for long context lengths more than 200k tokens and demonstrates performance comparable to proprietary models on long context evaluation sets.
57
  - **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
58
  - **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
59
  - **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.
@@ -321,7 +336,7 @@ Truly Useful Robots", OrionStar empowers more people through AI technology.
321
 
322
  **The core strengths of OrionStar lies in possessing end-to-end AI application capabilities,** including big data preprocessing, large model pretraining, fine-tuning, prompt engineering, agent, etc. With comprehensive end-to-end model training capabilities, including systematic data processing workflows and the parallel model training capability of hundreds of GPUs, it has been successfully applied in various industry scenarios such as government affairs, cloud services, international e-commerce, and fast-moving consumer goods.
323
 
324
- Companies with demands for deploying large-scale model applications are welcome to contact us.
325
  **Enquiry Hotline: 400-898-7779**<br>
326
  **E-mail: [email protected]**
327
 
 
8
  text: "Hello! How can I help you today?"
9
  pipeline_tag: text-generation
10
  ---
 
11
  <!-- markdownlint-disable first-line-h1 -->
12
  <!-- markdownlint-disable html -->
13
+ <div align="center">
14
+ <img src="./assets/imgs/orion_start.PNG" alt="logo" width="50%" />
15
+ </div>
16
 
17
  <div align="center">
18
  <h1>
 
27
  <p>
28
  <b>🌐English</b> |
29
  <a href="https://huggingface.co/OrionStarAI/Orion-14B-Base-Int4/blob/main/README_zh.md">πŸ‡¨πŸ‡³δΈ­ζ–‡</a><br><br>
30
+ πŸ€— <a href="https://huggingface.co/OrionStarAI" target="_blank">HuggingFace Mainpage</a> | πŸ€– <a href="https://modelscope.cn/organization/OrionStarAI" target="_blank">ModelScope Mainpage</a><br>🎬 <a href="https://huggingface.co/spaces/OrionStarAI/Orion-14B-App-Demo" target="_blank">HuggingFace Demo</a> | 🎫 <a href="https://modelscope.cn/studios/OrionStarAI/Orion-14B-App-Demo/summary" target="_blank">ModelScope Demo</a><br>😺 <a href="https://github.com/OrionStarAI/Orion" target="_blank">GitHub</a><br>πŸ“– <a href="https://github.com/OrionStarAI/Orion/blob/master/doc/Orion14B_v3.pdf" target="_blank">Tech Report</a>
31
  <p>
32
  </h4>
33
 
 
41
  - [πŸ”— Model Download](#model-download)
42
  - [πŸ”– Model Benchmark](#model-benchmark)
43
  - [πŸ“Š Model Inference](#model-inference)
 
44
  - [πŸ“œ Declarations & License](#declarations-license)
45
+ - [πŸ₯‡ Company Introduction](#company-introduction)
46
 
47
  # 1. Model Introduction
48
 
49
+ - Orion-14B series models are open-source multilingual large language models trained from scratch by OrionStarAI. The base model is trained on 2.5T multilingual corpus, including Chinese, English, Japanese, Korean, etc, and it exhibits superior performance in these languages. For details, please refer to [tech report](https://github.com/OrionStarAI/Orion/blob/master/doc/Orion14B_v3.pdf).
50
+
51
+ - The Orion-14B series models exhibit the following features:
52
+ - Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
53
+ - Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
54
+ - The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
55
+ - The long-chat version supports extremely long texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
56
+ - The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
57
+ <table style="border-collapse: collapse; width: 100%;">
58
+ <tr>
59
+ <td style="border: none; padding: 10px; box-sizing: border-box;">
60
+ <img src="./assets/imgs/opencompass_en.png" alt="opencompass" style="width: 100%; height: auto;">
61
+ </td>
62
+ <td style="border: none; padding: 10px; box-sizing: border-box;">
63
+ <img src="./assets/imgs/model_cap_en.png" alt="modelcap" style="width: 100%; height: auto;">
64
+ </td>
65
+ </tr>
66
+ </table>
67
 
68
  - Orion-14B series models including:
69
  - **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
70
  - **Orion-14B-Chat:** A chat-model fine-tuned on a high-quality corpus aims to provide an excellence interactive experience for users in the large model community.
71
+ - **Orion-14B-LongChat:** The long-context version excels at handling extremely lengthy texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
72
  - **Orion-14B-Chat-RAG:** A chat-model fine-tuned on a custom retrieval augmented generation dataset, achieving superior performance in retrieval augmented generation tasks.
73
  - **Orion-14B-Chat-Plugin:** A chat-model specifically tailored for plugin and function calling tasks, ideal for agent-related scenarios where the LLM acts as a plugin and function call system.
74
  - **Orion-14B-Base-Int4:** A quantized base model utilizing 4-bit integer weights. It significantly reduces the model size by 70% and increases the inference speed by 30% while incurring a minimal performance loss of only 1%.
 
336
 
337
  **The core strengths of OrionStar lies in possessing end-to-end AI application capabilities,** including big data preprocessing, large model pretraining, fine-tuning, prompt engineering, agent, etc. With comprehensive end-to-end model training capabilities, including systematic data processing workflows and the parallel model training capability of hundreds of GPUs, it has been successfully applied in various industry scenarios such as government affairs, cloud services, international e-commerce, and fast-moving consumer goods.
338
 
339
+ Companies with demands for deploying large-scale model applications are welcome to contact us.<br>
340
  **Enquiry Hotline: 400-898-7779**<br>
341
  **E-mail: [email protected]**
342