File size: 5,003 Bytes
785af45
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
---
license: apache-2.0
language:
- en
- zh
tags:
- rk3588
- rkllm
- Rockchip
- internlm2
---

# internlm2_rkLLM

- [中文](#书生浦语-18b)
- [English](#internlm2-18b)

## 书生·浦语-1.8B

### 介绍

internlm2_rkLLM 是从 [internlm2-chat-1_8b](https://huggingface.co/internlm/internlm2-chat-1_8b) 转换而来的 RKLLM 模型,专为 Rockchip 设备优化。该模型运行于 RK3588 的 NPU 上。

- **模型名称**: internlm2-chat-1_8b
- **模型架构**: 与 internlm2-chat-1_8b 相同
- **发布者**: FydeOS
- **日期**: 2024-06-03

### 模型详情

书生·浦语-1.8B (InternLM2-1.8B) 是第二代浦语模型系列的18亿参数版本。通过在线 RLHF 在 InternLM2-Chat-1.8B-SFT 之上进一步对齐。 InternLM2-Chat-1.8B表现出更好的指令跟随、聊天体验和函数调用,推荐下游应用程序使用。

InternLM2 模型具备以下的技术特点:
- 有效支持20万字超长上下文:模型在20万字长输入中几乎完美地实现长文“大海捞针”,而且在 LongBench 和 L-Eval 等长文任务中的表现也达到开源模型中的领先水平。
- 综合性能全面提升:各能力维度相比上一代模型全面进步,在推理、数学、代码等方面的能力提升显著。
### 使用指南

> 此模型仅支持搭载 Rockchip RK3588/s 芯片的设备。请确认设备信息并确保 NPU 可用。

#### openFyde 系统

> 请确保你已将系统升级到最新版本。

1. 下载模型文件 `XXX.rkllm`2. 新建文件夹 `model/`,将模型文件放置于该文件夹内。
3. 启动 FydeOS AI,在设置页面进行相关配置。

#### 其它系统
> 请确保已完成 RKLLM 的 NPU 相关内核更新。

1. 下载模型文件 `XXX.rkllm`2. 按照官方文档进行配置:[官方文档](https://github.com/airockchip/rknn-llm)。

### 常见问题(FAQ)

如遇到问题,请先查阅 issue 区,若问题仍未解决,再提交新的 issue。

### 限制与注意事项

- 模型在某些情况下可能存在性能限制
- 使用时请遵循相关法律法规
- 可能需要进行适当的参数调优以达到最佳效果

### 许可证

本模型采用与 [internlm2-chat-1_8b](https://huggingface.co/internlm/internlm2-chat-1_8b) 相同的许可证。

### 联系方式

如需更多信息,请联系:

- **电子邮件**: [email protected]
- **主页**: [FydeOS AI](https://fydeos.ai/zh/)


## InternLM2-1.8B

### Introduction

internlm2_rkLLM is a RKLLM model derived from [internlm2-chat-1_8b](https://huggingface.co/internlm/internlm2-chat-1_8b), specifically optimized for Rockchip devices. This model operates on the NPU of the RK3588 chip.

- **Model Name**: internlm2-chat-1_8b
- **Architecture**: Identical to internlm2-chat-1_8b
- **Publisher**: FydeOS
- **Release Date**: 3 June 2024

### Model Details

InternLM2-1.8B is the 1.8 billion parameter version of the second generation InternLM series. Further aligned on top of InternLM2-Chat-1.8B-SFT through online RLHF. InternLM2-Chat-1.8B exhibits better instruction following, chat experience, and function calling, which is recommended for downstream applications. 

The InternLM2 has the following technical features:

- Effective support for ultra-long contexts of up to 200,000 characters: The model nearly perfectly achieves "finding a needle in a haystack" in long inputs of 200,000 characters. It also leads among open-source models in performance on long-text tasks such as LongBench and L-Eval.
- Comprehensive performance enhancement: Compared to the previous generation model, it shows significant improvements in various capabilities, including reasoning, mathematics, and coding.

### User Guide

This model is only supported on devices with the Rockchip RK3588/s chip. Please verify your device's chip information and ensure the NPU is operational.

#### openFyde System

> Ensure you have upgraded to the latest version of openFyde.

1. Download the model file `XXX.rkllm`.
2. Create a folder named `model/` and place the model file inside this folder.
3. Launch FydeOS AI and configure the settings on the settings page.

#### Other Systems

> Ensure you have updated the NPU kernel related to RKLLM.

1. Download the model file `XXX.rkllm`.
2. Follow the configuration guidelines provided in the [official documentation](https://github.com/airockchip/rknn-llm).

### FAQ

If you encounter issues, please refer to the issue section first. If your problem remains unresolved, submit a new issue.

### Limitations and Considerations

- The model may have performance limitations in certain scenarios.
- Ensure compliance with relevant laws and regulations during usage.
- Parameter tuning might be necessary to achieve optimal performance.

### License

This model is licensed under the same terms as [internlm2-chat-1_8b](https://huggingface.co/internlm/internlm2-chat-1_8b).

### Contact Information

For more information, please contact:

- **Email**: [email protected]
- **Homepage**: [FydeOS AI](https://fydeos.ai/en/)