File size: 2,141 Bytes
0990e7b
 
28c6723
 
 
 
 
0990e7b
28c6723
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4b77fea
 
 
 
28c6723
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
license: apache-2.0
language:
- en
library_name: transformers
datasets:
- budecosystem/intellecta
---

<div align="center"><img src="https://raw.githubusercontent.com/BudEcosystem/boomer/main/assets/boomer-logo.png" width=200></div>


<p align="center"><i>Democratizing access to LLMs for the open-source community.<br>Let's advance AI, together. </i></p>

----

## Introduction 🎉

We are thrilled to announce the open-sourcing of our boomer-634m model, an important milestone in our ongoing AI research. This model, with 634 million parameters, was meticulously pre-trained from scratch on a custom synthetic dataset comprising 12 billion tokens.



## Run the model

Here is a quick guide to get you started with boomer-634m:
Please note that, at the moment, `trust_remote_code=True` is required for running the model.

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("budecosystem/boomer-634m",
                                             trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("budecosystem/boomer-634m")

input_ids = tokenizer("Explain why the sky is blue.", return_tensors='pt').to(model.device)["input_ids"]
outputs = model.generate(input_ids, max_new_tokens=216)
print(tokenizer.batch_decode(outputs))

```

## Evaluations

The boomer-634m model has been rigorously evaluated on various benchmarks, showcasing its robust performance across different tasks:

| Model Name   | MMLU  | ARC   | Hellaswag | GSM8K | Winogrande | MathQA | logiqa |
|--------------|-------|-------|-----------|-------|------------|--------|--------|
| boomer-634m  | 25.91 | 29.86 | 39.24     | 1.67  | 50.67      | 23.55  | 28.42  |




### Final thought on Boomer!

Embarking on the journey with boomer-634m is just the beginning. We are committed to developing more advanced, efficient, and accessible AI models. Join us in this exciting adventure to shape the future of AI. 


### Aknowledgements

Our heartfelt thanks go to the open-source community and the trailblazers in AI research whose work has paved the way for innovations like boomer-634m.