File size: 3,574 Bytes
be310d2
cfab146
be310d2
cfab146
be310d2
 
 
 
 
cfab146
 
0aed41e
 
be310d2
 
cfab146
 
 
be310d2
 
cfab146
 
be310d2
cfab146
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
be310d2
cfab146
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
be310d2
 
cfab146
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
77b7d02
cfab146
 
 
 
 
 
 
 
 
0aed41e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
---
base_model: meta-llama/Meta-Llama-3.1-8B
language:
- es
license: apache-2.0
tags:
- unsloth
- llama
- trl
library_name: transformers
pipeline_tag: text-generation
datasets:
- mariagrandury/elgrancorpus-it
---

# Model Details

## Model Description

- **Developed by:** mariagrandury
- **Model type:** Language model, instruction model
- **Language(s) (NLP):** es
- **License:** apache-2.0
- **Model fine-tuned from:** meta-llama/Meta-Llama-3.1-8B
- **Dataset used:** mariagrandury/elgrancorpus-it

<!--
## Model Sources

- **Paper**: Coming soon! ✨
- **Demo**: Coming soon! ✨
-->

# 💡 Uses

## Direct Use

This model's fine-tuning on an instructions dataset enables it to follow natural language instructions in Spanish. The direct use cases include virtual assistants and content generation.

## Downstream Use

This model is an instruct model, it’s primarily intended for direct use and may not be ideal for further fine-tuning. It serves as a general model suitable for a wide range of applications. However, for specific use cases within certain domains, fine-tuning with domain-specific data may improve the model's performance.

## Out-of-Scope Use

This model should not be used for production purposes without conducting a thorough assessment of risks and mitigation strategies.

# ⚠️ Bias, Risks, and Limitations

This model has limitations associated with both the underlying language model and the instruction tuning data. It is crucial to acknowledge that predictions generated by the model may inadvertently exhibit common deficiencies of language models, including hallucination, toxicity, and perpetuate harmful stereotypes across protected classes, identity characteristics, and sensitive, social, and occupational groups.

## Recommendations

Please, when utilizing this model, exercise caution and critically assess the output to mitigate the potential impact of biased or inaccurate information.

If considering this model for production use, it is crucial to thoroughly evaluate the associated risks and adopt suitable precautions. Conduct a comprehensive assessment to address any potential biases and ensure compliance with legal and ethical standards.

# 📚 Training Details

## Training Data

This model is based on [Meta Llama 3.1 8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) and has been fine-tuned using [elgrancorpus-it]().

It was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.

[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)

<!--
# ✅ Evaluation

We are evaluating the model and will publish the results soon.

### Results

Paper coming soon!
-->

# ⚙️ Technical Specifications

<!--
## Model Architecture and Objective
-->

## Compute Infrastructure

### Hardware

This model was trained using a GPU L4 with 53 GB for 1h.

### Software

We used the following libraries:

- `unsloth`
- `transformers`
- `peft`
- `accelerate`
- `bitsandbytes`

# 🌳 Environmental Impact

Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).

- **Hardware Type:** 1 X L4 - 53 GB
- **Hours used:** 1
- **Cloud Provider:** Google
- **Compute Region:** Europe
- **Carbon Emitted:** 72W x 1h = 0.07 kWh x 0.27 kg eq. CO2/kWh = 0.02 kg eq. CO2

<!--
# 🔥 How to Get Started with

```
```

# 📝 Citation

```
```
-->