File size: 1,130 Bytes
b49f93c
 
5b75bf5
 
 
b49f93c
 
5b75bf5
b49f93c
5b75bf5
 
 
b49f93c
5b75bf5
b49f93c
5b75bf5
 
 
 
 
 
 
 
 
 
b49f93c
5b75bf5
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
---
library_name: transformers
license: openrail
datasets:
- nuprl/MultiPL-T
---

# MultiPL-T CodeLlama-70b

This repository holds several [CodeLlama-13b](https://huggingface.co/meta-llama/CodeLlama-70b-hf) fine-tunes, all fine-tuned on MultiPL-T data. 
Examine the commit message to determine the language. Due to disk space constraints, we were not able to save all checkpoints. Instead, we have
uploaded the best checkpoint for each language.

For more information the training process, see the MultiPL-T paper:

```
@misc{cassano:multipl-t,
      title={Knowledge Transfer from High-Resource to Low-Resource Programming Languages for Code LLMs}, 
      author={Federico Cassano and John Gouwar and Francesca Lucchetti and Claire Schlesinger and Anders Freeman and Carolyn Jane Anderson and Molly Q Feldman and Michael Greenberg and Abhinav Jangda and Arjun Guha},
      year={2024},
      eprint={2308.09895},
      archivePrefix={arXiv},
      primaryClass={cs.PL}
}
```

For usage instructions, see the model card for the original model. Replace the model name with the name of this repository, and set revision=COMMIT_HASH.