---
license: llama2
base_model:
- codellama/CodeLlama-34b-Python-hf
datasets:
- nvidia/OpenMathInstruct-1
language:
- en
library_name: nemo
tags:
- nvidia
- code
- math
---
# OpenMath-CodeLlama-34b-Python
OpenMath models were designed to solve mathematical problems by integrating text-based reasoning with code blocks
executed by Python interpreter. The models were trained on [OpenMathInstruct-1](https://huggingface.co/datasets/nvidia/OpenMathInstruct-1),
a math instruction tuning dataset with 1.8M problem-solution pairs generated using permissively licensed
[Mixtral-8x7B](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) model.
|
greedy |
majority@50 |
model |
GSM8K |
MATH |
GMS8K |
MATH |
OpenMath-CodeLlama-7B (nemo | HF) |
75.9 |
43.6 |
84.8 |
55.6 |
OpenMath-Mistral-7B (nemo | HF) |
80.2 |
44.5 |
86.9 |
57.2 |
OpenMath-CodeLlama-13B (nemo | HF) |
78.8 |
45.5 |
86.8 |
57.6 |
OpenMath-CodeLlama-34B (nemo | HF) |
80.7 |
48.3 |
88.0 |
60.2 |
OpenMath-Llama2-70B (nemo | HF) |
84.7 |
46.3 |
90.1 |
58.3 |
OpenMath-CodeLlama-70B (nemo | HF) |
84.6 |
50.7 |
90.8 |
60.4 |
The pipeline we used to produce these models is fully open-sourced!
- [Code](https://github.com/Kipok/NeMo-Skills)
- [Models](https://huggingface.co/collections/nvidia/openmath-65c5619de2ba059be0775014)
- [Dataset](https://huggingface.co/datasets/nvidia/OpenMathInstruct-1)
See our [paper](https://arxiv.org/abs/2402.10176) for more details!
# How to use the models?
Try to [run inference with our models](https://github.com/Kipok/NeMo-Skills/blob/main/docs/inference.md) with just a few commands!
# Reproducing our results
We provide [all instructions](https://github.com/Kipok/NeMo-Skills/blob/main/docs/reproducing-results.md) to fully reproduce our results.
# Improving other models
To improve other models or to learn more about our code, read through the docs below.
- [NeMo-Skills Pipeline](https://github.com/Kipok/NeMo-Skills)
- [Generating synthetic data](https://github.com/Kipok/NeMo-Skills/blob/main/docs/synthetic-data-generation.md)
- [Finetuning models](https://github.com/Kipok/NeMo-Skills/blob/main/docs/finetuning.md)
- [Evaluating models](https://github.com/Kipok/NeMo-Skills/blob/main/docs/evaluation.md)
In our pipeline we use [NVIDIA NeMo](https://www.nvidia.com/en-us/ai-data-science/generative-ai/nemo-framework/),
an end-to-end, cloud-native framework to build, customize, and deploy generative AI models anywhere.
It includes training and inferencing frameworks, guardrailing toolkits, data curation tools, and pretrained models,
offering enterprises an easy, cost-effective, and fast way to adopt generative AI.
# Citation
If you find our work useful, please consider citing us!
```bibtex
@article{toshnival2024openmath,
title = {OpenMathInstruct-1: A 1.8 Million Math Instruction Tuning Dataset},
author = {Shubham Toshniwal and Ivan Moshkov and Sean Narenthiran and Daria Gitman and Fei Jia and Igor Gitman},
year = {2024},
journal = {arXiv preprint arXiv: Arxiv-2402.10176}
}
```
# License
The use of this model is governed by the [Llama 2 Community License Agreement](https://ai.meta.com/llama/license/)