Edit model card

Eurus: A suit of open-source LLMs optimized for reasoning

Introduction β€’ Evaluation

Links

Introduction

Eurus-70B-SFT is fine-tuned from CodeLLaMA-70B on all correct actions in UltraInteract, mixing a small proportion of UltraChat, ShareGPT, and OpenOrca examples.

It achieves better performance than other open-source models of similar sizes and even outperforms specialized models in corresponding domains in many cases.

Usage

We apply tailored prompts for coding and math, consistent with UltraInteract data formats:

Coding

[INST] Write Python code to solve the task:
{Instruction} [/INST]

Math-CoT

[INST] Solve the following math problem step-by-step.
Simplify your answer as much as possible. Present your final answer as \\boxed{Your Answer}.
{Instruction} [/INST]

Math-PoT

[INST] Tool available:
[1] Python interpreter
When you send a message containing Python code to python, it will be executed in a stateful Jupyter notebook environment.
Solve the following math problem step-by-step.
Simplify your answer as much as possible.
{Instruction} [/INST]

Evaluation

  • Eurus, both the 7B and 70B variants, achieve the best overall performance among open-source models of similar sizes. Eurus even outperforms specialized models in corresponding domains in many cases. Notably, Eurus-7B outperforms baselines that are 5Γ— larger, and Eurus-70B achieves better performance than GPT-3.5 Turbo.
  • Preference learning with UltraInteract can further improve performance, especially in math and the multi-turn ability.stats

Citation

@misc{yuan2024advancing,
      title={Advancing LLM Reasoning Generalists with Preference Trees}, 
      author={Lifan Yuan and Ganqu Cui and Hanbin Wang and Ning Ding and Xingyao Wang and Jia Deng and Boji Shan and Huimin Chen and Ruobing Xie and Yankai Lin and Zhenghao Liu and Bowen Zhou and Hao Peng and Zhiyuan Liu and Maosong Sun},
      year={2024},
      eprint={2404.02078},
      archivePrefix={arXiv},
      primaryClass={cs.AI}
}
Downloads last month
31
Safetensors
Model size
69B params
Tensor type
FP16
Β·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for openbmb/Eurus-70b-sft

Finetunes
1 model

Datasets used to train openbmb/Eurus-70b-sft

Spaces using openbmb/Eurus-70b-sft 2

Collection including openbmb/Eurus-70b-sft