Safetensors
llama

Model Description

DTA_llama3_8b is from the paper "Divide-Then-Aggregate: An Efficient Tool Learning Method via Parallel Tool Invocation". It is a large language model capable of invoking tools and can parallel invoke multiple tools within a single round. The tool format it used is similar to OpenAI's Function Call.

Uses

The related code can be found in our GitHub repository.

Training Data

The training data comes from our specially constructed DTA-Tool, which is derived from ToolBench.

Evaluation

Testing Data

We evaluated the performance of DTA-Llama on StableToolBench.

Results

result

Citation

@misc{zhu2025dividethenaggregateefficienttoollearning,
      title={Divide-Then-Aggregate: An Efficient Tool Learning Method via Parallel Tool Invocation}, 
      author={Dongsheng Zhu and Weixian Shi and Zhengliang Shi and Zhaochun Ren and Shuaiqiang Wang and Lingyong Yan and Dawei Yin},
      year={2025},
      eprint={2501.12432},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2501.12432}, 
}
Downloads last month
3
Safetensors
Model size
8.03B params
Tensor type
F32
·
Inference API
Unable to determine this model's library. Check the docs .

Model tree for dongsheng/DTA_llama3_8b

Finetuned
(32)
this model

Dataset used to train dongsheng/DTA_llama3_8b

Collection including dongsheng/DTA_llama3_8b