Leia-Swallow-13B

LEIA is a training technique for autoregressive LLMs that effectively improves their performance in languages other than English by enhancing cross-lingual knowledge transfer from English to a target language. This model is constructed by applying LEIA to Swallow, a Japanese-English bilingual LLM based on LLaMA 2. The model achieves enhanced performance on four out of six Japanese question answering benchmarks and equivalent performance on the remaining two, as reported below.

Please refer to our paper or blog post (in Japanese) for further technical details.

Model List

Empirical Results

The model is assessed using the following six question answering benchmarks:

  • X-CODAH
  • X-CSQA
  • JCommonsenseQA
  • NIILC
  • JEMHopQA
  • JAQKET v2
Model X-CODAH X-CSQA JCommonsenseQA NIILC JEMHopQA JAQKET v2
Swallow 43.3 41.8 89.3 64.1 50.6 88.9
LEIA 44.0 41.9 89.3 65.8 50.6 89.6

For further details of this experiment, please refer to our paper.

Contributors

  • Ikuya Yamada (Studio Ousia, RIKEN)
  • Ryokan Ri (LY Corporation, SB Intuitions)
Downloads last month
501
Safetensors
Model size
13.1B params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for leia-llm/Leia-Swallow-13b

Quantizations
2 models