|
--- |
|
inference: false |
|
language: |
|
- ja |
|
- en |
|
--- |
|
# webbigdata/ALMA-7B-Ja-GPTQ-Ja-En |
|
|
|
Original ALMA Model [ALMA-7B](https://huggingface.co/haoranxu/ALMA-7B). (26.95GB) is a new paradigm translation model. |
|
|
|
[ALMA-7B-Ja-GPTQ-Ja-En](https://huggingface.co/webbigdata/ALMA-7B-Ja) is a machine translation model that uses ALMA's learning method to translate Japanese to English.(13.3GB) |
|
|
|
This model is GPTQ quantized version model that reduces model size(3.9GB) and memory usage, although the performance is probably lower. |
|
And translation ability for languages other than Japanese and English has deteriorated significantly. |
|
|
|
[Free Colab Sample](https://github.com/webbigdata-jp/python_sample/blob/main/ALMA_7B_Ja_GPTQ_Ja_En_Free_Colab_sample.ipynb) |
|
|
|
If you want to translate the entire file at once, try Colab below. |
|
[ALMA_7B_Ja_GPTQ_Ja_En_batch_translation_sample](https://github.com/webbigdata-jp/python_sample/blob/main/ALMA_7B_Ja_GPTQ_Ja_En_batch_translation_sample.ipynb) |
|
|
|
if you enconter error below. |
|
```RuntimeError: probability tensor contains either `inf`, `nan` or element < 0``` |
|
It's mean your memory is not enough. decrease your num_beams or token size. |
|
|
|
|
|
**ALMA** (**A**dvanced **L**anguage **M**odel-based tr**A**nslator) is an LLM-based translation model, which adopts a new translation model paradigm: it begins with fine-tuning on monolingual data and is further optimized using high-quality parallel data. This two-step fine-tuning process ensures strong translation performance. |
|
Please find more details in their [paper](https://arxiv.org/abs/2309.11674). |
|
``` |
|
@misc{xu2023paradigm, |
|
title={A Paradigm Shift in Machine Translation: Boosting Translation Performance of Large Language Models}, |
|
author={Haoran Xu and Young Jin Kim and Amr Sharaf and Hany Hassan Awadalla}, |
|
year={2023}, |
|
eprint={2309.11674}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CL} |
|
} |
|
``` |
|
|
|
## about this work |
|
- **This work was done by :** [webbigdata](https://webbigdata.jp/). |