michaelfeil commited on
Commit
8ccfcb7
·
1 Parent(s): b412b0a

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -1,34 +1,9 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
  *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
  *.tflite filter=lfs diff=lfs merge=lfs -text
29
- *.tgz filter=lfs diff=lfs merge=lfs -text
30
- *.wasm filter=lfs diff=lfs merge=lfs -text
31
- *.xz filter=lfs diff=lfs merge=lfs -text
32
- *.zip filter=lfs diff=lfs merge=lfs -text
33
- *.zst filter=lfs diff=lfs merge=lfs -text
34
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ *.bin.* filter=lfs diff=lfs merge=lfs -text
2
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
3
  *.bin filter=lfs diff=lfs merge=lfs -text
 
 
 
 
4
  *.h5 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  *.tflite filter=lfs diff=lfs merge=lfs -text
6
+ *.tar.gz filter=lfs diff=lfs merge=lfs -text
7
+ *.ot filter=lfs diff=lfs merge=lfs -text
8
+ *.onnx filter=lfs diff=lfs merge=lfs -text
9
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
 
 
README.md ADDED
@@ -0,0 +1,150 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - ctranslate2
4
+ - translation
5
+ license: cc-by-4.0
6
+ ---
7
+ # # Fast-Inference with Ctranslate2
8
+ Speedup inference by 2x-8x using int8 inference in C++
9
+
10
+ quantized version of [Helsinki-NLP/opus-mt-en-de](https://huggingface.co/Helsinki-NLP/opus-mt-en-de)
11
+ ```bash
12
+ pip install hf-hub-ctranslate2>=1.0.0 ctranslate2>=3.13.0
13
+ ```
14
+ Converted using
15
+ ```
16
+ ct2-transformers-converter --model Helsinki-NLP/opus-mt-en-de --output_dir /home/michael/tmp-ct2fast-opus-mt-en-de --force --copy_files README.md generation_config.json tokenizer_config.json vocab.json source.spm .gitattributes target.spm --quantization float16
17
+ ```
18
+
19
+ Checkpoint compatible to [ctranslate2](https://github.com/OpenNMT/CTranslate2) and [hf-hub-ctranslate2](https://github.com/michaelfeil/hf-hub-ctranslate2)
20
+ - `compute_type=int8_float16` for `device="cuda"`
21
+ - `compute_type=int8` for `device="cpu"`
22
+
23
+ ```python
24
+ from hf_hub_ctranslate2 import TranslatorCT2fromHfHub, GeneratorCT2fromHfHub
25
+ from transformers import AutoTokenizer
26
+
27
+ model_name = "michaelfeil/ct2fast-opus-mt-en-de"
28
+ # use either TranslatorCT2fromHfHub or GeneratorCT2fromHfHub here, depending on model.
29
+ model = TranslatorCT2fromHfHub(
30
+ # load in int8 on CUDA
31
+ model_name_or_path=model_name,
32
+ device="cuda",
33
+ compute_type="int8_float16",
34
+ tokenizer=AutoTokenizer.from_pretrained("Helsinki-NLP/opus-mt-en-de")
35
+ )
36
+ outputs = model.generate(
37
+ text=["How do you call a fast Flan-ingo?", "User: How are you doing?"],
38
+ )
39
+ print(outputs)
40
+ ```
41
+
42
+ # Licence and other remarks:
43
+ This is just a quantized version. Licence conditions are intended to be idential to original huggingface repo.
44
+
45
+ # Original description
46
+
47
+
48
+ ### opus-mt-en-de
49
+
50
+
51
+ ## Table of Contents
52
+ - [Model Details](#model-details)
53
+ - [Uses](#uses)
54
+ - [Risks, Limitations and Biases](#risks-limitations-and-biases)
55
+ - [Training](#training)
56
+ - [Evaluation](#evaluation)
57
+ - [Citation Information](#citation-information)
58
+ - [How to Get Started With the Model](#how-to-get-started-with-the-model)
59
+
60
+ ## Model Details
61
+ **Model Description:**
62
+ - **Developed by:** Language Technology Research Group at the University of Helsinki
63
+ - **Model Type:** Translation
64
+ - **Language(s):**
65
+ - Source Language: English
66
+ - Target Language: German
67
+ - **License:** CC-BY-4.0
68
+ - **Resources for more information:**
69
+ - [GitHub Repo](https://github.com/Helsinki-NLP/OPUS-MT-train)
70
+
71
+
72
+ ## Uses
73
+
74
+ #### Direct Use
75
+
76
+ This model can be used for translation and text-to-text generation.
77
+
78
+
79
+ ## Risks, Limitations and Biases
80
+
81
+
82
+
83
+ **CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.**
84
+
85
+ Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)).
86
+
87
+ Further details about the dataset for this model can be found in the OPUS readme: [en-de](https://github.com/Helsinki-NLP/OPUS-MT-train/blob/master/models/en-de/README.md)
88
+
89
+
90
+ #### Training Data
91
+ ##### Preprocessing
92
+ * pre-processing: normalization + SentencePiece
93
+
94
+ * dataset: [opus](https://github.com/Helsinki-NLP/Opus-MT)
95
+ * download original weights: [opus-2020-02-26.zip](https://object.pouta.csc.fi/OPUS-MT-models/en-de/opus-2020-02-26.zip)
96
+
97
+ * test set translations: [opus-2020-02-26.test.txt](https://object.pouta.csc.fi/OPUS-MT-models/en-de/opus-2020-02-26.test.txt)
98
+
99
+ ## Evaluation
100
+
101
+ #### Results
102
+
103
+ * test set scores: [opus-2020-02-26.eval.txt](https://object.pouta.csc.fi/OPUS-MT-models/en-de/opus-2020-02-26.eval.txt)
104
+
105
+
106
+ #### Benchmarks
107
+
108
+ | testset | BLEU | chr-F |
109
+ |-----------------------|-------|-------|
110
+ | newssyscomb2009.en.de | 23.5 | 0.540 |
111
+ | news-test2008.en.de | 23.5 | 0.529 |
112
+ | newstest2009.en.de | 22.3 | 0.530 |
113
+ | newstest2010.en.de | 24.9 | 0.544 |
114
+ | newstest2011.en.de | 22.5 | 0.524 |
115
+ | newstest2012.en.de | 23.0 | 0.525 |
116
+ | newstest2013.en.de | 26.9 | 0.553 |
117
+ | newstest2015-ende.en.de | 31.1 | 0.594 |
118
+ | newstest2016-ende.en.de | 37.0 | 0.636 |
119
+ | newstest2017-ende.en.de | 29.9 | 0.586 |
120
+ | newstest2018-ende.en.de | 45.2 | 0.690 |
121
+ | newstest2019-ende.en.de | 40.9 | 0.654 |
122
+ | Tatoeba.en.de | 47.3 | 0.664 |
123
+
124
+
125
+
126
+ ## Citation Information
127
+
128
+ ```bibtex
129
+ @InProceedings{TiedemannThottingal:EAMT2020,
130
+ author = {J{\"o}rg Tiedemann and Santhosh Thottingal},
131
+ title = {{OPUS-MT} — {B}uilding open translation services for the {W}orld},
132
+ booktitle = {Proceedings of the 22nd Annual Conferenec of the European Association for Machine Translation (EAMT)},
133
+ year = {2020},
134
+ address = {Lisbon, Portugal}
135
+ }
136
+ ```
137
+
138
+ ## How to Get Started With the Model
139
+ ```python
140
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
141
+
142
+ tokenizer = AutoTokenizer.from_pretrained("Helsinki-NLP/opus-mt-en-de")
143
+
144
+ model = AutoModelForSeq2SeqLM.from_pretrained("Helsinki-NLP/opus-mt-en-de")
145
+
146
+ ```
147
+
148
+
149
+
150
+
config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_source_bos": false,
3
+ "add_source_eos": false,
4
+ "bos_token": "<s>",
5
+ "decoder_start_token": "</s>",
6
+ "eos_token": "</s>",
7
+ "unk_token": "<unk>"
8
+ }
generation_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bad_words_ids": [
4
+ [
5
+ 58100
6
+ ]
7
+ ],
8
+ "bos_token_id": 0,
9
+ "decoder_start_token_id": 58100,
10
+ "eos_token_id": 0,
11
+ "forced_eos_token_id": 0,
12
+ "max_length": 512,
13
+ "num_beams": 4,
14
+ "pad_token_id": 58100,
15
+ "transformers_version": "4.27.0.dev0"
16
+ }
model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e19b3ee9697c56978f1fb8f001c5bfb31285547322be9897e8727e2c58308e96
3
+ size 148423101
shared_vocabulary.txt ADDED
The diff for this file is too large to render. See raw diff
 
source.spm ADDED
Binary file (768 kB). View file
 
target.spm ADDED
Binary file (797 kB). View file
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"target_lang": "de", "source_lang": "en"}
vocab.json ADDED
The diff for this file is too large to render. See raw diff