RioShiina commited on
Commit
f081638
·
verified ·
1 Parent(s): 31d664d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +77 -3
README.md CHANGED
@@ -1,3 +1,77 @@
1
- ---
2
- license: llama3.1
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3
3
+ base_model_relation: quantized
4
+ license:
5
+ - llama3.1
6
+ - gemma
7
+ language:
8
+ - ja
9
+ - en
10
+ ---
11
+
12
+ **[4.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-8B-Instruct-v0.3-exl2/tree/4.0bpw)**
13
+ **[5.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-8B-Instruct-v0.3-exl2/tree/5.0bpw)**
14
+ **[6.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-8B-Instruct-v0.3-exl2/tree/6.0bpw)**
15
+ **[7.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-8B-Instruct-v0.3-exl2/tree/7.0bpw)**
16
+ **[8.0bpw](https://huggingface.co/rioshiina/Llama-3.1-Swallow-8B-Instruct-v0.3-exl2/tree/8.0bpw)**
17
+
18
+ # Llama-3.1-Swallow-8B-Instruct-v0.3-exl2
19
+ - Model creator: [tokyotech-llm](https://huggingface.co/tokyotech-llm)
20
+ - Original model: [Llama-3.1-Swallow-8B-Instruct-v0.3](https://huggingface.co/tokyotech-llm/Llama-3.1-Swallow-8B-Instruct-v0.3)
21
+
22
+ ### License
23
+
24
+ [META LLAMA 3.1 COMMUNITY LICENSE](https://www.llama.com/llama3_1/license/) and [Gemma Terms of Use](https://ai.google.dev/gemma/terms)
25
+
26
+ ## Prompt template
27
+
28
+ ```
29
+ <|begin_of_text|><|start_header_id|>system<|end_header_id|>
30
+
31
+ あなたは誠実で優秀な日本人のアシスタントです。<|eot_id|><|start_header_id|>user<|end_header_id|>
32
+
33
+ 東京の紅葉した公園で、東京タワーと高層ビルを背景に、空を舞うツバメと草地に佇むラマが出会う温かな物語を書いてください。<|eot_id|><|start_header_id|>assistant<|end_header_id|>
34
+
35
+ ```
36
+
37
+ ### Citations
38
+
39
+ ```tex
40
+ @inproceedings{Fujii:COLM2024,
41
+ title={Continual Pre-Training for Cross-Lingual LLM Adaptation:
42
+ Enhancing Japanese Language Capabilities},
43
+ author={Kazuki Fujii and Taishi Nakamura and Mengsay Loem and Hiroki
44
+ Iida and Masanari Ohi and Kakeru Hattori and Hirai Shota and Sakae
45
+ Mizuki and Rio Yokota and Naoaki Okazaki},
46
+ booktitle="Proceedings of the First Conference on Language Modeling",
47
+ series={COLM},
48
+ pages="(to appear)",
49
+ year="2024",
50
+ month=oct,
51
+ address={University of Pennsylvania, USA},
52
+ }
53
+
54
+ @inproceedings{Okazaki:COLM2024,
55
+ title={Building a Large Japanese Web Corpus for Large Language Models},
56
+ author={Naoaki Okazaki and Kakeru Hattori and Hirai Shota and Hiroki
57
+ Iida and Masanari Ohi and Kazuki Fujii and Taishi Nakamura and Mengsay
58
+ Loem and Rio Yokota and Sakae Mizuki},
59
+ booktitle="Proceedings of the First Conference on Language Modeling",
60
+ series={COLM},
61
+ pages="(to appear)",
62
+ year="2024",
63
+ month=oct,
64
+ address={University of Pennsylvania, USA},
65
+ }
66
+
67
+ @misc{dubey2024llama3herdmodels,
68
+ title={The Llama 3 Herd of Models},
69
+ author={Abhimanyu Dubey and Abhinav Jauhri and Abhinav Pandey and Abhishek Kadian and Ahmad Al-Dahle and Aiesha Letman and Akhil Mathur and Alan Schelten and Amy Yang and Angela Fan et al.},
70
+ year={2024},
71
+ eprint={2407.21783},
72
+ archivePrefix={arXiv},
73
+ primaryClass={cs.AI},
74
+ url={https://arxiv.org/abs/2407.21783},
75
+ }
76
+
77
+ ```