wwe180 commited on
Commit
7eba0a9
1 Parent(s): e5445fb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -39
README.md CHANGED
@@ -1,56 +1,47 @@
1
  ---
2
  base_model:
3
- - NousResearch/Hermes-2-Theta-Llama-3-8B
4
- - camillop/Meta-Llama-3-8B-ORPO-ITA-llama-adapters
5
- - gradientai/Llama-3-8B-Instruct-Gradient-1048k
6
- - openchat/openchat-3.6-8b-20240522
7
- - hfl/llama-3-chinese-8b-instruct-v2-lora
8
- - Vdr1/L3-8B-sunfall-Stheno-v3.1
9
- - hfl/llama-3-chinese-8b-instruct-v2-lora
10
- - migtissera/Llama-3-8B-Synthia-v3.5
11
- - unstoppable123/LLaMA3-8B_chinese_lora_sft_v0.2
12
- - shenzhi-wang/Llama3-8B-Chinese-Chat
13
- - Jiar/Llama-3-8B-Chinese
14
- - Sao10K/L3-8B-Stheno-v3.2
15
- - shibing624/llama-3-8b-instruct-262k-chinese-lora
16
  library_name: transformers
17
  tags:
18
  - mergekit
19
  - merge
20
-
 
 
21
  ---
 
 
 
22
  # merge
23
 
24
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
25
 
26
- ## Merge Details
27
- ### Merge Method
28
 
29
- This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [gradientai/Llama-3-8B-Instruct-Gradient-1048k](https://huggingface.co/gradientai/Llama-3-8B-Instruct-Gradient-1048k) as a base.
30
 
31
- ### Models Merged
 
32
 
33
- The following models were included in the merge:
34
- * [NousResearch/Hermes-2-Theta-Llama-3-8B](https://huggingface.co/NousResearch/Hermes-2-Theta-Llama-3-8B) + [camillop/Meta-Llama-3-8B-ORPO-ITA-llama-adapters](https://huggingface.co/camillop/Meta-Llama-3-8B-ORPO-ITA-llama-adapters)
35
- * [openchat/openchat-3.6-8b-20240522](https://huggingface.co/openchat/openchat-3.6-8b-20240522) + [hfl/llama-3-chinese-8b-instruct-v2-lora](https://huggingface.co/hfl/llama-3-chinese-8b-instruct-v2-lora)
36
- * [Vdr1/L3-8B-sunfall-Stheno-v3.1](https://huggingface.co/Vdr1/L3-8B-sunfall-Stheno-v3.1) + [hfl/llama-3-chinese-8b-instruct-v2-lora](https://huggingface.co/hfl/llama-3-chinese-8b-instruct-v2-lora)
37
- * [migtissera/Llama-3-8B-Synthia-v3.5](https://huggingface.co/migtissera/Llama-3-8B-Synthia-v3.5) + [unstoppable123/LLaMA3-8B_chinese_lora_sft_v0.2](https://huggingface.co/unstoppable123/LLaMA3-8B_chinese_lora_sft_v0.2)
38
- * [shenzhi-wang/Llama3-8B-Chinese-Chat](https://huggingface.co/shenzhi-wang/Llama3-8B-Chinese-Chat) + [Jiar/Llama-3-8B-Chinese](https://huggingface.co/Jiar/Llama-3-8B-Chinese)
39
- * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2) + [shibing624/llama-3-8b-instruct-262k-chinese-lora](https://huggingface.co/shibing624/llama-3-8b-instruct-262k-chinese-lora)
40
 
41
- ### Configuration
 
42
 
43
- The following YAML configuration was used to produce this model:
44
-
45
- ```yaml
46
- models:
47
- - model: shenzhi-wang/Llama3-8B-Chinese-Chat+Jiar/Llama-3-8B-Chinese
48
- - model: NousResearch/Hermes-2-Theta-Llama-3-8B+camillop/Meta-Llama-3-8B-ORPO-ITA-llama-adapters
49
- - model: migtissera/Llama-3-8B-Synthia-v3.5+unstoppable123/LLaMA3-8B_chinese_lora_sft_v0.2
50
- - model: openchat/openchat-3.6-8b-20240522+hfl/llama-3-chinese-8b-instruct-v2-lora
51
- - model: Sao10K/L3-8B-Stheno-v3.2+shibing624/llama-3-8b-instruct-262k-chinese-lora
52
- - model: Vdr1/L3-8B-sunfall-Stheno-v3.1+hfl/llama-3-chinese-8b-instruct-v2-lora
53
- merge_method: model_stock
54
- base_model: gradientai/Llama-3-8B-Instruct-Gradient-1048k
55
- dtype: bfloat16
56
  ```
 
 
 
 
 
 
 
 
1
  ---
2
  base_model:
3
+ - wwe180/L3-8B-LingYang-v1
 
 
 
 
 
 
 
 
 
 
 
 
4
  library_name: transformers
5
  tags:
6
  - mergekit
7
  - merge
8
+ - Llama3
9
+ license:
10
+ - other
11
  ---
12
+
13
+ #The model is experimental, so the results cannot be guaranteed.
14
+
15
  # merge
16
 
17
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
 
 
 
19
 
20
+ ## 💻 Usage
21
 
22
+ ```python
23
+ !pip install -qU transformers accelerate
24
 
25
+ from transformers import AutoTokenizer
26
+ import transformers
27
+ import torch
 
 
 
 
28
 
29
+ model = "L3-8B-LingYang-v1"
30
+ messages = [{"role": "user", "content": "What is a large language model?"}]
31
 
32
+ tokenizer = AutoTokenizer.from_pretrained(model)
33
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
34
+ pipeline = transformers.pipeline(
35
+ "text-generation",
36
+ model=model,
37
+ torch_dtype=torch.float16,
38
+ device_map="auto",
39
+ )
 
 
 
 
 
40
  ```
41
+ ## Statement:
42
+
43
+ L3-8B-LingYang-v1 does not represent the views and positions of the model developers We will not be liable for any problems arising from the use of the L3-8B-LingYang-v1 open Source model, including but not limited to data security issues, risk of public opinion, or any risks and problems arising from the misdirection, misuse, dissemination or misuse of the model.
44
+
45
+
46
+
47
+