File size: 1,215 Bytes
4ad8cfb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
merge_method: linear
dtype: float32
out_dtype: bfloat16
models:
- model: /Users/jsarnecki/opt/Workspace/meta-llama/Llama-3.1-8B
parameters:
weight: 1.0
- model: /Users/jsarnecki/opt/Workspace/NousResearch/Hermes-3-Llama-3.1-8B
parameters:
weight: 0.0
tokenizer:
source: /Users/jsarnecki/opt/Workspace/NousResearch/Hermes-3-Llama-3.1-8B
tokens:
<|start_header_id|>:
source: /Users/jsarnecki/opt/Workspace/NousResearch/Hermes-3-Llama-3.1-8B
force: true
<|end_header_id|>:
source: /Users/jsarnecki/opt/Workspace/NousResearch/Hermes-3-Llama-3.1-8B
force: true
<|eot_id|>:
source: /Users/jsarnecki/opt/Workspace/NousResearch/Hermes-3-Llama-3.1-8B
force: true
<|begin_of_text|>:
source: /Users/jsarnecki/opt/Workspace/NousResearch/Hermes-3-Llama-3.1-8B
force: true
<|end_of_text|>:
source: /Users/jsarnecki/opt/Workspace/NousResearch/Hermes-3-Llama-3.1-8B
force: true
<|im_start|>:
source: /Users/jsarnecki/opt/Workspace/NousResearch/Hermes-3-Llama-3.1-8B
force: true
<|im_end|>:
source: /Users/jsarnecki/opt/Workspace/NousResearch/Hermes-3-Llama-3.1-8B
force: true
|