File size: 376 Bytes
ca466b7
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15

 models:
   - model: meta-llama/Meta-Llama-3-8B-Instruct
   - model: TsinghuaC3I/Llama-3-8B-UltraMedical
 merge_method: slerp
 base_model: meta-llama/Meta-Llama-3-8B-Instruct
 parameters:
   t:
     - filter: self_attn
       value: [0, 0.5, 0.3, 0.7, 1]
     - filter: mlp
       value: [1, 0.5, 0.7, 0.3, 0]
     - value: 0.5 # fallback for rest of tensors
 dtype: float16