slices: - sources: - model: ResplendentAI/Datura_7B layer_range: [0, 32] - model: Epiculous/Mika-7B layer_range: [0, 32] merge_method: slerp base_model: ResplendentAI/Datura_7B parameters: t: - filter: self_attn value: [0, 0.7, 0.4, 0.6, 1] # Adjusted weights for self attention layers - filter: mlp value: [0.8, 0.5, 0.7, 0.3, 0] # Adjusted weights for MLP layers - value: 0.6 # Adjusted default weight dtype: bfloat16