tmoe-exp-v1 / mergekit_moe_config.yml
ehristoforu's picture
Upload folder using huggingface_hub
8d78564 verified
raw
history blame contribute delete
246 Bytes
base_model: beomi/EXAONE-3.5-2.4B-Instruct-Llamafied
gate_mode: random
architecture: mixtral
experts_per_token: 2
dtype: bfloat16
experts:
- source_model: beomi/EXAONE-3.5-2.4B-Instruct-Llamafied
- source_model: unsloth/Phi-3.5-mini-instruct