--- base_model: - marketeam/Phi-Marketing - OEvortex/EMO-phi-128k library_name: transformers tags: - mergekit - merge --- # phi-3-market-emo This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the della merge method using [marketeam/Phi-Marketing](https://huggingface.co/marketeam/Phi-Marketing) as a base. ### Models Merged The following models were included in the merge: * [OEvortex/EMO-phi-128k](https://huggingface.co/OEvortex/EMO-phi-128k) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: marketeam/Phi-Marketing parameters: weight: 1.0 - model: OEvortex/EMO-phi-128k parameters: weight: 1.0 merge_method: della base_model: marketeam/Phi-Marketing parameters: density: 0.7 lambda: 1.1 epsilon: 0.2 ```