--- language: - multilingual thumbnail: url to a thumbnail used in social sharing tags: - coding - moe license: mit base_model: ContextualAI/Contextual_KTO_Mistral_PairRM pipeline_tag: text-generation --- ## Usage NebulaNet-v2: An MOE of 4 7b expert models. It is good at coding and multi language translation. It should be fluent at chat and math too. The 4x7b merged model performs much better than the original Contextual_KTO_Mistral_PairRM on both coding and multilingual text generation in my observation. ## mergekit config ``` base_model: ContextualAI/Contextual_KTO_Mistral_PairRM experts: - source_model: ContextualAI/Contextual_KTO_Mistral_PairRM positive_prompts: - "chat" - "assistant" - "tell me" - "explain" - "I want" - source_model: Nexusflow/Starling-LM-7B-beta positive_prompts: - "code" - "python" - "javascript" - "programming" - "algorithm" - source_model: snorkelai/Snorkel-Mistral-PairRM-DPO positive_prompts: - "" - source_model: mlabonne/NeuralDaredevil-7B positive_prompts: - "reason" - "math" - "mathematics" - "solve" - "count" ```