metadata
language:
- en
license: apache-2.0
library_name: transformers
tags:
- code
- QA
- reasoning
Model Card for Model ID
Model Details
Model Description
A power full MOE 4x7b mixtral of mistral models consists of for more accuracy and precision in general reasoning, QA and code. HuggingFaceH4/zephyr-7b-beta mistralai/Mistral-7B-Instruct-v0.2 teknium/OpenHermes-2.5-Mistral-7B Intel/neural-chat-7b-v3-3
- Developed by: NEXT AI
- Funded by : Zpay Labs Pvt Ltd.
- Model type: Mixtral of Mistral 4x7b
- Language(s) (NLP): Code-Reasoning-QA
Model Sources
- Demo : Https://nextai.co.in