AlekseiPravdin
commited on
Commit
•
058aaf8
1
Parent(s):
0cce991
Upload folder using huggingface_hub
Browse files
README.md
CHANGED
@@ -35,38 +35,20 @@ parameters:
|
|
35 |
dtype: float16
|
36 |
```
|
37 |
|
38 |
-
## Model Details
|
39 |
-
|
40 |
-
Hermes-2-Pro is an upgraded version of the Nous Hermes model, designed for general task and conversation capabilities, with a focus on function calling and structured outputs. It has been fine-tuned on a cleaned version of the OpenHermes 2.5 dataset, achieving high scores in function calling evaluations. Llama3-8B-Chinese-Chat is an instruction-tuned model specifically for Chinese and English users, excelling in roleplaying and tool-using tasks.
|
41 |
-
|
42 |
-
## Description
|
43 |
-
|
44 |
-
The merged model combines the advanced generative capabilities of Hermes-2-Pro with the specialized tuning of Llama3-8B-Chinese-Chat. This results in a versatile model that excels in both English and Chinese text generation, providing enhanced context understanding and nuanced responses across various NLP tasks.
|
45 |
-
|
46 |
-
## Use Cases
|
47 |
-
|
48 |
-
- **Conversational AI**: Engage users in natural dialogue in both English and Chinese.
|
49 |
-
- **Function Calling**: Execute predefined functions based on user queries, enhancing interactivity.
|
50 |
-
- **Roleplaying**: Simulate characters or scenarios in a conversational context.
|
51 |
-
- **Text Generation**: Generate creative content, including stories, poems, and structured outputs.
|
52 |
-
|
53 |
## Model Features
|
54 |
|
55 |
-
-
|
56 |
-
- **Function Calling**: Enhanced ability to perform actions based on user input, improving user experience.
|
57 |
-
- **Structured Outputs**: Capable of generating outputs in specific formats, such as JSON, for easier integration into applications.
|
58 |
|
59 |
## Evaluation Results
|
60 |
|
61 |
-
|
62 |
-
-
|
63 |
-
|
64 |
-
## Limitations
|
65 |
|
66 |
-
|
|
|
|
|
67 |
|
68 |
-
|
69 |
-
- **Contextual Understanding**: Although improved, the model may still struggle with highly nuanced or context-specific queries.
|
70 |
-
- **Performance Variability**: Performance may vary based on the complexity of the task and the language used.
|
71 |
|
72 |
-
|
|
|
35 |
dtype: float16
|
36 |
```
|
37 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
38 |
## Model Features
|
39 |
|
40 |
+
This fusion model combines the robust generative capabilities of [NousResearch/Hermes-2-Pro-Llama-3-8B](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B) with the refined tuning of [shenzhi-wang/Llama3-8B-Chinese-Chat](https://huggingface.co/shenzhi-wang/Llama3-8B-Chinese-Chat), creating a versatile model suitable for a variety of text generation tasks. Leveraging the strengths of both parent models, Hermes-2-Pro-Llama-3-8B-Llama3-8B-Chinese-Chat-slerp-merge provides enhanced context understanding, nuanced text generation, and improved performance across diverse NLP tasks, including multilingual capabilities and structured outputs.
|
|
|
|
|
41 |
|
42 |
## Evaluation Results
|
43 |
|
44 |
+
### Hermes-2-Pro-Llama-3-8B
|
45 |
+
- Scored 90% on function calling evaluation.
|
46 |
+
- Scored 84% on structured JSON output evaluation.
|
|
|
47 |
|
48 |
+
### Llama3-8B-Chinese-Chat
|
49 |
+
- Significant improvements in roleplay, function calling, and math capabilities compared to previous versions.
|
50 |
+
- Achieved high performance in both Chinese and English tasks, surpassing ChatGPT in certain benchmarks.
|
51 |
|
52 |
+
## Limitations
|
|
|
|
|
53 |
|
54 |
+
While the merged model inherits the strengths of both parent models, it may also carry over some limitations and biases. For instance, the model may exhibit inconsistencies in responses when handling complex queries or when generating content that requires deep contextual understanding. Additionally, the model's performance may vary based on the language used, with potential biases present in the training data affecting the quality of outputs in less represented languages or dialects. Users should remain aware of these limitations when deploying the model in real-world applications.
|