sarath-shekkizhar
commited on
Commit
•
6166b71
1
Parent(s):
3e0f2fa
Update README.md
Browse files
README.md
CHANGED
@@ -12,13 +12,16 @@ tags:
|
|
12 |
|
13 |
Introducing TenyxChat-8x7B-v1, part of our TenyxChat series trained to function as useful assistants through preference tuning, using Tenyx's recently released advanced fine-tuning technology ([VentureBeat article](https://venturebeat.com/ai/tenyx-aims-to-fix-llms-catastrophic-forgetting-problem/)). Our model is trained using the [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290) framework on the open-source AI feedback dataset [UltraFeedback](https://huggingface.co/datasets/HuggingFaceH4/ultrafeedback_binarized).
|
14 |
|
15 |
-
We fine-tune [Mixtral-8x7B-Instruct-v0.1](https://arxiv.org/pdf/2401.04088.pdf) with our proprietary approach ([blog](https://www.tenyx.com/post/forgetting-and-toxicity-in-llms-a-deep-dive-on-fine-tuning-methods)
|
|
|
|
|
|
|
16 |
|
17 |
# Model details
|
18 |
|
19 |
- Model type: Fine-tuned Mixture Of Expert 8x7B model for chat.
|
20 |
- License: Apache 2.0
|
21 |
-
- Base model:
|
22 |
- Demo: [spaces/tenyx/TenyxChat-8x7B-v1](https://huggingface.co/spaces/tenyx/TenyxChat-8x7B-v1)
|
23 |
|
24 |
## Usage
|
@@ -60,7 +63,7 @@ outputs = pipe(prompt, max_new_tokens=512, do_sample=False)
|
|
60 |
|
61 |
# Performance
|
62 |
|
63 |
-
At the time of release (Jan 2024), TenyxChat-8x7B-v1 is the highest-ranked
|
64 |
|
65 |
## MT-Bench
|
66 |
|
|
|
12 |
|
13 |
Introducing TenyxChat-8x7B-v1, part of our TenyxChat series trained to function as useful assistants through preference tuning, using Tenyx's recently released advanced fine-tuning technology ([VentureBeat article](https://venturebeat.com/ai/tenyx-aims-to-fix-llms-catastrophic-forgetting-problem/)). Our model is trained using the [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290) framework on the open-source AI feedback dataset [UltraFeedback](https://huggingface.co/datasets/HuggingFaceH4/ultrafeedback_binarized).
|
14 |
|
15 |
+
We fine-tune [Mixtral-8x7B-Instruct-v0.1](https://arxiv.org/pdf/2401.04088.pdf) with our proprietary approach ([blog](https://www.tenyx.com/post/forgetting-and-toxicity-in-llms-a-deep-dive-on-fine-tuning-methods), [service](https://www.tenyx.com/fine-tuning)),
|
16 |
+
similar to that of our [7B model](already applied to obtain TenyxChat-7B-v1 (https://huggingface.co/tenyx/TenyxChat-7B-v1), and show an increase in [MT-Bench](https://arxiv.org/abs/2306.05685) scores.
|
17 |
+
Our approach aims to mitigate forgetting in LLMs in a computationally efficient manner, thereby enabling continual fine-tuning capabilities without altering the pre-trained output distribution.
|
18 |
+
TenyxChat-8x7B-v1 was trained using eight A100s (80GB) for about eight hours, with a training setup obtained from HuggingFaceH4 ([GitHub](https://github.com/huggingface/alignment-handbook)).
|
19 |
|
20 |
# Model details
|
21 |
|
22 |
- Model type: Fine-tuned Mixture Of Expert 8x7B model for chat.
|
23 |
- License: Apache 2.0
|
24 |
+
- Base model: [Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1)
|
25 |
- Demo: [spaces/tenyx/TenyxChat-8x7B-v1](https://huggingface.co/spaces/tenyx/TenyxChat-8x7B-v1)
|
26 |
|
27 |
## Usage
|
|
|
63 |
|
64 |
# Performance
|
65 |
|
66 |
+
At the time of release (Jan 2024), TenyxChat-8x7B-v1 is the highest-ranked model, only superseded by GPT4, on the MT-Bench evaluation available for download and commercial use.
|
67 |
|
68 |
## MT-Bench
|
69 |
|