Librarian Bot: Add moe tag to model
Browse filesThis pull request aims to enrich the metadata of your model by adding an `moe` (Mixture of Experts) `tag` in the `YAML` block of your model's `README.md`.
How did we find this information? We infered that this model is a `moe` model based on the following criteria:
- The model's name contains the string `moe`.
- The model indicates it uses a `moe` architecture
- The model's base model is a `moe` model
**Why add this?** Enhancing your model's metadata in this way:
- **Boosts Discoverability** - It becomes easier to find mixture of experts models on the Hub
- **Helping understand the ecosystem** - It becomes easier to understand the ecosystem of mixture of experts models on the Hub and how they are used
This PR comes courtesy of [Librarian Bot](https://huggingface.co/librarian-bot). If you have any feedback, queries, or need assistance, please don't hesitate to reach out to [@davanstrien](https://huggingface.co/davanstrien).
@@ -1,16 +1,17 @@
|
|
1 |
---
|
2 |
-
inference: false
|
3 |
language:
|
4 |
- en
|
5 |
-
library_name: transformers
|
6 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
7 |
model_name: Mixtral 8X7B - bnb 4-bit
|
|
|
8 |
model_type: mixtral
|
9 |
pipeline_tag: text-generation
|
10 |
quantized_by: ybelkada
|
11 |
-
tags:
|
12 |
-
- mistral
|
13 |
-
- mixtral
|
14 |
---
|
15 |
|
16 |
# Mixtral 8x7B Instruct-v0.1 - `bitsandbytes` 4-bit
|
|
|
1 |
---
|
|
|
2 |
language:
|
3 |
- en
|
|
|
4 |
license: apache-2.0
|
5 |
+
library_name: transformers
|
6 |
+
tags:
|
7 |
+
- mistral
|
8 |
+
- mixtral
|
9 |
+
- moe
|
10 |
model_name: Mixtral 8X7B - bnb 4-bit
|
11 |
+
inference: false
|
12 |
model_type: mixtral
|
13 |
pipeline_tag: text-generation
|
14 |
quantized_by: ybelkada
|
|
|
|
|
|
|
15 |
---
|
16 |
|
17 |
# Mixtral 8x7B Instruct-v0.1 - `bitsandbytes` 4-bit
|