Update README.md
Browse files
README.md
CHANGED
@@ -32,7 +32,7 @@ Ladybird-base-7B-v8 is based on the Mistral architecture, which is known for its
|
|
32 |
- **Grouped-Query Attention**: Optimizes attention mechanisms by grouping queries, reducing computational complexity while maintaining model quality.
|
33 |
- **Sliding-Window Attention**: Improves the model's ability to handle long-range dependencies by focusing on relevant segments of input, enhancing understanding and coherence.
|
34 |
- **Byte-fallback BPE Tokenizer**: Offers robust tokenization by combining the effectiveness of Byte-Pair Encoding (BPE) with a fallback mechanism for out-of-vocabulary bytes, ensuring comprehensive language coverage.
|
35 |
-
|
36 |
## Prompt Template
|
37 |
|
38 |
The prompt template is ChatML.
|
|
|
32 |
- **Grouped-Query Attention**: Optimizes attention mechanisms by grouping queries, reducing computational complexity while maintaining model quality.
|
33 |
- **Sliding-Window Attention**: Improves the model's ability to handle long-range dependencies by focusing on relevant segments of input, enhancing understanding and coherence.
|
34 |
- **Byte-fallback BPE Tokenizer**: Offers robust tokenization by combining the effectiveness of Byte-Pair Encoding (BPE) with a fallback mechanism for out-of-vocabulary bytes, ensuring comprehensive language coverage.
|
35 |
+
|
36 |
## Prompt Template
|
37 |
|
38 |
The prompt template is ChatML.
|