mgonzs13 commited on
Commit
0bfa9e7
1 Parent(s): ed14edc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +43 -0
README.md CHANGED
@@ -1,3 +1,46 @@
1
  ---
 
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model: bobofrut/ladybird-base-7B-v8
3
  license: apache-2.0
4
+ language:
5
+ - en
6
+ pipeline_tag: text-generation
7
+ tags:
8
+ - gguf
9
+ library_name: llama.cpp
10
+ model_creator: bobofrut
11
+ model_name: ladybird base 7B v8
12
+ model_type: mistral
13
+ prompt_template: |
14
+ <|im_start|>system
15
+ {system_message}<|im_end|>
16
+ <|im_start|>user
17
+ {prompt}<|im_end|>
18
+ <|im_start|>assistant
19
+ quantized_by: mgonzs13
20
  ---
21
+
22
+ # ladybird-base-7B-v8
23
+
24
+ **Model creator:** [bobofrut](https://huggingface.co/bobofrut)<br>
25
+ **Original model**: [Mistroll-7B-v2.2](https://huggingface.co/bobofrut/ladybird-base-7B-v8)<br>
26
+ **GGUF quantization:** `llama.cpp` commit [b8c1476e44cc1f3a1811613f65251cf779067636](https://github.com/ggerganov/llama.cpp/tree/b8c1476e44cc1f3a1811613f65251cf779067636)<br>
27
+
28
+ ## Description
29
+
30
+ Ladybird-base-7B-v8 is based on the Mistral architecture, which is known for its efficiency and effectiveness in handling complex language understanding and generation tasks. The model incorporates several innovative architecture choices to enhance its performance:
31
+
32
+ - **Grouped-Query Attention**: Optimizes attention mechanisms by grouping queries, reducing computational complexity while maintaining model quality.
33
+ - **Sliding-Window Attention**: Improves the model's ability to handle long-range dependencies by focusing on relevant segments of input, enhancing understanding and coherence.
34
+ - **Byte-fallback BPE Tokenizer**: Offers robust tokenization by combining the effectiveness of Byte-Pair Encoding (BPE) with a fallback mechanism for out-of-vocabulary bytes, ensuring comprehensive language coverage.
35
+ -
36
+ ## Prompt Template
37
+
38
+ The prompt template is ChatML.
39
+
40
+ ```
41
+ <|im_start|>system
42
+ {system_message}<|im_end|>
43
+ <|im_start|>user
44
+ {prompt}<|im_end|>
45
+ <|im_start|>assistant
46
+ ```