--- base_model: bobofrut/ladybird-base-7B-v8 license: apache-2.0 language: - en pipeline_tag: text-generation tags: - mistral - gguf library_name: llama.cpp model_creator: bobofrut model_name: ladybird base 7B v8 model_type: mistral prompt_template: | <|im_start|>system {system_message}<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant quantized_by: mgonzs13 --- # ladybird-base-7B-v8 **Model creator:** [bobofrut](https://huggingface.co/bobofrut)
**Original model**: [Mistroll-7B-v2.2](https://huggingface.co/bobofrut/ladybird-base-7B-v8)
**GGUF quantization:** `llama.cpp` commit [b8c1476e44cc1f3a1811613f65251cf779067636](https://github.com/ggerganov/llama.cpp/tree/b8c1476e44cc1f3a1811613f65251cf779067636)
## Description Ladybird-base-7B-v8 is based on the Mistral architecture, which is known for its efficiency and effectiveness in handling complex language understanding and generation tasks. The model incorporates several innovative architecture choices to enhance its performance: - **Grouped-Query Attention**: Optimizes attention mechanisms by grouping queries, reducing computational complexity while maintaining model quality. - **Sliding-Window Attention**: Improves the model's ability to handle long-range dependencies by focusing on relevant segments of input, enhancing understanding and coherence. - **Byte-fallback BPE Tokenizer**: Offers robust tokenization by combining the effectiveness of Byte-Pair Encoding (BPE) with a fallback mechanism for out-of-vocabulary bytes, ensuring comprehensive language coverage. ## Prompt Template The prompt template is ChatML. ``` <|im_start|>system {system_message}<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant ```