Update README.md
Browse files
README.md
CHANGED
@@ -21,43 +21,39 @@ tags:
|
|
21 |
- gguf
|
22 |
---
|
23 |
|
24 |
-
#
|
25 |
-
|
26 |
-
##
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
|
31 |
-
- **
|
32 |
-
|
33 |
-
|
34 |
-
|
35 |
-
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
|
41 |
-
**Model
|
42 |
-
|
43 |
-
|
44 |
-
|
45 |
-
|
46 |
-
|
47 |
-
|
48 |
-
|
49 |
-
|
50 |
-
|
51 |
-
|
52 |
-
**Model
|
53 |
-
|
54 |
-
|
55 |
-
|
56 |
-
|
57 |
-
|
58 |
-
**
|
59 |
-
|
60 |
-
## **Ethical Considerations and Limitations**
|
61 |
-
LlaMa 3.2 1B is a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Llama 3.2’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of Llama 3.2, developers should perform safety testing and tuning tailored to their specific applications of the model.
|
62 |
-
|
63 |
-
Please see the Responsible Use Guide available at [https://ai.meta.com/llama/responsible-use-guide/](https://ai.meta.com/llama/responsible-use-guide/)
|
|
|
21 |
- gguf
|
22 |
---
|
23 |
|
24 |
+
# LLaMA 3.2 1B Instruct
|
25 |
+
|
26 |
+
## 1. **Model Title**
|
27 |
+
- **Name**: LLaMA 3.2 1B Instruct
|
28 |
+
- **Parameter Size**: 1B (1.23B)
|
29 |
+
|
30 |
+
## 2. **Quantization Information**
|
31 |
+
- **Available Formats**:
|
32 |
+
- **ggml-model-q8_0.gguf**: 8-bit quantization for resource efficiency and good performance.
|
33 |
+
- **ggml-model-f16.gguf**: Half-precision (16-bit) floating-point format for enhanced precision.
|
34 |
+
- **Quantization Library**: llama.cpp
|
35 |
+
- **Use Cases**: Recommended for tasks such as multilingual dialogue, text generation, and summarization.
|
36 |
+
|
37 |
+
## 3. **Model Brief**
|
38 |
+
LLaMA 3.2 1B Instruct is a multilingual instruction-tuned language model, optimized for various dialogue tasks. It has been trained on a diverse set of publicly available data and performs well on common NLP benchmarks. The model architecture leverages improved transformer optimizations, making it effective for both text-only and code tasks.
|
39 |
+
|
40 |
+
- **Purpose**: Multilingual dialogue generation and summarization.
|
41 |
+
- **Model Family**: LLaMA 3.2
|
42 |
+
- **Architecture**: Auto-regressive Transformer with Grouped-Query Attention (GQA)
|
43 |
+
- **Training Data**: A mix of publicly available multilingual data, covering up to 9T tokens.
|
44 |
+
- **Supported Languages**: English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
|
45 |
+
- **Release Date**: September 25, 2024
|
46 |
+
- **Context Length**: 128k tokens
|
47 |
+
- **Knowledge Cutoff**: December 2023
|
48 |
+
|
49 |
+
## 4. **Core Library Information**
|
50 |
+
- **Library**: llama.cpp
|
51 |
+
- *[Repository Link](https://github.com/ggerganov/llama.cpp)*
|
52 |
+
- **Model Base**: [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct)
|
53 |
+
|
54 |
+
## 5. **Safety and Responsible Use**
|
55 |
+
LLaMA 3.2 1B is designed with safety in mind but still carries inherent risks due to its generative nature. It may produce biased, harmful, or unpredictable responses, especially for less-tested languages or sensitive prompts.
|
56 |
+
|
57 |
+
- **Testing and Risk Assessment**: Initial testing has focused on English outputs, and coverage for other languages is ongoing.
|
58 |
+
- **Limitations**: As with most LLMs, LLaMA 3.2 may not fully adhere to user instructions or safety guidelines and might exhibit unexpected behavior.
|
59 |
+
- **Responsible Use Guidelines**: For deployment, thorough testing is advised to align outputs with application-specific safety requirements. Refer to the [Responsible Use Guide](https://ai.meta.com/llama/responsible-use-guide/) for more details.
|
|
|
|
|
|
|
|