rogkesavan's picture
Update README.md
38ef389 verified
|
raw
history blame
5.65 kB
---
license: apache-2.0
base_model: nidum/Nidum-Llama-3.2-3B-Uncensored
library_name: adapter-transformers
tags:
- chemistry
- biology
- legal
- code
- medical
- finance
- mlx
pipeline_tag: text-generation
---
### **Nidum-Llama-3.2-3B-Uncensored-MLX-8bit**
### **Welcome to Nidum!**
At Nidum, our mission is to bring cutting-edge AI capabilities to everyone with unrestricted access to innovation. With **Nidum-Llama-3.2-3B-Uncensored-MLX-8bit**, you get an optimized, efficient, and versatile AI model for diverse applications.
---
[![GitHub Icon](https://upload.wikimedia.org/wikipedia/commons/thumb/9/95/Font_Awesome_5_brands_github.svg/232px-Font_Awesome_5_brands_github.svg.png)](https://github.com/NidumAI-Inc)
**Discover Nidum's Open-Source Projects on GitHub**: [https://github.com/NidumAI-Inc](https://github.com/NidumAI-Inc)
---
### **Key Features**
1. **Efficient and Compact**: Developed in **MLX-8bit format** for improved performance and reduced memory demands.
2. **Wide Applicability**: Suitable for technical problem-solving, educational content, and conversational tasks.
3. **Advanced Context Awareness**: Handles long-context conversations with exceptional coherence.
4. **Streamlined Integration**: Optimized for use with the **mlx-lm library** for effortless development.
5. **Unrestricted Responses**: Offers uncensored answers across all supported domains.
---
### **How to Use**
To use **Nidum-Llama-3.2-3B-Uncensored-MLX-8bit**, install the **mlx-lm** library and follow these steps:
#### **Installation**
```bash
pip install mlx-lm
```
#### **Usage**
```python
from mlx_lm import load, generate
# Load the model and tokenizer
model, tokenizer = load("nidum/Nidum-Llama-3.2-3B-Uncensored-MLX-8bit")
# Create a prompt
prompt = "hello"
# Apply the chat template if available
if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None:
messages = [{"role": "user", "content": prompt}]
prompt = tokenizer.apply_chat_template(
messages, tokenize=False, add_generation_prompt=True
)
# Generate the response
response = generate(model, tokenizer, prompt=prompt, verbose=True)
# Print the response
print(response)
```
---
### **About the Model**
The **nidum/Nidum-Llama-3.2-3B-Uncensored-MLX-8bit** model, converted using **mlx-lm version 0.19.2**, brings:
- **Memory Efficiency**: Tailored for systems with limited hardware.
- **Performance Optimization**: Matches the capabilities of the original model while delivering faster inference.
- **Plug-and-Play**: Easily integrates with the **mlx-lm** library for deployment ease.
---
### **Use Cases**
- **Problem Solving in Tech and Science**
- **Educational and Research Assistance**
- **Creative Writing and Brainstorming**
- **Extended Dialogues**
- **Uninhibited Knowledge Exploration**
---
### **Datasets and Fine-Tuning**
Derived from **Nidum-Llama-3.2-3B-Uncensored**, the MLX-8bit version inherits:
- **Uncensored Fine-Tuning**: Delivers detailed and open-ended responses.
- **RAG-Based Optimization**: Enhances retrieval-augmented generation for data-driven tasks.
- **Math Reasoning Support**: Precise mathematical computations and explanations.
- **Long-Context Training**: Ensures relevance and coherence in extended conversations.
---
### **Quantized Model Download**
The **MLX-8bit** format strikes the perfect balance between memory optimization and performance.
---
#### **Benchmark**
| **Benchmark** | **Metric** | **LLaMA 3B** | **Nidum 3B** | **Observation** |
|-------------------|-----------------------------------|--------------|--------------|-----------------------------------------------------------------------------------------------------|
| **GPQA** | Exact Match (Flexible) | 0.3 | 0.5 | Nidum 3B achieves notable improvement in **generative tasks**. |
| | Accuracy | 0.4 | 0.5 | Demonstrates strong performance, especially in **zero-shot** tasks. |
| **HellaSwag** | Accuracy | 0.3 | 0.4 | Excels in **common-sense reasoning** tasks. |
| | Normalized Accuracy | 0.3 | 0.4 | Strong contextual understanding in sentence completion tasks. |
| | Normalized Accuracy (Stderr) | 0.15275 | 0.1633 | Enhanced consistency in normalized accuracy. |
| | Accuracy (Stderr) | 0.15275 | 0.1633 | Demonstrates robustness in reasoning accuracy compared to LLaMA 3B. |
---
### **Insights**
1. **High Performance, Low Resource**: The MLX-8bit format is ideal for environments with limited memory and processing power.
2. **Seamless Integration**: Designed for smooth integration into lightweight systems and workflows.
---
### **Contributing**
Join us in enhancing the **MLX-8bit** model's capabilities. Contact us for collaboration opportunities.
---
### **Contact**
For questions, support, or feedback, email **info@nidum.ai**.
---
### **Experience the Future**
Harness the power of **Nidum-Llama-3.2-3B-Uncensored-MLX-8bit** for a perfect blend of performance and efficiency.
---