--- license: apache-2.0 base_model: nidum/Nidum-Llama-3.2-3B-Uncensored library_name: adapter-transformers tags: - chemistry - biology - legal - code - medical - finance - mlx pipeline_tag: text-generation --- ### **Nidum-Llama-3.2-3B-Uncensored-MLX-8bit** ### **Welcome to Nidum!** At Nidum, our mission is to bring cutting-edge AI capabilities to everyone with unrestricted access to innovation. With **Nidum-Llama-3.2-3B-Uncensored-MLX-8bit**, you get an optimized, efficient, and versatile AI model for diverse applications. --- [![GitHub Icon](https://upload.wikimedia.org/wikipedia/commons/thumb/9/95/Font_Awesome_5_brands_github.svg/232px-Font_Awesome_5_brands_github.svg.png)](https://github.com/NidumAI-Inc) **Discover Nidum's Open-Source Projects on GitHub**: [https://github.com/NidumAI-Inc](https://github.com/NidumAI-Inc) --- ### **Key Features** 1. **Efficient and Compact**: Developed in **MLX-8bit format** for improved performance and reduced memory demands. 2. **Wide Applicability**: Suitable for technical problem-solving, educational content, and conversational tasks. 3. **Advanced Context Awareness**: Handles long-context conversations with exceptional coherence. 4. **Streamlined Integration**: Optimized for use with the **mlx-lm library** for effortless development. 5. **Unrestricted Responses**: Offers uncensored answers across all supported domains. --- ### **How to Use** To use **Nidum-Llama-3.2-3B-Uncensored-MLX-8bit**, install the **mlx-lm** library and follow these steps: #### **Installation** ```bash pip install mlx-lm ``` #### **Usage** ```python from mlx_lm import load, generate # Load the model and tokenizer model, tokenizer = load("nidum/Nidum-Llama-3.2-3B-Uncensored-MLX-8bit") # Create a prompt prompt = "hello" # Apply the chat template if available if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None: messages = [{"role": "user", "content": prompt}] prompt = tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) # Generate the response response = generate(model, tokenizer, prompt=prompt, verbose=True) # Print the response print(response) ``` --- ### **About the Model** The **nidum/Nidum-Llama-3.2-3B-Uncensored-MLX-8bit** model, converted using **mlx-lm version 0.19.2**, brings: - **Memory Efficiency**: Tailored for systems with limited hardware. - **Performance Optimization**: Matches the capabilities of the original model while delivering faster inference. - **Plug-and-Play**: Easily integrates with the **mlx-lm** library for deployment ease. --- ### **Use Cases** - **Problem Solving in Tech and Science** - **Educational and Research Assistance** - **Creative Writing and Brainstorming** - **Extended Dialogues** - **Uninhibited Knowledge Exploration** --- ### **Datasets and Fine-Tuning** Derived from **Nidum-Llama-3.2-3B-Uncensored**, the MLX-8bit version inherits: - **Uncensored Fine-Tuning**: Delivers detailed and open-ended responses. - **RAG-Based Optimization**: Enhances retrieval-augmented generation for data-driven tasks. - **Math Reasoning Support**: Precise mathematical computations and explanations. - **Long-Context Training**: Ensures relevance and coherence in extended conversations. --- ### **Quantized Model Download** The **MLX-8bit** format strikes the perfect balance between memory optimization and performance. --- #### **Benchmark** | **Benchmark** | **Metric** | **LLaMA 3B** | **Nidum 3B** | **Observation** | |-------------------|-----------------------------------|--------------|--------------|-----------------------------------------------------------------------------------------------------| | **GPQA** | Exact Match (Flexible) | 0.3 | 0.5 | Nidum 3B achieves notable improvement in **generative tasks**. | | | Accuracy | 0.4 | 0.5 | Demonstrates strong performance, especially in **zero-shot** tasks. | | **HellaSwag** | Accuracy | 0.3 | 0.4 | Excels in **common-sense reasoning** tasks. | | | Normalized Accuracy | 0.3 | 0.4 | Strong contextual understanding in sentence completion tasks. | | | Normalized Accuracy (Stderr) | 0.15275 | 0.1633 | Enhanced consistency in normalized accuracy. | | | Accuracy (Stderr) | 0.15275 | 0.1633 | Demonstrates robustness in reasoning accuracy compared to LLaMA 3B. | --- ### **Insights** 1. **High Performance, Low Resource**: The MLX-8bit format is ideal for environments with limited memory and processing power. 2. **Seamless Integration**: Designed for smooth integration into lightweight systems and workflows. --- ### **Contributing** Join us in enhancing the **MLX-8bit** model's capabilities. Contact us for collaboration opportunities. --- ### **Contact** For questions, support, or feedback, email **info@nidum.ai**. --- ### **Experience the Future** Harness the power of **Nidum-Llama-3.2-3B-Uncensored-MLX-8bit** for a perfect blend of performance and efficiency. ---