metadata
license: apache-2.0
datasets:
- amphora/QwQ-LongCoT-130K
language:
- en
base_model:
- prithivMLmods/QwQ-LCoT-7B-Instruct
pipeline_tag: text-generation
library_name: transformers
tags:
- QwQ
- 4B
- Adapter
- safetensors
- 4bit
- Qwen2.5
- text-generation-inference
QwQ-4B-Instruct-Model-Files
The QwQ-4B-Instruct is a lightweight and efficient fine-tuned language model for instruction-following tasks and reasoning. It is based on a quantized version of the Qwen2.5-7B model, optimized for inference speed and reduced memory consumption, while retaining robust capabilities for complex tasks.
File Name | Size | Description | Upload Status |
---|---|---|---|
.gitattributes |
1.57 kB | Tracks files stored with Git LFS. | Uploaded |
README.md |
271 Bytes | Basic project documentation. | Updated |
added_tokens.json |
657 Bytes | Specifies additional tokens for the tokenizer. | Uploaded |
config.json |
1.26 kB | Detailed model configuration file. | Uploaded |
generation_config.json |
281 Bytes | Configuration for text generation settings. | Uploaded |
merges.txt |
1.82 MB | Byte pair encoding (BPE) merge rules for tokenizer. | Uploaded |
model-00001-of-00002.safetensors |
4.46 GB | Part 1 of the model weights in safetensors format. | Uploaded (LFS) |
model-00002-of-00002.safetensors |
1.09 GB | Part 2 of the model weights in safetensors format. | Uploaded (LFS) |
model.safetensors.index.json |
124 kB | Index file for safetensors model sharding. | Uploaded |
special_tokens_map.json |
644 Bytes | Mapping of special tokens (e.g., , ). | Uploaded |
tokenizer.json |
11.4 MB | Complete tokenizer configuration. | Uploaded (LFS) |
tokenizer_config.json |
7.73 kB | Settings for the tokenizer integration. | Uploaded |
vocab.json |
2.78 MB | Vocabulary file containing token-to-id mappings. | Uploaded |
Key Features:
Model Size:
- 4.46B parameters.
Precision Support:
- Available in multiple tensor types:
- FP16
- F32
- U8 (Quantized)
- Available in multiple tensor types:
Model Sharding:
- The model weights are stored in two parts for efficient download:
model-00001-of-00002.safetensors
(4.46 GB)model-00002-of-00002.safetensors
(1.09 GB)
- Indexed with
model.safetensors.index.json
.
- The model weights are stored in two parts for efficient download:
Tokenizer:
- Uses Byte-Pair Encoding (BPE).
- Includes:
vocab.json
(2.78 MB)merges.txt
(1.82 MB)tokenizer.json
(11.4 MB, pre-trained configuration).
- Special tokens mapped in
special_tokens_map.json
(e.g.,<pad>
,<eos>
).
Configuration Files:
config.json
: Defines the architecture, hyperparameters, and settings.generation_config.json
: Specifies text generation behavior (e.g., max length, temperature).
Training Dataset:
- Dataset Name: amphora/QwQ-LongCoT-130K
- Size: 133k examples.
- Focus: Chain-of-Thought reasoning for detailed and logical outputs.
Use Cases:
Instruction-Following:
- Excels in handling concise and multi-step instructions.
Reasoning:
- Well-suited for tasks requiring logical deductions and detailed explanations.
Text Generation:
- Generates coherent and contextually aware responses across various domains.
Resource-Constrained Applications:
- Optimized for scenarios requiring lower computational resources due to its smaller model size and quantization.