shouryashashank
commited on
Commit
•
224cb5c
1
Parent(s):
6da0f58
Update README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,7 @@ pipeline_tag: text-generation
|
|
8 |
---
|
9 |
# Model Card for Model ID
|
10 |
|
11 |
-
The model `Precacons/ReasonGPT-
|
12 |
|
13 |
|
14 |
## Model Details
|
@@ -47,7 +47,7 @@ The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based
|
|
47 |
|
48 |
### Limitations
|
49 |
|
50 |
-
**ReasonGPT-
|
51 |
|
52 |
1. **Calculation Accuracy**:
|
53 |
- Due to its small size, the model may not perform complex calculations with high accuracy. It is optimized for reasoning and explanations rather than precise numerical computations.
|
@@ -59,7 +59,7 @@ The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based
|
|
59 |
- With a smaller parameter size, the model may have limitations in understanding and generating contextually rich and nuanced responses compared to larger models.
|
60 |
|
61 |
4. **Bias and Fairness**:
|
62 |
-
- Like all language models, ReasonGPT-
|
63 |
|
64 |
5. **Resource Constraints**:
|
65 |
- While the model is designed to be efficient, it still requires a GPU for optimal performance. Users with limited computational resources may experience slower inference times.
|
@@ -70,7 +70,7 @@ The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based
|
|
70 |
import predacons
|
71 |
|
72 |
# Load the model and tokenizer
|
73 |
-
model_path = "ReasonGPT-
|
74 |
model = predacons.load_model(model_path = model_path)
|
75 |
tokenizer = predacons.load_tokenizer(model_path)
|
76 |
|
@@ -87,7 +87,7 @@ generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
|
87 |
print(generated_text)
|
88 |
```
|
89 |
|
90 |
-
This example demonstrates how to load the `ReasonGPT-
|
91 |
|
92 |
|
93 |
|
|
|
8 |
---
|
9 |
# Model Card for Model ID
|
10 |
|
11 |
+
The model `Precacons/ReasonGPT-2B-4bit` is a lightweight language model based on the GEMMA architecture. It is designed to provide reasoning and explanations for any given problem. Despite its powerful capabilities, it is very compact, with a size of just 2.16 GB, making it efficient for deployment and use in various applications.
|
12 |
|
13 |
|
14 |
## Model Details
|
|
|
47 |
|
48 |
### Limitations
|
49 |
|
50 |
+
**ReasonGPT-2B-4bit** is a compact model designed for efficiency, but it comes with certain limitations:
|
51 |
|
52 |
1. **Calculation Accuracy**:
|
53 |
- Due to its small size, the model may not perform complex calculations with high accuracy. It is optimized for reasoning and explanations rather than precise numerical computations.
|
|
|
59 |
- With a smaller parameter size, the model may have limitations in understanding and generating contextually rich and nuanced responses compared to larger models.
|
60 |
|
61 |
4. **Bias and Fairness**:
|
62 |
+
- Like all language models, ReasonGPT-2B-4bit may exhibit biases present in the training data. Users should be cautious of potential biases in the generated outputs.
|
63 |
|
64 |
5. **Resource Constraints**:
|
65 |
- While the model is designed to be efficient, it still requires a GPU for optimal performance. Users with limited computational resources may experience slower inference times.
|
|
|
70 |
import predacons
|
71 |
|
72 |
# Load the model and tokenizer
|
73 |
+
model_path = "ReasonGPT-2B-4bit"
|
74 |
model = predacons.load_model(model_path = model_path)
|
75 |
tokenizer = predacons.load_tokenizer(model_path)
|
76 |
|
|
|
87 |
print(generated_text)
|
88 |
```
|
89 |
|
90 |
+
This example demonstrates how to load the `ReasonGPT-2B-4bit` model and use it to generate an explanation for a given query, keeping in mind the limitations mentioned above.
|
91 |
|
92 |
|
93 |
|