shouryashashank commited on
Commit
224cb5c
1 Parent(s): 6da0f58

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -8,7 +8,7 @@ pipeline_tag: text-generation
8
  ---
9
  # Model Card for Model ID
10
 
11
- The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based on the GEMMA architecture. It is designed to provide reasoning and explanations for any given problem. Despite its powerful capabilities, it is very compact, with a size of just 2.16 GB, making it efficient for deployment and use in various applications.
12
 
13
 
14
  ## Model Details
@@ -47,7 +47,7 @@ The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based
47
 
48
  ### Limitations
49
 
50
- **ReasonGPT-2.5B-4bit** is a compact model designed for efficiency, but it comes with certain limitations:
51
 
52
  1. **Calculation Accuracy**:
53
  - Due to its small size, the model may not perform complex calculations with high accuracy. It is optimized for reasoning and explanations rather than precise numerical computations.
@@ -59,7 +59,7 @@ The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based
59
  - With a smaller parameter size, the model may have limitations in understanding and generating contextually rich and nuanced responses compared to larger models.
60
 
61
  4. **Bias and Fairness**:
62
- - Like all language models, ReasonGPT-2.5B-4bit may exhibit biases present in the training data. Users should be cautious of potential biases in the generated outputs.
63
 
64
  5. **Resource Constraints**:
65
  - While the model is designed to be efficient, it still requires a GPU for optimal performance. Users with limited computational resources may experience slower inference times.
@@ -70,7 +70,7 @@ The model `Precacons/ReasonGPT-2.5B-4bit` is a lightweight language model based
70
  import predacons
71
 
72
  # Load the model and tokenizer
73
- model_path = "ReasonGPT-2.5B-4bit"
74
  model = predacons.load_model(model_path = model_path)
75
  tokenizer = predacons.load_tokenizer(model_path)
76
 
@@ -87,7 +87,7 @@ generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
87
  print(generated_text)
88
  ```
89
 
90
- This example demonstrates how to load the `ReasonGPT-2.5B-4bit` model and use it to generate an explanation for a given query, keeping in mind the limitations mentioned above.
91
 
92
 
93
 
 
8
  ---
9
  # Model Card for Model ID
10
 
11
+ The model `Precacons/ReasonGPT-2B-4bit` is a lightweight language model based on the GEMMA architecture. It is designed to provide reasoning and explanations for any given problem. Despite its powerful capabilities, it is very compact, with a size of just 2.16 GB, making it efficient for deployment and use in various applications.
12
 
13
 
14
  ## Model Details
 
47
 
48
  ### Limitations
49
 
50
+ **ReasonGPT-2B-4bit** is a compact model designed for efficiency, but it comes with certain limitations:
51
 
52
  1. **Calculation Accuracy**:
53
  - Due to its small size, the model may not perform complex calculations with high accuracy. It is optimized for reasoning and explanations rather than precise numerical computations.
 
59
  - With a smaller parameter size, the model may have limitations in understanding and generating contextually rich and nuanced responses compared to larger models.
60
 
61
  4. **Bias and Fairness**:
62
+ - Like all language models, ReasonGPT-2B-4bit may exhibit biases present in the training data. Users should be cautious of potential biases in the generated outputs.
63
 
64
  5. **Resource Constraints**:
65
  - While the model is designed to be efficient, it still requires a GPU for optimal performance. Users with limited computational resources may experience slower inference times.
 
70
  import predacons
71
 
72
  # Load the model and tokenizer
73
+ model_path = "ReasonGPT-2B-4bit"
74
  model = predacons.load_model(model_path = model_path)
75
  tokenizer = predacons.load_tokenizer(model_path)
76
 
 
87
  print(generated_text)
88
  ```
89
 
90
+ This example demonstrates how to load the `ReasonGPT-2B-4bit` model and use it to generate an explanation for a given query, keeping in mind the limitations mentioned above.
91
 
92
 
93