Update README.md
Browse files
README.md
CHANGED
@@ -13,12 +13,11 @@ library_name: transformers
|
|
13 |
<!-- Provide a quick summary of what the model is/does. -->
|
14 |
|
15 |
AI Squared's `dlite-v1-124m` ([blog post](https://medium.com/ai-squared/introducing-dlite-a-lightweight-chatgpt-like-model-based-on-dolly-deaa49402a1f)) is a large language
|
16 |
-
model which is derived from OpenAI's smallest [GPT-2](https://huggingface.co/gpt2) model and fine-tuned on a single T4 GPU on a
|
17 |
([Stanford Alpaca](https://crfm.stanford.edu/2023/03/13/alpaca.html)) to help it exhibit chat-based capabilities.
|
18 |
|
19 |
While `dlite-v1-124m` is **not a state-of-the-art model**, we believe that the level of interactivity that can be achieved on such a small model that is trained so cheaply
|
20 |
-
is important to showcase, as it continues to demonstrate that creating powerful AI capabilities
|
21 |
-
## Model Details
|
22 |
|
23 |
|
24 |
### Model Description
|
@@ -45,11 +44,16 @@ Just as with any other LLM, we advise users of this technology to exercise good
|
|
45 |
|
46 |
## Usage
|
47 |
|
|
|
|
|
|
|
48 |
### Load Model and Tokenizer from this Repository Using the `transformers` Package
|
49 |
|
50 |
```python
|
51 |
-
from
|
|
|
52 |
model_id = 'aisquared/dlite-v1-124m'
|
|
|
53 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
54 |
model = AutoModelForCausalLM.from_pretrained(model_id)
|
55 |
```
|
|
|
13 |
<!-- Provide a quick summary of what the model is/does. -->
|
14 |
|
15 |
AI Squared's `dlite-v1-124m` ([blog post](https://medium.com/ai-squared/introducing-dlite-a-lightweight-chatgpt-like-model-based-on-dolly-deaa49402a1f)) is a large language
|
16 |
+
model which is derived from OpenAI's smallest [GPT-2](https://huggingface.co/gpt2) model and fine-tuned on a single T4 GPU on a corpus of 50k records
|
17 |
([Stanford Alpaca](https://crfm.stanford.edu/2023/03/13/alpaca.html)) to help it exhibit chat-based capabilities.
|
18 |
|
19 |
While `dlite-v1-124m` is **not a state-of-the-art model**, we believe that the level of interactivity that can be achieved on such a small model that is trained so cheaply
|
20 |
+
is important to showcase, as it continues to demonstrate that creating powerful AI capabilities may be much more accessible than previously thought.
|
|
|
21 |
|
22 |
|
23 |
### Model Description
|
|
|
44 |
|
45 |
## Usage
|
46 |
|
47 |
+
The code below shows how to use `dlite-v1-124m` in the way which it was trained. While the model can be used "out of the box" using the
|
48 |
+
`transformers` library, using the function defined below to create a response from the model will achieve better results.
|
49 |
+
|
50 |
### Load Model and Tokenizer from this Repository Using the `transformers` Package
|
51 |
|
52 |
```python
|
53 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
54 |
+
|
55 |
model_id = 'aisquared/dlite-v1-124m'
|
56 |
+
|
57 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
58 |
model = AutoModelForCausalLM.from_pretrained(model_id)
|
59 |
```
|