Update README.md
Browse files
README.md
CHANGED
@@ -46,7 +46,7 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
46 |
)
|
47 |
adapters.init(model)
|
48 |
|
49 |
-
adapter_name = model.load_adapter(adapter_id, set_active=True)
|
50 |
|
51 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
52 |
```
|
@@ -76,7 +76,8 @@ def prompt_model(model, text: str):
|
|
76 |
output_tokens = model.generate(**batch, stopping_criteria=[EosListStoppingCriteria()])
|
77 |
|
78 |
# skip prompt when decoding
|
79 |
-
|
|
|
80 |
```
|
81 |
|
82 |
Now, to prompt the model:
|
|
|
46 |
)
|
47 |
adapters.init(model)
|
48 |
|
49 |
+
adapter_name = model.load_adapter(adapter_id, source="hf", set_active=True)
|
50 |
|
51 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
52 |
```
|
|
|
76 |
output_tokens = model.generate(**batch, stopping_criteria=[EosListStoppingCriteria()])
|
77 |
|
78 |
# skip prompt when decoding
|
79 |
+
decoded = tokenizer.decode(output_tokens[0, batch["input_ids"].shape[1]:], skip_special_tokens=True)
|
80 |
+
return decoded[:-10] if decoded.endswith("### Human:") else decoded
|
81 |
```
|
82 |
|
83 |
Now, to prompt the model:
|