license: bigscience-bloom-rail-1.0 | |
language: | |
- en | |
pipeline_tag: text-generation | |
tags: | |
- nlp | |
- code | |
## Model Summary | |
bloomz-3b-mmlu-lora is a LORA model which fine-tuned on mmlu dataset. The base model is `bigscience/bloomz-3b` . | |
## How to Use | |
```python | |
import torch | |
from transformers import AutoTokenizer | |
from peft import AutoPeftModelForCausalLM | |
torch.set_default_device("cuda") | |
model = AutoPeftModelForCausalLM.from_pretrained("liuchanghf/bloomz-3b-mmlu-lora") | |
tokenizer = AutoTokenizer.from_pretrained("bigscience/bloomz-3b") | |
inputs = tokenizer('''def print_prime(n): | |
""" | |
Print all primes between 1 and n | |
"""''', return_tensors="pt", return_attention_mask=False) | |
outputs = model.generate(**inputs, max_length=200) | |
text = tokenizer.batch_decode(outputs)[0] | |
print(text) | |
``` |