--- license: apache-2.0 language: - en --- GPT-J for preference modeling - Dataset: https://huggingface.co/datasets/reciprocate/oasst_hh_shp_hellaswag_webgpt_rm_dataset - Logs: https://wandb.ai/sorry/autocrit/runs/y9des5kz?workspace=user-sorry Usage: ```python from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("reciprocate/gpt-j_rm_format-oa", revision="501f895") model = AutoModelForSequenceClassification.from_pretrained("reciprocate/gpt-j_rm_format-oa", revision="501f895") input = "<|prompter|>Are you lying right now?<|assistant|>I am :)<|endoftext|>" model(**tokenizer(input, return_tensors="pt"))[0].item() ``` Output: ```python -4.877448558807373 ```