--- language: en thumbnail: https://example.com/thumbnail.png tags: - paraphrasing - T5 - text generation - NLP - transformers license: mit datasets: - mteb/quora metrics: - accuracy base_model: - humarin/chatgpt_paraphraser_on_T5_base library_name: transformers --- # ChatGPT and T5 Base Paraphraser This model is a fine-tuned version of the T5 transformer model designed for paraphrasing questions using the ChatGPT architecture. ## Model Description The `chat_gpt_and_t5_base_paraphraser` model is trained to generate paraphrased versions of input questions by utilizing a sequence-to-sequence approach. The model leverages the T5 architecture and has been fine-tuned on the Quora Question-Answer dataset to improve its ability to create diverse and meaningful paraphrases. ## Intended Use This model is intended for applications where paraphrasing of text is required, such as: - Chatbots - Question-answering systems - Content generation - Educational tools ## How to Use To use the model, install the Hugging Face `transformers` library and follow these steps: ```python from transformers import AutoTokenizer, AutoModelForSeq2SeqLM # Load the model and tokenizer model_name = "jaesani/chat_gpt_and_t5_base_paraphraser" tokenizer = AutoTokenizer.from_pretrained(model_name) model = AutoModelForSeq2SeqLM.from_pretrained(model_name) def paraphrase(question, max_length=128): input_ids = tokenizer(f'paraphrase: {question}', return_tensors="pt", padding="longest", max_length=max_length, truncation=True).input_ids outputs = model.generate(input_ids, max_length=max_length) return tokenizer.decode(outputs[0], skip_special_tokens=True) # Example usage paraphrased_text = paraphrase("What are the best places to see in New York?") print(paraphrased_text) ``` ## Training Data The model was fine-tuned using the Quora Question-Answer Dataset, which consists of pairs of questions that may or may not be paraphrases of each other. ## Evaluation The model's performance can be evaluated based on the diversity and coherence of the paraphrases it generates. Specific metrics can include BLEU scores and human evaluations for semantic similarity. ## Limitations The model may produce paraphrases that are not contextually relevant. It may struggle with highly technical or domain-specific language. Generated paraphrases might be similar for closely related input questions. ## License This model is licensed under MIT License.