Edit model card

ibleducation/ibl-tutoring-7B-32k

ibleducation/ibl-tutoring-7B-32k is a tutor model that is fine-tuned against amazon/Mistrallite. The model is trained to respond like a professional teacher.

Example Conversations

  1. Question: I'm having trouble understanding this art movement. Can you help me?
    Answer:

    Absolutely! Let's explore the movement together. I'm here to support you. We can analyze the artists' techniques, their motivations, and find creative ways to connect with the movement. I'm here to provide guidance and encouragement.  Remember, it's important to be open to different perspectives and interpretations. Let's have fun exploring the art movement!
    
  2. Question: I am thinking of stopping school do you think it is right?
    Answer:

    I understand that school can be challenging, but I believe that education is important. Let's work together to find ways to make school more enjoyable and less stressful. I'm here to support you
    
  3. Question: How do I analyze algebraic expressions?
    Answer:

    Analyzing algebraic expressions can be challenging, But with patience and practice, you'll become proficient! Let's start by simplifying the expressions and identifying the terms. I'll be patient in guiding you through the steps and explaining the reasoning behind each simplification. We'll conquer this together. Remember, I'm here to support you and help you develop a deeper understanding of algebraic expressions. If you have any doubts or need further clarification, feel free to ask. My mission is to help you reach your goal of becoming proficient in analyzing algebraic expressions. Let's start by simplifying the expression 3x + 5y - 2x + 3y. We can combine like terms and simplify it to 5x + 3y. By continually adding tools to your toolbox, such as simplifying and combining like terms, you'll become a proficient analyzer of algebraic expressions. I'm here to provide guidance and support along the way. I'm excited to see your progress!
    

Motivation of Developing ibl-tutoring-7B-32k

Students today use llm's in their learning and research. However, most models are not trained to behave and respond to conversations with the virtues a teacher must possess. ibl-tutoring-7B-32k is fine tuned on top of amazon/Mistrallite to alter its behaviour to converse the way a teacher should Mistrallite was chosen because its it's performance when compared to Mistral-7B-instruct especially with reference to token length

Model Details

How to Use ibl-tutoring-7B-32k from Python Code (HuggingFace transformers)

Install the necessary packages

Requires: transformers 4.34.0 or later, flash-attn 2.3.1.post1 or later, and accelerate 0.23.0 or later.

pip install transformers==4.34.0
pip install flash-attn==2.3.1.post1 --no-build-isolation
pip install accelerate==0.23.0

You can then try the following example code

from transformers import AutoModelForCausalLM, AutoTokenizer
import transformers
import torch

model_id = "ibleducation/ibl-tutoring-7B-32k"

tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id,
                                             torch_dtype=torch.bfloat16,
                                             use_flash_attention_2=True,
                                             device_map="auto",)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    tokenizer=tokenizer,
)
prompt = "<|prompter|>What makes a good teacher?</s><|assistant|>"

sequences = pipeline(
    prompt,
    max_new_tokens=400,
    do_sample=False,
    return_full_text=False,
    num_return_sequences=1,
    eos_token_id=tokenizer.eos_token_id,
)
for seq in sequences:
    print(f"{seq['generated_text']}")

Important - Use the prompt template below for ibl-tutoring-7B-32k:

<|prompter|>{prompt}</s><|assistant|>
Downloads last month
32
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train iblai/ibl-tutoring-7B-32k

Evaluation results