Model Card for Model ID
This model was trained to choose between RAG and COT techniques for especific domain chat aplications. Depending on users questions, the model may choose what is the best way to generate the response.
- Sometimes, questions are domain specific and can be answered by performing a simple RAG.
- Sometimes, we may get complex questions that require a step by step approach.
We performed a simple prompt tunning over a low-parameters base model so that we can create a basic low parameter model capable of few-shot classification with really low dataset of nearly ~100 samples.
Base Model Sources
Prompt tunned version from bigscience/bloom-560m on a bnb configuration of 4bits.
Uses
This model aims to start to perform a especific task by choosing Retrieval Augmented Generation-RAG or Chain of Thought-COT
How to Get Started with the Model
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
def make_inference(query, model):
prompt = """\
### Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
Categorize this question into one of this two categories:
RAG
COT
Input:
{Question}
### Response:
"""
batch = tokenizer(prompt.format(Question=query), return_tensors='pt').to("cuda")
with torch.cuda.amp.autocast():
output_tokens = model.generate(**batch, max_new_tokens=10)
return output_tokens
query = "{your_question_goes_here}"
output_tokens = make_inference(query, model)
response = tokenizer.decode(output_tokens[0])
print(response)
Training Details
Training Data
The dataset used is a sinthetic dataset that contains pairs, values of quentions, techniques.
Training Prompt
### Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
Categorize this question into one of this two categories:
RAG
COT
Input:
{Question}
### Response:
{Category}
### End
"""
Training Hyperparameters
- evaluation_strategy="steps",
- eval_steps=1,
- logging_strategy="steps",
- per_device_train_batch_size=6,
- gradient_accumulation_steps=4,
- warmup_steps=50,
- max_steps=100,
- learning_rate=1e-3,
- fp16=True,
- logging_steps=1,
Evaluation
Metrics
- Accuracy
Results
Train:
Validation:
Test:
Model Card Contact
Linkedin: www.linkedin.com/in/jrodriguez130
- Downloads last month
- 11
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.