Edit model card

PyCodeGen 350M

This model is finetuned version of codegen-350M-mono by Salesforce trained on python code dataset using QLORA method.

Pretrained model description

codegen-350M-mono

Codegen-350M-mono comes from the family of autoregressive models for program synthesis developed by Salesforce. This model was first trained on ThePile dataset which is 825.18 GiB English text corpus. It was then adapted to generate code by training on a set of GitQuery with source codes. Finally model has been adapted to the Python language by training on the BigPython dataset.

Training Data

python_code_instructions_18k_alpaca

The dataset contains problem descriptions and code in python language. This dataset is taken from sahil2801/code_instructions_120k, which adds a prompt column in alpaca style.

Intended uses

The model can be used to generate python code that solves task with optionally given input data.

Example of usage

from transformers import AutoModelForCausalLM, AutoTokenizer


model = AutoModelForCausalLM.from_pretrained('chincyk/PyCodeGen')
tokenizer = AutoTokenizer.from_pretrained('chincyk/PyCodeGen')

instruction = "Write a python class that represents a calculator, then use it to add two numbers."
input = "a = 5, b = 2"

prompt = f"""
    ### Instruction:
    Use the Task below and the Input given to write the Response, which is a programming code that can solve the Task.

    ### Task:
    {instruction}

    ### Input:
    {input}
    
    ### Response:
    """

input_ids = tokenizer(prompt, truncation=True, return_tensors="pt")['input_ids']
output = model.generate(input_ids=input_ids, max_length=200)

print(tokenizer.decode(output[0], skip_special_tokens=True))

Training parameters

BitsAndBytes:

  • load_in_4bit: True,
  • bnb_4bit_quant_type: nf4,
  • bnb_4bit_use_double_quant: True,
  • bnb_4bit_compute_dtype: torch.bfloat16

LoraConfig:

  • r: 32,
  • lora_alpha: 16,
  • target_modules: all-linear,
  • lora_dropout: 0.1,
  • bias: none,
  • task_type: CASUAL_LM

Finetuning:

  • num_epochs: 15
  • train_batch_size: 4
  • eval_batch_size: 8
  • gradient_accumulation_steps: 8
  • learning_rate: 3e-4
  • weight_decay: 0.01
  • lr_scheduler_name: cosine
  • num_warmup_steps: 190
Downloads last month
10
Safetensors
Model size
357M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train chincyk/PyCodeGen