Text Generation
Transformers
English
llama
Generated from Trainer
8-bit precision
bitsandbytes

Built with Axolotl

no-robots-y34b-lora

This model is a Yi-34B-Llama training on the HuggingFaceH4/no_robots. It uses my converted dataset in ShareGPT format with a few minor corrections (https://huggingface.co/datasets/Doctor-Shotgun/no-robots-sharegpt).

The Yi-34B-Llama model is a modified 01-ai/Yi-34B with keys renamed to match those used in Llama models, eliminating the need for remote code and ensuring compatibility with existing training and inference repositories. Architecturally this is similar to a Llama 2 34B model with an expanded vocab size of 64000.

Model description

No Robots is a high-quality dataset of 10,000 instructions and demonstrations created by skilled human annotators. This data can be used for supervised fine-tuning (SFT) to make language models follow instructions better. No Robots was modelled after the instruction dataset described in OpenAI's InstructGPT paper, and is comprised mostly of single-turn instructions across the following categories:

Category Count
Generation 4560
Open QA 1240
Brainstorm 1120
Chat 850
Rewrite 660
Summarize 420
Coding 350
Classify 350
Closed QA 260
Extract 190

This lora was trained using a modified multi-turn Alpaca prompt format:

### Instruction:
Below is a message that describes a task. Write a response that appropriately completes the request.

### Input:
{human prompt}

### Response:
{bot response}

Some chat examples have alternate system prompts that differ from the default provided above.

Intended uses & limitations

The intended use is to add instruction-following capabilities to the base model based on curated human examples. Outputs may exhibit biases observed in the base model, and have not been filtered for explicit or harmful content and hallucinations.

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 8e-05
  • train_batch_size: 2
  • eval_batch_size: 2
  • seed: 42
  • gradient_accumulation_steps: 4
  • total_train_batch_size: 8
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: cosine
  • lr_scheduler_warmup_steps: 10
  • num_epochs: 3

Framework versions

  • Transformers 4.34.1
  • Pytorch 2.0.1+cu118
  • Datasets 2.14.6
  • Tokenizers 0.14.1

Citation data

@misc{no_robots,
  author = {Nazneen Rajani and Lewis Tunstall and Edward Beeching and Nathan Lambert and Alexander M. Rush and Thomas Wolf},
  title = {No Robots},
  year = {2023},
  publisher = {Hugging Face},
  journal = {Hugging Face repository},
  howpublished = {\url{https://huggingface.co/datasets/HuggingFaceH4/no_robots}}
}
Downloads last month
13
Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for Doctor-Shotgun/no-robots-y34b-lora

Quantizations
1 model

Datasets used to train Doctor-Shotgun/no-robots-y34b-lora