|
--- |
|
license: bigscience-openrail-m |
|
datasets: |
|
- iamplus/Instruction_Tuning |
|
- iamplus/Conversational_Data |
|
--- |
|
Instruction Tuned GPT-NeoXT-20B model on Instruction Tuning dataset as listed below (~5.2M data) using ***Colossal AI*** |
|
|
|
**Base Model:** togethercomputer/GPT-NeoXT-Chat-Base-20B (GPT-NeoXT-Chat-Base-20B-v0.16 - fine-tuned on feedback data) |
|
|
|
**Training Details :** |
|
* Epochs: 2 |
|
* Batch Size : 5 instantaneous per device x 1 gradient accumulation steps x 8 gpus = 40 |
|
* Block Size : 2020 |
|
* Weight Decay : 0 |
|
* Learning Rate : 1e-6 |
|
* Learning Rate Scheduler Type : Cosine |
|
* Number of warmup steps : 600 |
|
* Machine : 8xA100 80GB |
|
|
|
**Training Data Specifics :** |
|
* Labels and Input ids are exactly the same. |
|
* Block Size is 2020, Multiple instructions are clubbed together in each data. |
|
* "###" is the EOS Token used in the data. |