metadata
license: llama3
Alpesteibock-Llama-3-8B-Alpha
Alpesteibock-Llama-3-8B-Alpha is an experimental QLoRA fine-tune of NousResearch/Hermes-2-Pro-Llama-3-8B on a dataset of more than 28 million tokens of Swiss German text from multiple sources.
Dataset
Training Details
Hardware: 1x RTX 4090 Duration: 30 hours in total (2 hours for first phase and 28 hours for second phase)
Hyperparameters
Adapter: QLoRA Precision: 4 bit Optimizer: adamw_bnb_8bit LoRA Rank: 256 LoRA Alpha: 256 Learning Rate: 1e-5 Context Length: 4096 tokens Batch Size: 1 Gradient Accumulation Steps: 1 Sample Packing: Off for first phase, on for second phase Epochs: 2