metadata
license: apache-2.0
datasets:
- tatsu-lab/alpaca
language:
- en
model_name_or_path = yahma/llama-13b-hf
learning_rate = 1e-4
max_steps = 10000
epoch = 3.08
train_loss = 0.7169
train_runtime = 1 day, 7:23:32.67
train_samples_per_second = 1.416
train_steps_per_second = 0.088