Llama-2-7b-hf / flyte_training_config.json
nielsbantilan's picture
Upload folder using huggingface_hub
1aa01d1
raw
history blame
562 Bytes
{"base_model": "meta-llama/Llama-2-7b-hf", "data_path": "wikipedia", "data_name": "20220301.simple", "num_epochs": 1, "max_steps": 100, "learning_rate": 2e-05, "weight_decay": 0.02, "warmup_ratio": 0.03, "lr_scheduler_type": "cosine", "batch_size": 4, "micro_batch_size": 1, "val_set_size": 0, "group_by_length": false, "instruction_key": "instruction", "input_key": "input", "output_key": "output", "device_map": "auto", "cache_dir": null, "optim": "adamw_torch", "model_max_length": 512, "debug_mode": false, "debug_train_data_size": 1024, "wandb_project": ""}