{
  "dataset": {
    "align_stage_components": [
      "download/llava-laion-cc-sbu-558k/chat.json",
      "download/llava-laion-cc-sbu-558k"
    ],
    "dataset_id": "llava-v15",
    "dataset_root_dir": "data",
    "finetune_stage_components": [
      "download/llava-v1.5-instruct/llava_v1_5_mix665k.json",
      "download/llava-v1.5-instruct"
    ],
    "type": "llava-v15"
  },
  "hf_token": ".hf_token",
  "model": {
    "align_epochs": 1,
    "align_global_batch_size": 256,
    "align_learning_rate": 0.001,
    "align_lr_scheduler_type": "linear-warmup+cosine-decay",
    "align_max_grad_norm": 1.0,
    "align_max_steps": null,
    "align_per_device_batch_size": 16,
    "align_train_strategy": "fsdp-shard-grad-op",
    "align_warmup_ratio": 0.03,
    "align_weight_decay": 0.0,
    "arch_specifier": "no-align+gelu-mlp",
    "enable_gradient_checkpointing": true,
    "enable_mixed_precision_training": true,
    "finetune_epochs": 1,
    "finetune_global_batch_size": 128,
    "finetune_learning_rate": 2e-05,
    "finetune_lr_scheduler_type": "linear-warmup+cosine-decay",
    "finetune_max_grad_norm": 1.0,
    "finetune_max_steps": 7800,
    "finetune_per_device_batch_size": 16,
    "finetune_train_strategy": "fsdp-full-shard",
    "finetune_warmup_ratio": 0.03,
    "finetune_weight_decay": 0.1,
    "image_resize_strategy": "letterbox",
    "llm_backbone_id": "vicuna-v15-7b",
    "llm_max_length": 2048,
    "model_id": "train-1.5-epochs+7b",
    "reduce_in_full_precision": false,
    "type": "train-1.5-epochs+7b",
    "vision_backbone_id": "clip-vit-l-336px"
  },
  "pretrained_checkpoint": null,
  "run_id": "train-1.5-epochs+7b",
  "run_root_dir": "runs",
  "seed": 7,
  "stage": "finetune",
  "trackers": [
    "jsonl",
    "wandb"
  ],
  "wandb_entity": null,
  "wandb_project": "prismatic"
}