tiny_model / mlp_map_test /M0_S-10_R16_P5_config.json
noanabeshima's picture
Upload folder using huggingface_hub
84e485d verified
raw
history blame contribute delete
925 Bytes
{
"n_features": 25000,
"d_model": 768,
"lr_exp": -10,
"disable_comet": false,
"per_neuron_reinit_interval": 0,
"reservoir_time_discount": 0.995,
"reinit_interval": 800,
"max_reinit_neurons": 5000,
"reservoir_size": 5000,
"n_piles": 292,
"log_interval": 200,
"reinit_input_norm": "target_scaled",
"reinit_input": "x",
"reinit_norm_alpha": 0.3,
"data_loc": "mlp_data",
"reinit_threshold": -6,
"scheduler": "wsd",
"layer_idx": 0,
"l1_exp": -10,
"neuron_reinit_percent": 0.85,
"beta1": 1,
"beta2": 4,
"reinit_target": "y",
"sparse_adam": false,
"run_template": "M{layer_idx}_S{l1_exp}_R{l1_ratio}_P{l1_p}",
"project_name": "mlp_map_test",
"decoder_bias": true,
"l1_beta": 0.99,
"alt_sparsity_loss": "log",
"l1_ratio": 16,
"l1_p": 5,
"optimizer": "sparse_adam",
"model_type": "mlp_map",
"adam_beta1": 0.5,
"adam_beta2": 0.9375,
"run_name": "M0_S-10_R16_P5"
}