|
_name_or_path: "TheBloke/Llama-2-7B-fp16", |
|
architectures: [ |
|
"LlamaForCausalLM" |
|
], |
|
bos_token_id: 1, |
|
eos_token_id: 2, |
|
hidden_act: "silu", |
|
hidden_size: 4096, |
|
initializer_range: 0.02, |
|
intermediate_size: 11008, |
|
max_position_embeddings: 2048, |
|
model_type: "llama", |
|
num_attention_heads: 32, |
|
num_hidden_layers: 32, |
|
num_key_value_heads: 32, |
|
pad_token_id: 0, |
|
pretraining_tp: 1, |
|
rms_norm_eps: 0.00001, |
|
rope_scaling: null, |
|
tie_word_embeddings: false, |
|
torch_dtype: "float32", |
|
transformers_version: "4.31.0", |
|
use_cache: true, |
|
vocab_size: 32000 |
|
} |