dacorvo HF staff commited on
Commit
04c5b2a
1 Parent(s): 8c690bf

Synchronizing local compiler cache.

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/502c174d959be18e9532.json +1 -0
  2. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/de8fa96c7bfc636c1862.json +1 -0
  3. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/f8993574c5ad6dd46126.json +1 -0
  4. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/gpt2/gpt2/2f32aee440dc0531b512.json +1 -0
  5. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/27b58df3bfc4df73d037.json +1 -0
  6. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/89abea0c2d8d0238662e.json +1 -0
  7. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/acae9d99b368f7e213f8.json +1 -0
  8. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/7d494ae543777c6ceced.json +1 -0
  9. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/ae94dd17cae6a47e6149.json +1 -0
  10. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/e1b2ea90dc6ea1de8bfd.json +1 -0
  11. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/030abc51a5f2f76b79e7.json +1 -0
  12. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/a8de305007c583a8655b.json +1 -0
  13. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/b5387a4261fa49a90350.json +1 -0
  14. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/0d04f4d4b987d19e2758.json +1 -0
  15. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/81569bedad5c44215faa.json +1 -0
  16. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/8739d715a7c9acaf4c0b.json +1 -0
  17. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/6823d660d178eb94eca0.json +1 -0
  18. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/bb6453044adc53f42f22.json +1 -0
  19. neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/e48297c242b96882b93b.json +1 -0
  20. neuronxcc-2.14.213.0+013d129b/MODULE_0b9060298e5901893068+2c2d707e/compile_flags.txt +1 -0
  21. neuronxcc-2.14.213.0+013d129b/MODULE_0b9060298e5901893068+2c2d707e/model.hlo_module.pb +3 -0
  22. neuronxcc-2.14.213.0+013d129b/MODULE_0b9060298e5901893068+2c2d707e/model.neff +0 -0
  23. neuronxcc-2.14.213.0+013d129b/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/compile_flags.txt +1 -0
  24. neuronxcc-2.14.213.0+013d129b/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/model.hlo_module.pb +3 -0
  25. neuronxcc-2.14.213.0+013d129b/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/model.neff +0 -0
  26. neuronxcc-2.14.213.0+013d129b/MODULE_1f3d02b2530b83c4abf7+2c2d707e/compile_flags.txt +1 -0
  27. neuronxcc-2.14.213.0+013d129b/MODULE_1f3d02b2530b83c4abf7+2c2d707e/model.hlo_module.pb +3 -0
  28. neuronxcc-2.14.213.0+013d129b/MODULE_1f3d02b2530b83c4abf7+2c2d707e/model.neff +0 -0
  29. neuronxcc-2.14.213.0+013d129b/MODULE_23ced3c113e97afdb062+2c2d707e/compile_flags.txt +1 -0
  30. neuronxcc-2.14.213.0+013d129b/MODULE_23ced3c113e97afdb062+2c2d707e/model.hlo_module.pb +3 -0
  31. neuronxcc-2.14.213.0+013d129b/MODULE_23ced3c113e97afdb062+2c2d707e/model.neff +0 -0
  32. neuronxcc-2.14.213.0+013d129b/MODULE_257d339d60d0fce6e129+2c2d707e/compile_flags.txt +1 -0
  33. neuronxcc-2.14.213.0+013d129b/MODULE_257d339d60d0fce6e129+2c2d707e/model.hlo_module.pb +3 -0
  34. neuronxcc-2.14.213.0+013d129b/MODULE_257d339d60d0fce6e129+2c2d707e/model.neff +0 -0
  35. neuronxcc-2.14.213.0+013d129b/MODULE_2aa937204f63179c0db3+2c2d707e/compile_flags.txt +1 -0
  36. neuronxcc-2.14.213.0+013d129b/MODULE_2aa937204f63179c0db3+2c2d707e/model.hlo_module.pb +3 -0
  37. neuronxcc-2.14.213.0+013d129b/MODULE_2aa937204f63179c0db3+2c2d707e/model.neff +0 -0
  38. neuronxcc-2.14.213.0+013d129b/MODULE_34480d97464f3bf86d90+2c2d707e/compile_flags.txt +1 -0
  39. neuronxcc-2.14.213.0+013d129b/MODULE_34480d97464f3bf86d90+2c2d707e/model.hlo_module.pb +3 -0
  40. neuronxcc-2.14.213.0+013d129b/MODULE_34480d97464f3bf86d90+2c2d707e/model.neff +0 -0
  41. neuronxcc-2.14.213.0+013d129b/MODULE_46948ddb997e0e72bed0+2c2d707e/compile_flags.txt +1 -0
  42. neuronxcc-2.14.213.0+013d129b/MODULE_46948ddb997e0e72bed0+2c2d707e/model.hlo_module.pb +3 -0
  43. neuronxcc-2.14.213.0+013d129b/MODULE_46948ddb997e0e72bed0+2c2d707e/model.neff +0 -0
  44. neuronxcc-2.14.213.0+013d129b/MODULE_4a486888f48de6666282+2c2d707e/compile_flags.txt +1 -0
  45. neuronxcc-2.14.213.0+013d129b/MODULE_4a486888f48de6666282+2c2d707e/model.hlo_module.pb +3 -0
  46. neuronxcc-2.14.213.0+013d129b/MODULE_4a486888f48de6666282+2c2d707e/model.neff +0 -0
  47. neuronxcc-2.14.213.0+013d129b/MODULE_4fd66239b44449a4f2f7+2c2d707e/compile_flags.txt +1 -0
  48. neuronxcc-2.14.213.0+013d129b/MODULE_4fd66239b44449a4f2f7+2c2d707e/model.hlo_module.pb +3 -0
  49. neuronxcc-2.14.213.0+013d129b/MODULE_4fd66239b44449a4f2f7+2c2d707e/model.neff +0 -0
  50. neuronxcc-2.14.213.0+013d129b/MODULE_667f1d540de4ddfad7c8+2c2d707e/compile_flags.txt +1 -0
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/502c174d959be18e9532.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/de8fa96c7bfc636c1862.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/bloom/hf-internal-testing/tiny-random-BloomForCausalLM/f8993574c5ad6dd46126.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"apply_residual_connection_post_layernorm": false, "architectures": ["BloomForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 1, "dtype": "float32", "eos_token_id": 2, "gradient_checkpointing": false, "hidden_dropout": 0.1, "hidden_size": 32, "id2label": {"0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2"}, "initializer_range": 0.02, "is_decoder": true, "label2id": {"LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2}, "layer_norm_epsilon": 1e-05, "model_type": "bloom", "n_head": 4, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-BloomForCausalLM", "checkpoint_revision": "92b07e9b7b4f986fa7c54e2ac3b9201b4ba5212e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 3, "pretraining_tp": 1, "seq_length": 7, "slow_but_exact": true, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 1024}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/gpt2/gpt2/2f32aee440dc0531b512.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "architectures": ["GPT2LMHeadModel"], "attn_pdrop": 0.1, "bos_token_id": 50256, "embd_pdrop": 0.1, "eos_token_id": 50256, "initializer_range": 0.02, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 1024, "n_embd": 768, "n_head": 12, "n_inner": null, "n_layer": 12, "n_positions": 1024, "neuron": {"auto_cast_type": "fp16", "batch_size": 4, "checkpoint_id": "gpt2", "checkpoint_revision": "607a30d783dfa663caf39e06633721c8d4cfcd7e", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 1024, "task": "text-generation"}, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "task_specific_params": {"text-generation": {"do_sample": true, "max_length": 50}}, "use_cache": true, "vocab_size": 50257}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/27b58df3bfc4df73d037.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/89abea0c2d8d0238662e.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/gpt2/hf-internal-testing/tiny-random-gpt2/acae9d99b368f7e213f8.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation_function": "gelu_new", "attention_probs_dropout_prob": 0.1, "attn_pdrop": 0.1, "bos_token_id": 98, "embd_pdrop": 0.1, "eos_token_id": 98, "gradient_checkpointing": false, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "initializer_range": 0.02, "intermediate_size": 37, "layer_norm_epsilon": 1e-05, "model_type": "gpt2", "n_ctx": 512, "n_embd": 32, "n_head": 4, "n_inner": null, "n_layer": 5, "n_positions": 512, "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-gpt2", "checkpoint_revision": "71034c5d8bde858ff824298bdedc65515b97d2b9", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "pad_token_id": 98, "reorder_and_upcast_attn": false, "resid_pdrop": 0.1, "scale_attn_by_inverse_layer_idx": false, "scale_attn_weights": true, "summary_activation": null, "summary_first_dropout": 0.1, "summary_proj_to_labels": true, "summary_type": "cls_index", "summary_use_proj": true, "type_vocab_size": 16, "use_cache": true, "vocab_size": 1000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/7d494ae543777c6ceced.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/ae94dd17cae6a47e6149.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/llama/dacorvo/tiny-random-llama/e1b2ea90dc6ea1de8bfd.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "attention_bias": false, "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 128, "initializer_range": 0.02, "intermediate_size": 256, "max_position_embeddings": 512, "mlp_bias": false, "model_type": "llama", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-llama", "checkpoint_revision": "7fdafd2fe6a2d31c6abb72ae60db606d8bb23196", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 1, "num_hidden_layers": 1, "num_key_value_heads": 1, "pretraining_tp": 1, "rms_norm_eps": 1e-06, "rope_scaling": null, "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/030abc51a5f2f76b79e7.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/a8de305007c583a8655b.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mistral/dacorvo/tiny-random-MistralForCausalLM/b5387a4261fa49a90350.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MistralForCausalLM"], "attention_dropout": 0.0, "attention_probs_dropout_prob": 0.1, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "gelu", "hidden_dropout_prob": 0.1, "hidden_size": 32, "initializer_range": 0.02, "intermediate_size": 37, "is_decoder": true, "max_position_embeddings": 512, "model_type": "mistral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/tiny-random-MistralForCausalLM", "checkpoint_revision": "81d453e3c8985649e9ee3d4c9378461029d1c73a", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 2, "num_key_value_heads": 2, "pad_token_id": 0, "rms_norm_eps": 1e-06, "rope_theta": 10000.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float32", "type_vocab_size": 16, "use_cache": true, "vocab_size": 32000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/0d04f4d4b987d19e2758.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/81569bedad5c44215faa.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/mixtral/dacorvo/Mixtral-tiny/8739d715a7c9acaf4c0b.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["MixtralForCausalLM"], "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 1024, "initializer_range": 0.02, "intermediate_size": 3584, "max_position_embeddings": 1024, "model_type": "mixtral", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "dacorvo/Mixtral-tiny", "checkpoint_revision": "c557ba205ddff6ea911f4719e0d543d6c08356b6", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 32, "num_experts_per_tok": 2, "num_hidden_layers": 2, "num_key_value_heads": 8, "num_local_experts": 8, "output_router_logits": false, "rms_norm_eps": 1e-05, "rope_theta": 10000.0, "router_aux_loss_coef": 0.001, "router_jitter_noise": 0.0, "sliding_window": 4096, "tie_word_embeddings": false, "torch_dtype": "float16", "use_cache": true, "vocab_size": 32000}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/6823d660d178eb94eca0.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 2, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/bb6453044adc53f42f22.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp32", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
neuronxcc-2.14.213.0+013d129b/0_REGISTRY/0.0.24.dev0/inference/opt/hf-internal-testing/tiny-random-OPTForCausalLM/e48297c242b96882b93b.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_remove_final_layer_norm": false, "activation_function": "relu", "architectures": ["OPTForCausalLM"], "attention_dropout": 0.1, "bos_token_id": 2, "do_layer_norm_before": true, "dropout": 0.1, "embed_dim": 16, "enable_bias": true, "eos_token_id": 2, "ffn_dim": 4, "hidden_size": 16, "init_std": 0.02, "is_decoder": true, "layer_norm_elementwise_affine": true, "layerdrop": 0.0, "max_position_embeddings": 100, "model_type": "opt", "neuron": {"auto_cast_type": "fp16", "batch_size": 1, "checkpoint_id": "hf-internal-testing/tiny-random-OPTForCausalLM", "checkpoint_revision": "190d1f4fc0011d2eaeaa05282e0fbd2445e4b11f", "compiler_type": "neuronx-cc", "compiler_version": "2.14.213.0+013d129b", "num_cores": 2, "sequence_length": 100, "task": "text-generation"}, "num_attention_heads": 4, "num_hidden_layers": 5, "pad_token_id": 1, "torch_dtype": "float32", "use_cache": true, "vocab_size": 50265, "word_embed_proj_dim": 16}
neuronxcc-2.14.213.0+013d129b/MODULE_0b9060298e5901893068+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_0b9060298e5901893068+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7d78190172a5596e4c18b735e6f3b730c56c05c9c05685feb3bb3da6de729c7
3
+ size 64090
neuronxcc-2.14.213.0+013d129b/MODULE_0b9060298e5901893068+2c2d707e/model.neff ADDED
Binary file (400 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0beb3bc81714d1f70e77131f347a87578e245d93457b5aa875e18761c434e459
3
+ size 1900
neuronxcc-2.14.213.0+013d129b/MODULE_1ddfe7bb03e6b5141bca+2c2d707e/model.neff ADDED
Binary file (42 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_1f3d02b2530b83c4abf7+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_1f3d02b2530b83c4abf7+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c016cb9348980e5ceadfeb94168eab02c144bb5693c61751c2d3367d09ef5b5f
3
+ size 57433
neuronxcc-2.14.213.0+013d129b/MODULE_1f3d02b2530b83c4abf7+2c2d707e/model.neff ADDED
Binary file (155 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_23ced3c113e97afdb062+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_23ced3c113e97afdb062+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76ee8fcd669c0344010019e40d232b2beb49ae91140f62839d3464f9d6c4bd3a
3
+ size 60871
neuronxcc-2.14.213.0+013d129b/MODULE_23ced3c113e97afdb062+2c2d707e/model.neff ADDED
Binary file (165 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_257d339d60d0fce6e129+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_257d339d60d0fce6e129+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15567e4d8acb5d4d60c377ea8defce5048d45c48f339a6185772a293a8afa793
3
+ size 49801
neuronxcc-2.14.213.0+013d129b/MODULE_257d339d60d0fce6e129+2c2d707e/model.neff ADDED
Binary file (175 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_2aa937204f63179c0db3+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_2aa937204f63179c0db3+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d86873653e3f68563af1f244d97fd91d90568d1dfc5ede3cdceb914bb08460a7
3
+ size 57064
neuronxcc-2.14.213.0+013d129b/MODULE_2aa937204f63179c0db3+2c2d707e/model.neff ADDED
Binary file (155 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_34480d97464f3bf86d90+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_34480d97464f3bf86d90+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:634c3398b71dcfaabba007cabe8d9a72a6020285edef8ab2bd1cde5422a1bf5f
3
+ size 46625
neuronxcc-2.14.213.0+013d129b/MODULE_34480d97464f3bf86d90+2c2d707e/model.neff ADDED
Binary file (493 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_46948ddb997e0e72bed0+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_46948ddb997e0e72bed0+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6b93438bd3b5a59ce97c1ae510ec134652a81d958b6ca06f2122467a6569b1f
3
+ size 1900
neuronxcc-2.14.213.0+013d129b/MODULE_46948ddb997e0e72bed0+2c2d707e/model.neff ADDED
Binary file (52.2 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_4a486888f48de6666282+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_4a486888f48de6666282+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:401229f357b8f66524080f98dd7eb6a847f069c7ac6908aa712968abc981b181
3
+ size 11805
neuronxcc-2.14.213.0+013d129b/MODULE_4a486888f48de6666282+2c2d707e/model.neff ADDED
Binary file (134 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_4fd66239b44449a4f2f7+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none
neuronxcc-2.14.213.0+013d129b/MODULE_4fd66239b44449a4f2f7+2c2d707e/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:befe42fedd8b9425909c5facb895e1105a46e1910c668329289cc3d4a17ee184
3
+ size 11430
neuronxcc-2.14.213.0+013d129b/MODULE_4fd66239b44449a4f2f7+2c2d707e/model.neff ADDED
Binary file (134 kB). View file
 
neuronxcc-2.14.213.0+013d129b/MODULE_667f1d540de4ddfad7c8+2c2d707e/compile_flags.txt ADDED
@@ -0,0 +1 @@
 
 
1
+ --model-type=transformer --auto-cast=none