Julien Simon commited on
Commit
bf572e3
1 Parent(s): fe190a7
Files changed (2) hide show
  1. results.py +1 -0
  2. results_arcee_supernova.py +19 -0
results.py CHANGED
@@ -27,6 +27,7 @@ instance_type_mappings = {
27
  "c6i.2xlarge": {"cloud": "AWS", "gpu": "-", "gpuRAM": "-"},
28
  "c6i.4xlarge": {"cloud": "AWS", "gpu": "-", "gpuRAM": "-"},
29
  "c7i.4xlarge": {"cloud": "AWS", "gpu": "-", "gpuRAM": "-"},
 
30
  "inf2.48xlarge": {"cloud": "AWS", "gpu": "24xNeuronCore v2", "gpuRAM": "384 GB"},
31
  "inf2.*": {"cloud": "AWS", "gpu": "-", "gpuRAM": "-"},
32
  }
 
27
  "c6i.2xlarge": {"cloud": "AWS", "gpu": "-", "gpuRAM": "-"},
28
  "c6i.4xlarge": {"cloud": "AWS", "gpu": "-", "gpuRAM": "-"},
29
  "c7i.4xlarge": {"cloud": "AWS", "gpu": "-", "gpuRAM": "-"},
30
+ "inf2.24xlarge": {"cloud": "AWS", "gpu": "12xNeuronCore v2", "gpuRAM": "192 GB"},
31
  "inf2.48xlarge": {"cloud": "AWS", "gpu": "24xNeuronCore v2", "gpuRAM": "384 GB"},
32
  "inf2.*": {"cloud": "AWS", "gpu": "-", "gpuRAM": "-"},
33
  }
results_arcee_supernova.py CHANGED
@@ -12,6 +12,25 @@ results_arcee_supernova = {
12
  "tokensPerSecond": "33",
13
  "notes": "",
14
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  {
16
  "instanceType": "inf2.48xlarge",
17
  "quantization": "none",
 
12
  "tokensPerSecond": "33",
13
  "notes": "",
14
  },
15
+ {
16
+ "instanceType": "inf2.24xlarge",
17
+ "configurations": [
18
+ {
19
+ "quantization": "none",
20
+ "container": "transformers-neuronx",
21
+ "status": "KO",
22
+ "tokensPerSecond": "-",
23
+ "notes": "OOM bs=2,seqlen=4096",
24
+ },
25
+ {
26
+ "quantization": "none",
27
+ "container": "transformers-neuronx",
28
+ "status": "KO",
29
+ "tokensPerSecond": "28.4",
30
+ "notes": "OOM bs=2,seqlen=2048",
31
+ },
32
+ ],
33
+ },
34
  {
35
  "instanceType": "inf2.48xlarge",
36
  "quantization": "none",