Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: other
|
3 |
+
license_name: nvidia-open-model-license
|
4 |
+
license_link: >-
|
5 |
+
https://developer.download.nvidia.com/licenses/nvidia-open-model-license-agreement-june-2024.pdf
|
6 |
+
inference: false
|
7 |
+
fine-tuning: false
|
8 |
+
tags:
|
9 |
+
- vllm
|
10 |
+
base_model: nvidia/Nemotron-4-340B-Base
|
11 |
+
---
|
12 |
+
|
13 |
+
## Nemotron-4-340B-Base-hf
|
14 |
+
|
15 |
+
Converted checkpoint of [nvidia/Nemotron-4-340B-Base](https://huggingface.co/nvidia/Nemotron-4-340B-Base). Specifically it was produced from the [v1.2 .nemo checkpoint on NGC](https://catalog.ngc.nvidia.com/orgs/nvidia/teams/nemo/models/nemotron-4-340b-base/files?version=1.2).
|
16 |
+
|
17 |
+
This runs in vLLM with this PR: https://github.com/vllm-project/vllm/pull/6611. Support in transformers is still pending.
|
18 |
+
|
19 |
+
### Evaluations
|
20 |
+
|
21 |
+
Please see the [FP8 checkpoint](https://huggingface.co/mgoin/Nemotron-4-340B-Base-hf-FP8) for evaluations since I only have done single-node inference.
|