Update README.md
Browse files
README.md
CHANGED
@@ -6,7 +6,6 @@ datasets:
|
|
6 |
# OLMo-Bitnet-1B
|
7 |
|
8 |
OLMo-Bitnet-1B is a 1B parameter model trained using the method described in [The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits](https://arxiv.org/abs/2402.17764).
|
9 |
-
The result of this is that all of the parameter weights take only the values -1, 0, or 1.
|
10 |
|
11 |
It was trained on a 60B subset of the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset, so it is merely a research proof-of-concept to test out the methodolgy.
|
12 |
|
|
|
6 |
# OLMo-Bitnet-1B
|
7 |
|
8 |
OLMo-Bitnet-1B is a 1B parameter model trained using the method described in [The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits](https://arxiv.org/abs/2402.17764).
|
|
|
9 |
|
10 |
It was trained on a 60B subset of the [Dolma](https://huggingface.co/datasets/allenai/dolma) dataset, so it is merely a research proof-of-concept to test out the methodolgy.
|
11 |
|