Lewdiculous
commited on
Commit
•
98e783d
1
Parent(s):
cc67229
Update README.md
Browse files
README.md
CHANGED
@@ -9,6 +9,7 @@ tags:
|
|
9 |
> [!IMPORTANT]
|
10 |
> **Updated!** <br>
|
11 |
> Version (**v2**) files added! With imatrix data generated from the FP16 and conversions directly from the BF16. <br>
|
|
|
12 |
> Hopefully avoiding any losses in the model conversion, as has been the recently discussed topic on Llama-3 and GGUF lately. <br>
|
13 |
> If you are able to test them and notice any issues let me know in the discussions.
|
14 |
|
|
|
9 |
> [!IMPORTANT]
|
10 |
> **Updated!** <br>
|
11 |
> Version (**v2**) files added! With imatrix data generated from the FP16 and conversions directly from the BF16. <br>
|
12 |
+
> This is a more disk and compute intensive so lets hope we get GPU inference support for BF16 models in llama.cpp. <br>
|
13 |
> Hopefully avoiding any losses in the model conversion, as has been the recently discussed topic on Llama-3 and GGUF lately. <br>
|
14 |
> If you are able to test them and notice any issues let me know in the discussions.
|
15 |
|