jartine commited on
Commit
072f53a
1 Parent(s): ce86e9e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -49,7 +49,7 @@ It uses Cosmopolitan Libc to turn LLM weights into runnable llama.cpp
49
  binaries that run on the stock installs of six OSes for both ARM64 and
50
  AMD64.
51
 
52
- ## About Quantization Formats
53
 
54
  Your choice of quantization format depends on three things:
55
 
@@ -67,7 +67,7 @@ computation speed (flops) so simpler quants help.
67
 
68
  Note: BF16 is currently only supported on CPU.
69
 
70
- ## Hardware Choices
71
 
72
  Any Macbook with 32GB should be able to run
73
  Meta-Llama-3-70B-Instruct.Q2\_K.llamafile reasonably well. At this
 
49
  binaries that run on the stock installs of six OSes for both ARM64 and
50
  AMD64.
51
 
52
+ ## About Quantization Formats (General Advice)
53
 
54
  Your choice of quantization format depends on three things:
55
 
 
67
 
68
  Note: BF16 is currently only supported on CPU.
69
 
70
+ ## Hardware Choices (LLaMA3 70B Specific)
71
 
72
  Any Macbook with 32GB should be able to run
73
  Meta-Llama-3-70B-Instruct.Q2\_K.llamafile reasonably well. At this