Yi-34B-exl2 / README.md
turboderp's picture
Update README.md with license information (#1)
b9bdf4d verified
|
raw
history blame
972 Bytes
metadata
language:
  - en
license: apache-2.0

Information

This is a Exl2 quantized version of Yi-34B

Please refer to the original creator for more information.

Calibration dataset: wikitext

Branches:

  • main: Measurement files
  • 4bpw: 4 bits per weight
  • 4.5bpw: 4.5 bits per weight
  • 4.6bpw: 4.6 bits per weight
  • 6bpw: 6 bits per weight

Notes

  • 4.5 and 4.6 bpw are added for experimental purposes to see how far a 24GB card can be pushed.
  • 6bpw is recommended for the best quality to vram usage ratio (assuming you have enough vram).
  • Please ask for more bpws in the community tab if necessary.

Donate?

All my infrastructure and cloud expenses are paid out of pocket. If you'd like to donate, you can do so here: https://ko-fi.com/kingbri

You should not feel obligated to donate, but if you do, I'd appreciate it.