Jeonghwan Park PRO

maywell

AI & ML interests

None yet

Recent Activity

Articles

Organizations

maywell's activity

New activity in maywell/koVast 4 days ago

License 문의

1
#1 opened 9 days ago by dyo85
Reacted to yongchanghao's post with πŸ”₯ 21 days ago
view post
Post
3724
We just released a paper (NeuZip) that compresses VRAM in a lossless manner to run larger models. This should be particularly useful when VRAM is insufficient during training/inference. Specifically, we look inside each floating number and find that the exponents are highly compressible (as shown in the figure below).

Read more about the work at NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks (2410.20650)