Thrilled to introduce Adam-mini, an optimizer that achieves on-par or better performance than AdamW with 45% to 50% less memory footprint. Adam-mini can also achieve 49.5% higher throughput than AdamW on Llama2-7B pre-training.
The design of Adam-mini is inspired by certain Hessian structures we observed on Transformers.
Feel free to try it out! Try switching to Adam-mini with the same hyperparams of AdamW, it would work with only half memory. Hope Adam-mini can help save time, cost, and energy in your tasks!
We are happy to introduce our InstantStyle, which is a framework that employs straightforward yet potent techniques for achieving effective disentanglement of style and content from reference images.
After giving GPU Programming a hands-on try, I have come to appreciate the level of complexity in AI compute:
- Existing/leading frameworks (CUDA, OpenCL, DSLs, even Triton), still fall at the mercy of low-level compute that requires deeper understanding and experience. - Ambiguous optimizations methods that will literally drive you mad π€― - Triton is cool but not cool enough (high level abstractions that fall back to low level compute issues as you build more specialized kernels) - As for CUDA, optimization requires considering all major components of the GPU (DRAM, SRAM, ALUs) π€ - Models today require stallion written GPU kernels to reduce storage and compute cost. - GPTQ was a big save ππΌ
@karpathy is right expertise in this area is scarce and the reason is quite obvious - uncertainties: we are still struggling to get peak performance from multi-connected GPUs while maintaining precision and reducing cost.
π Major Update: OpenLLM Turkish Benchmarks & Leaderboard Launch! π
Exciting news for the Hugging Face community! I'm thrilled to announce the launch of my fully translated OpenLLM Benchmarks in Turkish, accompanied by my innovative leaderboard, ready to highlight the capabilities of Turkish language models. This marks a landmark achievement in supporting and advancing Turkish AI research.
Whatβs New:
π Complete OpenLLM Benchmarks in Turkish: Dive into my comprehensive suite of benchmarks, now available for thorough evaluation of Turkish LLMs.
π Live Leaderboard: Explore my live leaderboard showcasing the progress and excellence in Turkish language AI. (Note: Current evaluations are conducted manually but are consistently updated.)
Partnership Invitation:
π€ Join My Automation Mission: I'm on the lookout for partners to help transition from manual to automated leaderboard evaluations. Your support can catalyze real-time, streamlined assessments, pushing Turkish LLMs to new heights. Key Resources:
π‘ Share Your Models: Contribute to the burgeoning field of Turkish AI, showcasing your work and contributing to the collective progress.
Let's unite to propel Turkish AI forward and set a precedent for the global community. Stay tuned as I plan to expand these efforts to other languages, further enriching the AI ecosystem!
Join this groundbreaking endeavor and letβs shape the future of AI together! π