if you use Google Kubernetes Engine to host you ML workloads, I think this series of videos is a great way to kickstart your journey of deploying LLMs, in less than 10 minutes! Thank you @wietse-venema-demo !
I'd like to share here a bit more about our Deep Learning Containers (DLCs) we built with Google Cloud, to transform the way you build AI with open models on this platform!
With pre-configured, optimized environments for PyTorch Training (GPU) and Inference (CPU/GPU), Text Generation Inference (GPU), and Text Embeddings Inference (CPU/GPU), the Hugging Face DLCs offer:
⚡ Optimized performance on Google Cloud's infrastructure, with TGI, TEI, and PyTorch acceleration. 🛠️ Hassle-free environment setup, no more dependency issues. 🔄 Seamless updates to the latest stable versions. 💼 Streamlined workflow, reducing dev and maintenance overheads. 🔒 Robust security features of Google Cloud. ☁️ Fine-tuned for optimal performance, integrated with GKE and Vertex AI. 📦 Community examples for easy experimentation and implementation. 🔜 TPU support for PyTorch Training/Inference and Text Generation Inference is coming soon!
IBM & NASA just released open-source AI model for weather & climate on Hugging Face.
Prithvi WxC offers insights beyond forecasting, tackling challenges from local weather to global climate. Potential apps: targeted forecasts, severe weather detection & more. https://huggingface.co/Prithvi-WxC
This is impressive. Check out this comparison of the Ida hurricane between ground truth and the AI model's prediction.
Thanks to the Hugging Face DLCs for TGI and Google Cloud Vertex AI, deploying a high-performance text generation container for serving Large Language Models (LLMs) has never been easier. And we’re not going to stop here – stay tuned as we enable more experiences to build AI with open models on Google Cloud!
Just crossed 200,000 free public AI datasets shared by the community on Hugging Face! Text, image, video, audio, time-series & many more... Thanks everyone!