Arth's picture

Arth

SilentWraith

AI & ML interests

None yet

Recent Activity

published a Space 2 days ago
SilentWraith/Image
liked a model 6 months ago
meta-llama/Llama-3.1-405B-Instruct-FP8
liked a Space 7 months ago
somosnlp/nlp-conferences
View all activity

Organizations

fast.ai community's profile picture SomosNLP's profile picture ONNXConfig for all's profile picture Open-Source AI Meetup's profile picture Platzi Community's profile picture Stable Diffusion Dreambooth Concepts Library's profile picture Blog-explorers's profile picture huggingPartyParis's profile picture That Time I got Reincarnated as a Hugging Face Organization's profile picture ZeroGPU Explorers's profile picture Q's profile picture Women on Hugging Face's profile picture MLX Community's profile picture INNOVA AI's profile picture Social Post Explorers's profile picture Dev Mode Explorers's profile picture Paris AI Running Club's profile picture Hugging Face Discord Community's profile picture

SilentWraith's activity

published a Space 2 days ago
liked a Space 7 months ago
reacted to felfri's post with πŸš€ 7 months ago
view post
Post
2283
πŸš€ Excited to announce the release of our new research paper, "LLAVAGUARD: VLM-based Safeguards for Vision Dataset Curation and Safety Assessment"!
In this work, we introduce LLAVAGUARD, a family of cutting-edge Vision-Language Model (VLM) judges designed to enhance the safety and integrity of vision datasets and generative models. Our approach leverages flexible policies for assessing safety in diverse settings. This context awareness ensures robust data curation and model safeguarding alongside comprehensive safety assessments, setting a new standard for vision datasets and models. We provide three versions (7B, 13B, and 34B) and our data, see below. This achievement wouldn't have been possible without the incredible teamwork and dedication of my great colleagues @LukasHug , @PSaiml , @mbrack . πŸ™ Together, we've pushed the boundaries of what’s possible at the intersection of large generative models and safety.
πŸ” Dive into our paper to explore:
Innovative methodologies for dataset curation and model safeguarding.
State-of-the-art safety assessments.
Practical implications for AI development and deployment.
Find more at AIML-TUDA/llavaguard-665b42e89803408ee8ec1086 and https://ml-research.github.io/human-centered-genai/projects/llavaguard/index.html
  • 2 replies
Β·
reacted to albertvillanova's post with πŸš€ 9 months ago
view post
Post
1664
πŸš€ We recently released datasets 2.19.0! πŸ“¦

πŸ”₯ What's New:
- Polars integration πŸ»β€β„οΈ
- fsspec support for conversion to JSON, CSV, and Parquet
- Mode parameter for Image feature
- CLI function to convert script-datasets to Parquet
- Dataset.take and Dataset.skip

Plus, a bunch of general improvements & bug fixes!

Check out the release notes: https://github.com/huggingface/datasets/releases/tag/2.19.0

Upgrade now and power up your data workflows! πŸ’₯
  • 2 replies
Β·
reacted to lewtun's post with πŸ€—β€οΈ 9 months ago
view post
Post
5059
Introducing Zephyr 141B-A35B πŸͺ:

HuggingFaceH4/zephyr-orpo-141b-A35b-v0.1

Yesterday, Mistral released their latest base model (via magnet link of course πŸ˜…) and the community quickly converted it to transformers format and pushed it to the Hub: mistral-community/Mixtral-8x22B-v0.1

Early evals of this model looked extremely strong, so we teamed up with Argilla and KAIST AI to cook up a Zephyr recipe with a few new alignment techniques that came out recently:

πŸ§‘β€πŸ³ Align the base model with Odds Ratio Preference Optimisation (ORPO). This novel algorithm developed by @JW17 and @nlee-208 and @j6mes and does not require an SFT step to achieve high performance and is thus much more computationally efficient than methods like DPO and PPO.

🦫 Use a brand new dataset of 7k high-quality, multi-turn preferences that has been developed by our friends at Argilla. To create this dataset, they took the excellent Capybara SFT dataset from @LDJnr LDJnr/Capybara and converted it into a preference dataset by augmenting the final turn with responses from new LLMs that were then ranked by GPT-4.

What we find especially neat about this approach is that training on 7k samples only takes ~1.3h on 4 H100 nodes, yet produces a model that is very strong on chat benchmarks like IFEval and BBH.

Kudos to @alvarobartt @JW17 and @nlee-208 for this very nice and fast-paced collab!

For more details on the paper and dataset, checkout our collection: HuggingFaceH4/zephyr-orpo-6617eba2c5c0e2cc3c151524
updated a Space 10 months ago