Arthur Zucker

ArthurZ

AI & ML interests

None yet

Recent Activity

View all activity

Articles

Organizations

Language Technology Research Group at the University of Helsinki's profile picture Hugging Face's profile picture Google's profile picture BigScience Workshop's profile picture Hugging Face Internal Testing Organization's profile picture Beijing Academy of Artificial Intelligence's profile picture HuggingFaceM4's profile picture HFLeoArthurYounes's profile picture Famous's profile picture Hugging Face OSS Metrics's profile picture Polytech Sorbonne X Hugging Face's profile picture Code Llama's profile picture Music Gen Sprint's profile picture huggingPartyParis's profile picture adept-hf-collab's profile picture gg-hf's profile picture Unofficial Mistral Community's profile picture Mistral AI EAP's profile picture State Space Models's profile picture Llava Hugging Face's profile picture Hugging Face Assignments's profile picture mx-test's profile picture On-device Squad's profile picture Social Post Explorers's profile picture hsramall's profile picture Paris AI Running Club's profile picture gg-tt's profile picture Hugging Face Discord Community's profile picture LLHF's profile picture SLLHF's profile picture blhf's profile picture Meta Llama's profile picture nltpt's profile picture Hugging Face Party @ PyTorch Conference's profile picture s0409's profile picture wut?'s profile picture kernels-community's profile picture

ArthurZ's activity

reacted to MonsterMMORPG's post with πŸš€β€οΈ 6 days ago
view post
Post
1804
FLUX Redux is a hidden Gem

I am still doing huge research to publish an amazing fully Public - no paywalled Tutorial, but this is generated via SwarmUI

Style Model Merge Strength : 0.5

FLUX Guidance Scale is : 6

Used base model is my FLUX fine tuned model with 256 images via Kohya SS GUI as shown in tutorial ( https://youtu.be/FvpWy1x5etM ) - 70 epoch

Prompt : anime ohwx man walking in a jungle <segment:yolo-face_yolov9c.pt-1,0.7,0.5> ohwx man, anime
  • 2 replies
Β·
reacted to Xenova's post with πŸ”₯ 14 days ago
view post
Post
5006
Have you tried out πŸ€— Transformers.js v3? Here are the new features:
⚑ WebGPU support (up to 100x faster than WASM)
πŸ”’ New quantization formats (dtypes)
πŸ› 120 supported architectures in total
πŸ“‚ 25 new example projects and templates
πŸ€– Over 1200 pre-converted models
🌐 Node.js (ESM + CJS), Deno, and Bun compatibility
🏑 A new home on GitHub and NPM

Get started with npm i @huggingface/transformers.

Learn more in our blog post: https://huggingface.co/blog/transformersjs-v3
  • 3 replies
Β·
reacted to davidberenstein1957's post with πŸ‘€ 14 days ago
view post
Post
1909
For anyone who struggles with NER or information extraction with LLM.

We showed an efficient workflow for token classification including zero-shot suggestions and model fine-tuning with Argilla, GliNER, the NuMind NuExtract LLM and SpanMarker. @argilla

Video: https://youtu.be/JvLpaYgNd84?feature=shared
Notebooks and slides included to try it yourself πŸ™‚
reacted to LukeNeumann's post with 🀯 14 days ago
view post
Post
1206
Nine years ago, I uploaded the first 8K resolution video to YouTube and I've been stockpiling 8K footage ever since: https://www.youtube.com/watch?v=sLprVF6d7Ug&t

Should @Overlaiapp release the first open-source 8K video dataset?

Could anyone even fine tune a model with this?πŸ˜…
Β·
reacted to their post with ❀️ 14 days ago
reacted to AkimfromParis's post with β€οΈπŸ‘ 14 days ago
view post
Post
1421
πŸ‡―πŸ‡΅ The Open Japanese LLM Leaderboard created by LLM-jp 🌸 in partnership with HuggingFace πŸ€— was released today!

Blog: https://huggingface.co/blog/leaderboard-japanese
Space: llm-jp/open-japanese-llm-leaderboard

🌍 The leaderboard is available in both Japanese and English
πŸ“š Based on the evaluation tool, llm-jp-eval with more than 20 datasets for Japanese LLMs
πŸ“Š The leaderboard showcases all the metrics for NLP experts, plus averages for NLP beginners
πŸ’» For the comfort of users, we chose a horizontal UI, and implemented it in a light and dark theme on Gradio
πŸ”¬ The radar chart provides a very interesting visualization of metrics!
🌱 We are using the Japanese research platform, MDX, so please be patient!
⚑ LLMs bigger than +70B will be evaluated soon…

How do you say β€œGPUs Go Brrr” in Japanese - > GPUγŒγƒ–γƒ³γƒ–γƒ³ο½ž! (To pronounce "GPU ga bunbun!") πŸ”₯
  • 4 replies
Β·
reacted to monsoon-nlp's post with πŸ‘€ 14 days ago
view post
Post
1384
Great to see Tatta Bio release an embeddings version of their DNA/protein language model 🧬: tattabio/gLM2_650M_embed
  • 2 replies
Β·
reacted to AdinaY's post with πŸ‘ 14 days ago
reacted to jsulz's post with πŸš€ 14 days ago
view post
Post
1998
In August, the XetHub team joined Hugging Face
- https://huggingface.co/blog/xethub-joins-hf - and we’ve been rolling up our sleeves to bring the best of both worlds together. We started with a deep dive into the current state of files stored with Git LFS on the Hub.

Getting this information was no small feat. We had to:
* Analyze a complete database dump of all repositories and files stored in Git LFS across Hugging Face.
* Parse through metadata on file sizes and types to accurately map the storage breakdown across Spaces, Models, and Datasets.

You can read more about the findings (with some jaw-dropping stats + charts) here https://www.linkedin.com/feed/update/urn:li:activity:7244486280351285248
reacted to jsulz's post with 🧠 14 days ago
view post
Post
2885
When the XetHub crew joined Hugging Face this fall, @erinys and I started brainstorming how to share our work to replace Git LFS on the Hub. Uploading and downloading large models and datasets takes precious time. That’s where our chunk-based approach comes in.

Instead of versioning files (like Git and Git LFS), we version variable-sized chunks of data. For the Hugging Face community, this means:

⏩ Only upload the chunks that changed.
πŸš€ Download just the updates, not the whole file.
🧠 We store your file as deduplicated chunks

In our benchmarks, we found that using CDC to store iterative model and dataset version led to transfer speedups of ~2x, but this isn’t just a performance boost. It’s a rethinking of how we manage models and datasets on the Hub.

We're planning on our new storage backend to the Hub in early 2025 - check out our blog to dive deeper, and let us know: how could this improve your workflows?

https://huggingface.co/blog/from-files-to-chunks
New activity in mistralai/Pixtral-Large-Instruct-2411 15 days ago

Upload transformers version

6
#3 opened 16 days ago by ArthurZ
posted an update 16 days ago