Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
45.5
TFLOPS
162
13
62
Jean Louis
JLouisBiz
Follow
1234aurel's profile picture
d2evid's profile picture
Meetpatel018's profile picture
78 followers
Β·
115 following
https://www.StartYourOwnGoldMine.com
YourOwnGoldMine
gnusupport
AI & ML interests
- LLM for sales, marketing, promotion - LLM for Website Revision System - increasing quality of communication with customers - helping clients access information faster - saving people from financial troubles
Recent Activity
replied
to
Jiaqi-hkust
's
post
about 20 hours ago
We have open-sourced Robust-R1 (AAAI 2026 Oral), a new paradigm in the field of anti-degradation and robustness enhancement for multimodal large models. Multimodal Large Language Models struggle to maintain reliable performance under extreme real-world visual degradations, which impede their practical robustness. Existing robust MLLMs predominantly rely on implicit training/adaptation that focuses solely on visual encoder generalization, suffering from limited interpretability and isolated optimization. To overcome these limitations, we propose Robust-R1, a novel framework that explicitly models visual degradations through structured reasoning chains. Our approach integrates: (i) supervised fine-tuning for degradation-aware reasoning foundations, (ii) reward-driven alignment for accurately perceiving degradation parameters, and (iii) dynamic reasoning depth scaling adapted to degradation intensity. To facilitate this approach, we introduce a specialized 11K dataset featuring realistic degradations synthesized across four critical real-world visual processing stages, each annotated with structured chains connecting degradation parameters, perceptual influence, pristine semantic reasoning chain, and conclusion. Comprehensive evaluations demonstrate state-of-the-art robustness: Robust-R1 outperforms all general and robust baselines on the real-world degradation benchmark R-Bench, while maintaining superior anti-degradation performance under multi-intensity adversarial degradations on MMMB, MMStar, and RealWorldQA. We have made all of our papers, codes, data, model weights and demos fully open-source: Paper: https://huggingface.co/papers/2512.17532 (help us to upvote) GitHub code: https://github.com/jqtangust/Robust-R1 (help us to star) HF model: https://huggingface.co/Jiaqi-hkust/Robust-R1 HF data: https://huggingface.co/datasets/Jiaqi-hkust/Robust-R1 HF Space: https://huggingface.co/spaces/Jiaqi-hkust/Robust-R1 We sincerely invite everyone to give it a try.
reacted
to
inoculatemedia
's
post
with π
about 20 hours ago
Iβm opening the waitlist for what I believe to be the most advanced multimodal bridge for A/V professionals. Txt2img, img2video, editing, export to ProRes, apply Luts, Pexels and TouchDesigner integrations, music and voice gen, multichannel mixing. Announcing: Lilikoi by Haawke AI Teaser video made entirely with Lilikoi: https://youtu.be/-O7DH7vFkYg?si=q2t5t6WjQCk2Cp0w Https://Lilikoi.haawke.com Technical brief: https://haawke.com/technical_brief.html
new
activity
2 days ago
Kibalama/lugandaSTT:
How do I convert the file for whispper.cpp compatibility?
View all activity
Organizations
JLouisBiz
's Spaces
1
Sort:Β Recently updated
Running
1
GNU LLM Integration
π
Empowering GNU/Linux users with NLP