Adina Yakefu

AdinaY

AI & ML interests

None yet

Recent Activity

View all activity

Organizations

Hugging Face's profile picture Hugging Face Chinese Localization's profile picture Huggingface Projects's profile picture Blog-explorers's profile picture ICCV2023's profile picture Open LLM Leaderboard's profile picture huggingPartyParis's profile picture Qwen's profile picture Women on Hugging Face's profile picture Journalists on Hugging Face's profile picture Social Post Explorers's profile picture Chinese LLMs on Hugging Face's profile picture Hugging Face for Legal's profile picture

AdinaY's activity

reacted to lin-tan's post with ๐Ÿ”ฅ about 22 hours ago
view post
Post
1340
๐Ÿš€ Excited to share that our paper, "SELP: Generating Safe and Efficient Task Plans for Robot Agents with Large Language Models", has been accepted to #ICRA2025! ๐Ÿ”— Preprint: https://arxiv.org/pdf/2409.19471

We introduce SELP (Safe Efficient LLM Planner), a novel approach for generating plans that adhere to user-specified constraints while optimizing for time-efficient execution. By leveraging linear temporal logic (LTL) to interpret natural language commands, SELP effectively handles complex commands and long-horizon tasks. ๐Ÿค–

๐Ÿ’กSELP presents three key insights:
1๏ธโƒฃ Equivalence Voting: Ensures robust translations from natural language instructions into LTL specifications.
2๏ธโƒฃ Constrained Decoding: Uses the generated LTL formula to guide the autoregressive inference of plans, ensuring the generated plans conform to the LTL.
3๏ธโƒฃ Domain-Specific Fine-Tuning: Customizes LLMs for specific robotic tasks, boosting both safety and efficiency.

๐Ÿ“Š Experiment: Our experiments demonstrate SELPโ€™s effectiveness and generalizability across diverse tasks. In drone navigation, SELP outperforms state-of-the-art LLM planners by 10.8% in safety rate and by 19.8% in plan efficiency. For robot manipulation, SELP achieves a 20.4% improvement in safety rate.

@yiwu @jiang719

#ICRA2025 #LLM #Robotics #Agent #LLMPlanner
reacted to their post with ๐Ÿ”ฅ about 22 hours ago
view post
Post
1139
Xwen ๐Ÿ”ฅ a series of open models based on Qwen2.5 models, developed by a brilliant research team of PhD students from the Chinese community.
shenzhi-wang/xwen-chat-679e30ab1f4b90cfa7dbc49e
โœจ 7B/72B
โœจ Apache 2.0
โœจ Xwen-72B-Chat outperformed DeepSeek V3 on Arena Hard Auto
posted an update about 22 hours ago
view post
Post
1139
Xwen ๐Ÿ”ฅ a series of open models based on Qwen2.5 models, developed by a brilliant research team of PhD students from the Chinese community.
shenzhi-wang/xwen-chat-679e30ab1f4b90cfa7dbc49e
โœจ 7B/72B
โœจ Apache 2.0
โœจ Xwen-72B-Chat outperformed DeepSeek V3 on Arena Hard Auto
reacted to m-ric's post with ๐Ÿš€๐Ÿ”ฅ 1 day ago
view post
Post
5530
Introducing ๐—ผ๐—ฝ๐—ฒ๐—ป ๐——๐—ฒ๐—ฒ๐—ฝ-๐—ฅ๐—ฒ๐˜€๐—ฒ๐—ฎ๐—ฟ๐—ฐ๐—ต by Hugging Face! ๐Ÿ’ฅ

OpenAI's latest agentic app Deep Research seems really good... But it's closed, as usual.

โฑ๏ธ So with a team of cracked colleagues, we set ourselves a 24hours deadline to replicate and open-source Deep Research! โฑ๏ธ

โžก๏ธ We built open-Deep-Research, an entirely open agent that can: navigate the web autonomously, scroll and search through pages, download and manipulate files, run calculation on data...

We aimed for the best performance: are the agent's answers really rigorous?

On GAIA benchmark, Deep Research had 67% accuracy on the validation set.
โžก๏ธ open Deep Research is at 55% (powered by o1), it is:
- the best pass@1 solution submitted
- the best open solution ๐Ÿ’ช๐Ÿ’ช

And it's only getting started ! Please jump in, drop PRs, and let's bring it to the top !

Read the blog post ๐Ÿ‘‰ https://huggingface.co/blog/open-deep-research
reacted to albertvillanova's post with ๐Ÿค— 1 day ago
view post
Post
1727
๐Ÿš€ Introducing @huggingface Open Deep-Research๐Ÿ’ฅ

In just 24 hours, we built an open-source agent that:
โœ… Autonomously browse the web
โœ… Search, scroll & extract info
โœ… Download & manipulate files
โœ… Run calculations on data

55% on GAIA validation set! Help us improve it!๐Ÿ’ก
https://huggingface.co/blog/open-deep-research
  • 1 reply
ยท
reacted to victor's post with ๐Ÿ”ฅโค๏ธ 1 day ago
view post
Post
2775
Hey everyone, we've given https://hf.co/spaces page a fresh update!

Smart Search: Now just type what you want to doโ€”like "make a viral meme" or "generate music"โ€”and our search gets it.

New Categories: Check out the cool new filter bar with icons to help you pick a category fast.

Redesigned Space Cards: Reworked a bit to really show off the app descriptions, so you know what each Space does at a glance.

Random Prompt: Need ideas? Hit the dice button for a burst of inspiration.

Weโ€™d love to hear what you thinkโ€”drop us some feedback plz!
ยท
reacted to ZhengPeng7's post with ๐Ÿ”ฅ๐Ÿ‘ 1 day ago
view post
Post
1353
We just released the [BiRefNet_HR]( ZhengPeng7/BiRefNet_HR) for general use on higher resolution images, which was trained with images in 2048x2048. If your images are mostly larger than 1024x1024, use BiRefNet_HR for better results! Thanks to @Freepik for the kind support of H200s for this huge training.

HF Model: ZhengPeng7/BiRefNet_HR.
HF Demo: ZhengPeng7/BiRefNet_demo, where you need to choose General-HR and set high resolution.
PyTorch weights & ONNX: in Google Drive and the GitHub release.

Here is a comparison between the results of the original one and the new HR one on HR inputs:

And, the performance of this new HR one and the previous one trained in 1024x1024 on val set:
posted an update 9 days ago
replied to their post 10 days ago
reacted to clem's post with ๐Ÿค—๐Ÿ”ฅ 10 days ago
view post
Post
6906
AI is not a zero-sum game. Open-source AI is the tide that lifts all boats!
replied to their post 10 days ago
view reply

Good catch! Qwen team also mentioned there'll be some surprise. ๐Ÿ‘€

posted an update 10 days ago
view post
Post
2602
๐Ÿ”ฅSo many exciting releases coming from the Chinese community this month!
zh-ai-community/2025-january-6786b054f492fb223591269e

LLMs:
โœจ Qwen2.5 -1M by Alibaba
Qwen/qwen25-1m-679325716327ec07860530ba
โœจ InternLM3-8B-Instruct by Shanghai AI Lab
internlm/internlm3-8b-instruct
โœจ MiniMax-Text-01 by MiniMax AI
MiniMaxAI/MiniMax-Text-01
โœจ RWKV-7 by BlinkDL -- RNN + Transformer ๐Ÿ‘€
BlinkDL/rwkv-7-world
โœจ DeepSeek-R1 by DeepSeek -- THE ONE ๐Ÿ™Œ
https://huggingface.co/deepseek-ai
โœจ Baichuan-M1-14B by Baichuan - Medical ๐Ÿฉบ
baichuan-inc/Baichuan-M1-14B-Base
โœจ Qwen2.5-Math-PRM by Alibaba - Math ๐Ÿ”ข
Qwen/Qwen2.5-Math-PRM-7B

Code:
โœจ Tare by Bytedance
https://trae.ai

TTS:
โœจ T2A-01-HD by MiniMax AI
https://hailuo.ai/audio
โœจ LLaSA by HKUST Audio
HKUSTAudio/Llasa-3B

MLLM:
โœจ Kimi k1.5 by Moonshot AI
https://kimi.ai
โœจ MiniCPM-o-2_6 by OpenBMB
openbmb/MiniCPM-o-2_6
โœจ Sa2VA-4B by ByteDance
ByteDance/Sa2VA-4B
โœจ VideoLLaMA 3 by Alibaba DAMO
DAMO-NLP-SG/videollama3-678cdda9281a0e32fe79af15
โœจ LLaVA-Mini by Chinese Academy of Sciences
ICTNLP/llava-mini-llama-3.1-8b
โœจHunyuan-7B by Tencent
tencent/Hunyuan-7B-Instruct
โœจ Hunyuan 3D 2.0 by Tencent
tencent/Hunyuan3D-2
โœจMiniMax-VL-01 by MiniMax AI - A non transformer based VLM ๐Ÿ‘€
MiniMaxAI/MiniMax-VL-01

Agent:
โœจ UI-TARS by Bytedance
bytedance-research/UI-TARS-7B-SFT
โœจ GLM-PC by Zhipu AI
https://cogagent.aminer.cn

Dataset:
โœจ Fineweb-Edu-Chinese by Opencsg
opencsg/Fineweb-Edu-Chinese-V2.1
โœจ Multimodal_textbook by Alibaba
DAMO-NLP-SG/multimodal_textbook
โœจ MME-Finance by Hithink AI
ยท
reacted to Kseniase's post with ๐Ÿ”ฅ๐Ÿš€ 11 days ago
view post
Post
3005
7 Open-source Methods to Improve Video Generation and Understanding

AI community is making great strides toward achieving the full potential of multimodality in video generation and understanding. Last week studies showed that working with videos is now one of the main focuses for improving AI models. Another highlight of the week is that open source, once again, proves its value. For those who were impressed by DeepSeek-R1, weโ€™re with you!

Today, weโ€™re combining these two key focuses and bringing you a list of open-source methods for better video generation and understanding:

1. VideoLLaMA 3 model: Excels in various video and image tasks thanks to vision-centric training approach. VideoLLaMA 3: Frontier Multimodal Foundation Models for Image and Video Understanding (2501.13106)

2. FILMAGENT framework assigns roles to multiple AI agents, like a director, screenwriter, actor, and cinematographer, to automate the filmmaking process in 3D virtual environments. FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in Virtual 3D Spaces (2501.12909)

3. Improving Video Generation with Human Feedback (2501.13918) proposes a new VideoReward Model and approach that uses human feedback to refine video generation models.

4. DiffuEraser video inpainting model, based on stable diffusion, is designed to fill in missing areas with detailed, realistic content and to ensure consistent structures across frames. DiffuEraser: A Diffusion Model for Video Inpainting (2501.10018)

5. MAGI is a hybrid video gen model that combines masked and casual modeling. Its key innovation, Complete Teacher Forcing (CTF), conditions masked frames on fully visible frames. Taming Teacher Forcing for Masked Autoregressive Video Generation (2501.12389)

6. Go-with-the-Flow: Motion-Controllable Video Diffusion Models Using Real-Time Warped Noise (2501.08331) proposes motion control, allowing users to guide how objects or the camera move in generated videos. Its noise warping algorithm replaces random noise in videos with structured noise based on motion info.

7. Video Depth Anything model estimates depth consistently in super-long videos (several minutes or more) without sacrificing quality or speed. Video Depth Anything: Consistent Depth Estimation for Super-Long Videos (2501.12375)
  • 1 reply
ยท
reacted to clem's post with ๐Ÿ”ฅ 13 days ago
posted an update 13 days ago