Jade Ethan Terblanche

SugoiLoki

AI & ML interests

dio

Recent Activity

liked a model 16 days ago
coqui/XTTS-v2
liked a model about 2 months ago
OrionStarAI/Orion-14B-Base
View all activity

Organizations

None yet

SugoiLoki's activity

New activity in meta-llama/Llama-3.2-11B-Vision-Instruct about 2 months ago
reacted to csabakecskemeti's post with 👍 about 2 months ago
view post
Post
1244
Some time ago, I built a predictive LLM router that routes chat requests between small and large LLM models based on prompt classification. It dynamically selects the most suitable model depending on the complexity of the user input, ensuring optimal performance while maintaining conversation context. I also fine-tuned a RoBERTa model to use with the package, but you can plug and play any classifier of your choice.

Project's homepage:
https://devquasar.com/llm-predictive-router/
Pypi:
https://pypi.org/project/llm-predictive-router/
Model:
DevQuasar/roberta-prompt_classifier-v0.1
Training data:
DevQuasar/llm_router_dataset-synth
Git:
https://github.com/csabakecskemeti/llm_predictive_router_package

Feel free to check it out, and/or contribute.