Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
---
|
2 |
-
title:
|
3 |
-
emoji:
|
4 |
colorFrom: blue
|
5 |
colorTo: purple
|
6 |
sdk: gradio
|
@@ -11,4 +11,106 @@ license: mit
|
|
11 |
short_description: 'Compact LLM Battle Arena: Frugal AI Face-Off!'
|
12 |
---
|
13 |
|
14 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
title: GPU Poor LLM Arena
|
3 |
+
emoji: ๐
|
4 |
colorFrom: blue
|
5 |
colorTo: purple
|
6 |
sdk: gradio
|
|
|
11 |
short_description: 'Compact LLM Battle Arena: Frugal AI Face-Off!'
|
12 |
---
|
13 |
|
14 |
+
# ๐ GPU-Poor LLM Gladiator Arena ๐
|
15 |
+
|
16 |
+
Welcome to the GPU-Poor LLM Gladiator Arena, where frugal meets fabulous in the world of AI! This project pits compact language models (maxing out at 9B parameters) against each other in a battle of wits and words.
|
17 |
+
|
18 |
+
## ๐ค Starting from "Why?"
|
19 |
+
|
20 |
+
In the recent months, We've seen a lot of these "Tiny" models released, and some of them are really impressive.
|
21 |
+
|
22 |
+
- **Gradio Exploration**: This project serves me as a playground for experimenting with Gradio app development, I am learning how to create interactive AI interfaces with it.
|
23 |
+
|
24 |
+
- **Tiny Model Evaluation**: I wanted to develop a personal (and now public) stats system for evaluating tiny language models. It's not too serious, but it provides valuable insights into the capabilities of these compact powerhouses.
|
25 |
+
|
26 |
+
- **Accessibility**: Built on Ollama, this arena allows pretty much anyone to experiment with these models themselves. No need for expensive GPUs or cloud services!
|
27 |
+
|
28 |
+
- **Pure Fun**: At its core, this project is about having fun with AI. It's a lighthearted way to explore and compare different models. So, haters, feel free to chill โ we're just here for a good time!
|
29 |
+
|
30 |
+
|
31 |
+
## ๐ Features
|
32 |
+
|
33 |
+
- **Battle Arena**: Pit two mystery models against each other and decide which pint-sized powerhouse reigns supreme.
|
34 |
+
- **Leaderboard**: Track the performance of different models over time.
|
35 |
+
- **Performance Chart**: Visualize model performance with interactive charts.
|
36 |
+
- **Privacy-Focused**: Uses local Ollama API, avoiding pricey commercial APIs and keeping data close to home.
|
37 |
+
- **Customizable**: Easy to add new models and prompts.
|
38 |
+
|
39 |
+
## ๐ Getting Started
|
40 |
+
|
41 |
+
### Prerequisites
|
42 |
+
|
43 |
+
- Python 3.7+
|
44 |
+
- Gradio
|
45 |
+
- Plotly
|
46 |
+
- Ollama (running locally)
|
47 |
+
|
48 |
+
### Installation
|
49 |
+
|
50 |
+
1. Clone the repository:
|
51 |
+
```
|
52 |
+
git clone https://github.com/yourusername/gpu-poor-llm-gladiator-arena.git
|
53 |
+
cd gpu-poor-llm-gladiator-arena
|
54 |
+
```
|
55 |
+
|
56 |
+
2. Install the required packages:
|
57 |
+
```
|
58 |
+
pip install gradio plotly requests
|
59 |
+
```
|
60 |
+
|
61 |
+
3. Ensure Ollama is running locally or via a remote server.
|
62 |
+
|
63 |
+
4. Run the application:
|
64 |
+
```
|
65 |
+
python app.py
|
66 |
+
```
|
67 |
+
|
68 |
+
## ๐ฎ How to Use
|
69 |
+
|
70 |
+
1. Open the application in your web browser (typically at `http://localhost:7860`).
|
71 |
+
2. In the "Battle Arena" tab:
|
72 |
+
- Enter a prompt or use the random prompt generator (๐ฒ button).
|
73 |
+
- Click "Generate Responses" to see outputs from two random models.
|
74 |
+
- Vote for the better response.
|
75 |
+
3. Check the "Leaderboard" tab to see overall model performance.
|
76 |
+
4. View the "Performance Chart" tab for a visual representation of model wins and losses.
|
77 |
+
|
78 |
+
## ๐ Configuration
|
79 |
+
|
80 |
+
You can customize the arena by modifying the `arena_config.py` file:
|
81 |
+
|
82 |
+
- Add or remove models from the `APPROVED_MODELS` list.
|
83 |
+
- Adjust the `API_URL` and `API_KEY` if needed.
|
84 |
+
- Customize `example_prompts` for more variety in random prompts.
|
85 |
+
|
86 |
+
## ๐ Leaderboard
|
87 |
+
|
88 |
+
The leaderboard data is stored in `leaderboard.json`. This file is automatically updated after each battle.
|
89 |
+
|
90 |
+
## ๐ค Models
|
91 |
+
|
92 |
+
The arena currently supports various compact models, including:
|
93 |
+
|
94 |
+
- LLaMA 3.2 (1B and 3B versions)
|
95 |
+
- LLaMA 3.1 (8B version)
|
96 |
+
- Gemma 2 (2B and 9B versions)
|
97 |
+
- Qwen 2.5 (0.5B, 1.5B, 3B, and 7B versions)
|
98 |
+
- Mistral 0.3 (7B version)
|
99 |
+
- Phi 3.5 (3.8B version)
|
100 |
+
- Hermes 3 (8B version)
|
101 |
+
- Aya 23 (8B version)
|
102 |
+
|
103 |
+
## ๐ค Contributing
|
104 |
+
|
105 |
+
Contributions are welcome! Feel free to suggest a model, which is supported by Ollama. Some results are already quite surprising.
|
106 |
+
|
107 |
+
## ๐ License
|
108 |
+
|
109 |
+
This project is open-source and available under the MIT License
|
110 |
+
|
111 |
+
## ๐ Acknowledgements
|
112 |
+
|
113 |
+
- Thanks to the Ollama team for providing that amazing tool.
|
114 |
+
- Shoutout to all the AI researchers and compact language models teams, making this frugal AI arena possible!
|
115 |
+
|
116 |
+
Enjoy the battles in the GPU-Poor LLM Gladiator Arena! May the best compact model win! ๐
|