Spaces:
Runtime error
Runtime error
Update src/assets/text_content.py
Browse files
src/assets/text_content.py
CHANGED
@@ -6,6 +6,10 @@ INTRODUCTION_TEXT = f"""
|
|
6 |
🐨 KG LLM Leaderboard aims to track, rank, and evaluate the performance of released Large Language Models on traditional KBQA/KGQA datasets.
|
7 |
|
8 |
The data on this page is sourced from a research paper. If you intend to use the data from this page, please remember to cite the following source: https://arxiv.org/abs/2303.07992
|
|
|
|
|
|
|
|
|
9 |
"""
|
10 |
|
11 |
LLM_BENCHMARKS_TEXT = f"""
|
|
|
6 |
🐨 KG LLM Leaderboard aims to track, rank, and evaluate the performance of released Large Language Models on traditional KBQA/KGQA datasets.
|
7 |
|
8 |
The data on this page is sourced from a research paper. If you intend to use the data from this page, please remember to cite the following source: https://arxiv.org/abs/2303.07992
|
9 |
+
|
10 |
+
We compare the current SOTA traditional KBQA models (fine-tuned (FT) and zero-shot (ZS)),
|
11 |
+
LLMs in the GPT family, and Other Non-GPT LLM. In QALD-9 and LC-quad2, the evaluation metric used is F1, while other datasets use Accuracy (Exact match).
|
12 |
+
|
13 |
"""
|
14 |
|
15 |
LLM_BENCHMARKS_TEXT = f"""
|