Update README.md
Browse files
README.md
CHANGED
@@ -15,7 +15,7 @@ We evaluate the code solution generation ability of TLLM on three benchmarks: Wi
|
|
15 |
| :------------------- | :----: | :----: | :----: | :-----: | :-----: | :----: | :----------: | :-----: |
|
16 |
| TaPEX | 38.5 | β | β | β | 83.9 | 15.0 | / | 45.8 |
|
17 |
| TaPas | 31.5 | β | β | β | 74.2 | 23.1 | / | 42.92 |
|
18 |
-
| TableLlama | 24.0 | 22.2 |
|
19 |
| GPT3.5 | 58.5 |<ins>72.1</ins>| 71.2 | 60.8 | 81.7 | 67.4 | 77.1 | 69.8 |
|
20 |
| GPT4 |**74.1**|**77.1**|**78.4**|**69.5** | 84.0 | 69.5 | 77.8 | **75.8**|
|
21 |
| Llama2-Chat (13B) | 48.8 | 49.6 | 67.7 | 61.5 | β | β | β | 56.9 |
|
|
|
15 |
| :------------------- | :----: | :----: | :----: | :-----: | :-----: | :----: | :----------: | :-----: |
|
16 |
| TaPEX | 38.5 | β | β | β | 83.9 | 15.0 | / | 45.8 |
|
17 |
| TaPas | 31.5 | β | β | β | 74.2 | 23.1 | / | 42.92 |
|
18 |
+
| TableLlama | 24.0 | 22.2 | 20.5 | 6.4 | 43.7 | 9.0 | / | 20.7 |
|
19 |
| GPT3.5 | 58.5 |<ins>72.1</ins>| 71.2 | 60.8 | 81.7 | 67.4 | 77.1 | 69.8 |
|
20 |
| GPT4 |**74.1**|**77.1**|**78.4**|**69.5** | 84.0 | 69.5 | 77.8 | **75.8**|
|
21 |
| Llama2-Chat (13B) | 48.8 | 49.6 | 67.7 | 61.5 | β | β | β | 56.9 |
|