Update README.md
Browse files
README.md
CHANGED
@@ -26,11 +26,12 @@ need to check whether this phenomenon is repeated in larger models (3B, 8B).
|
|
26 |
|
27 |
## Model Performance Comparison (BFCL)
|
28 |
|
29 |
-
| task name | minpeter/Llama-3.2-1B-chatml-tool-v2 | meta-llama/Llama-3.2-1B-Instruct |
|
30 |
-
|
31 |
-
| parallel_multiple | 0.000 | 0.025 |
|
32 |
-
| parallel | 0.000 | 0.035 |
|
33 |
-
| simple | 0.72 | 0.215 |
|
34 |
-
| multiple | 0.695 | 0.17 |
|
|
|
35 |
|
36 |
*Parallel calls are not taken into account. 0 points are expected. We plan to fix this in v3.
|
|
|
26 |
|
27 |
## Model Performance Comparison (BFCL)
|
28 |
|
29 |
+
| task name | minpeter/Llama-3.2-1B-chatml-tool-v2 | meta-llama/Llama-3.2-1B-Instruct (measure) | meta-llama/Llama-3.2-1B-Instruct (Reported) |
|
30 |
+
|-----------------|-----------------------------------|-----------------------------------|---------------------------------------|
|
31 |
+
| parallel_multiple | 0.000 | 0.025 | **0.15** |
|
32 |
+
| parallel | 0.000 | 0.035 | **0.36** |
|
33 |
+
| simple | **0.72** | 0.215 | 0.2925 |
|
34 |
+
| multiple | **0.695** | 0.17 | 0.335 |
|
35 |
+
|
36 |
|
37 |
*Parallel calls are not taken into account. 0 points are expected. We plan to fix this in v3.
|