Commit
·
687e701
1
Parent(s):
f7b44cb
Update README.md
Browse files
README.md
CHANGED
@@ -113,7 +113,7 @@ We then conducted their [automatic evaluation](https://github.com/FreedomIntelli
|
|
113 |
This prompt was designed to elicit assessments of answers in terms of helpfulness, relevance, accuracy, and level of detail.
|
114 |
[Additional prompts](https://github.com/FreedomIntelligence/LLMZoo/blob/main/llmzoo/eval/prompts/order/prompt_all.json) are provided for assessing overall performance on different perspectives.
|
115 |
|
116 |
-
Follows the results against GPT-3.5
|
117 |
|
118 |
* Automatic Evaluation **in Portuguese**:
|
119 |
|
|
|
113 |
This prompt was designed to elicit assessments of answers in terms of helpfulness, relevance, accuracy, and level of detail.
|
114 |
[Additional prompts](https://github.com/FreedomIntelligence/LLMZoo/blob/main/llmzoo/eval/prompts/order/prompt_all.json) are provided for assessing overall performance on different perspectives.
|
115 |
|
116 |
+
Follows the results against GPT-3.5 and two of the highest performing open-source models at the moment, Vicuna (13B) and Falcon (40B):
|
117 |
|
118 |
* Automatic Evaluation **in Portuguese**:
|
119 |
|