Update README.md
Browse files
README.md
CHANGED
@@ -26,7 +26,7 @@ As same as 42dot-PLM, the model is built upon a Transformer decoder architecture
|
|
26 |
(\* unit: tokens)
|
27 |
### Supervised Fine-tuning
|
28 |
|
29 |
-
Fine-tuning took about
|
30 |
|
31 |
### Evaluation
|
32 |
Inspired by recent attempts like [Vicuna](https://lmsys.org/blog/2023-03-30-vicuna/#how-good-is-vicuna), we evaluate 42dot-PLM with other proprietary/open-sourced chatbots using GPT-4 for assesing various aspects of responses. The sample of evaluation dataset and prompt template can be downloaded from our [GitHub repo](https://github.com/42dot/42dot_LLM).
|
|
|
26 |
(\* unit: tokens)
|
27 |
### Supervised Fine-tuning
|
28 |
|
29 |
+
Fine-tuning took about 20 hours using 8 * NVIDIA A100 GPUs. For training dataset, we manually constructed (question/insturuction) and response pairs, which can either be single- or multi-turn.
|
30 |
|
31 |
### Evaluation
|
32 |
Inspired by recent attempts like [Vicuna](https://lmsys.org/blog/2023-03-30-vicuna/#how-good-is-vicuna), we evaluate 42dot-PLM with other proprietary/open-sourced chatbots using GPT-4 for assesing various aspects of responses. The sample of evaluation dataset and prompt template can be downloaded from our [GitHub repo](https://github.com/42dot/42dot_LLM).
|