ykhwang commited on
Commit
0af0692
·
1 Parent(s): 13052b8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -26,7 +26,7 @@ As same as 42dot-PLM, the model is built upon a Transformer decoder architecture
26
  (\* unit: tokens)
27
  ### Supervised Fine-tuning
28
 
29
- Fine-tuning took about 4 hours using 8 * NVIDIA A100 GPUs. For training dataset, we manually constructed (question/insturuction) and response pairs, which can either be single- or multi-turn.
30
 
31
  ### Evaluation
32
  Inspired by recent attempts like [Vicuna](https://lmsys.org/blog/2023-03-30-vicuna/#how-good-is-vicuna), we evaluate 42dot-PLM with other proprietary/open-sourced chatbots using GPT-4 for assesing various aspects of responses. The sample of evaluation dataset and prompt template can be downloaded from our [GitHub repo](https://github.com/42dot/42dot_LLM).
 
26
  (\* unit: tokens)
27
  ### Supervised Fine-tuning
28
 
29
+ Fine-tuning took about 10 hours using 8 * NVIDIA A100 GPUs. For training dataset, we manually constructed (question/insturuction) and response pairs, which can either be single- or multi-turn.
30
 
31
  ### Evaluation
32
  Inspired by recent attempts like [Vicuna](https://lmsys.org/blog/2023-03-30-vicuna/#how-good-is-vicuna), we evaluate 42dot-PLM with other proprietary/open-sourced chatbots using GPT-4 for assesing various aspects of responses. The sample of evaluation dataset and prompt template can be downloaded from our [GitHub repo](https://github.com/42dot/42dot_LLM).