Text Generation
Transformers
PyTorch
Japanese
llama
text-generation-inference
Inference Endpoints
ptrdvn commited on
Commit
f72c6f1
·
1 Parent(s): c2fc322

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -0
README.md CHANGED
@@ -10,6 +10,8 @@ language:
10
  # About
11
  This model is Lightblue's QLoRA finetune of OpenOrca's [Open-Orca/OpenOrcaxOpenChat-Preview2-13B](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) model on Japanese fine-tuning datasets.
12
 
 
 
13
  We trained on equal samples of the following three datasets:
14
  * [SNOW](https://huggingface.co/datasets/snow_simplified_japanese_corpus)
15
  * [TyDiQA (Ja)](https://huggingface.co/datasets/khalidalt/tydiqa-goldp)
@@ -20,6 +22,18 @@ which resulted in a dataset of 13,167 samples total.
20
  These three datasets were chosen as they represent three distinct fine-tuning tasks (Text simplification, question answering, and text summarization, respectively) which we hypothesize can help to improve the language models suitability for dealing with Japanese data.
21
  These three datasets make up the model name: STX.
22
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  # How to use
24
 
25
  ```python
 
10
  # About
11
  This model is Lightblue's QLoRA finetune of OpenOrca's [Open-Orca/OpenOrcaxOpenChat-Preview2-13B](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) model on Japanese fine-tuning datasets.
12
 
13
+ This model specialises on answering **Closed Question Answering** in Japanese. Input a piece of reference text, ask a question, and see the model answer based on the reference text.
14
+
15
  We trained on equal samples of the following three datasets:
16
  * [SNOW](https://huggingface.co/datasets/snow_simplified_japanese_corpus)
17
  * [TyDiQA (Ja)](https://huggingface.co/datasets/khalidalt/tydiqa-goldp)
 
22
  These three datasets were chosen as they represent three distinct fine-tuning tasks (Text simplification, question answering, and text summarization, respectively) which we hypothesize can help to improve the language models suitability for dealing with Japanese data.
23
  These three datasets make up the model name: STX.
24
 
25
+ With these datasets, we achieve the following scores on the JGLUE benchmark:
26
+
27
+ | Model Name | Open-Orca/OpenOrcaxOpenChat-Preview2-13B | lightblue/openorca_stx |
28
+ |------------------------|------------------------------------------|------------------------|
29
+ | jsquad-1.1-0.3 | 0.692 | 0.836 |
30
+ | jcommonsenseqa-1.1-0.3 | 0.831 | 0.782 |
31
+ | jnli-1.1-0.3 | 0.504 | 0.48 |
32
+ | marc_ja-1.1-0.3 | 0.936 | 0.959 |
33
+
34
+ Our model achieves much better results on the question answering benchmark (JSQuAD) than the base checkpoint without monstrous degradation of performance on multi-choice question benchmarks (JCommonSense, JNLI, MARC-Ja) purely through QLoRA training.
35
+ This shows the potential for applying strong language models such as [Open-Orca/OpenOrcaxOpenChat-Preview2-13B](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) to minimal QLoRA fine-tuning using Japanese fine-tuning datasets to achieve better results at narrow NLP tasks.
36
+
37
  # How to use
38
 
39
  ```python