Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,8 @@ language:
|
|
10 |
# About
|
11 |
This model is Lightblue's QLoRA finetune of OpenOrca's [Open-Orca/OpenOrcaxOpenChat-Preview2-13B](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) model on Japanese fine-tuning datasets.
|
12 |
|
|
|
|
|
13 |
We trained on equal samples of the following three datasets:
|
14 |
* [SNOW](https://huggingface.co/datasets/snow_simplified_japanese_corpus)
|
15 |
* [TyDiQA (Ja)](https://huggingface.co/datasets/khalidalt/tydiqa-goldp)
|
@@ -20,6 +22,18 @@ which resulted in a dataset of 13,167 samples total.
|
|
20 |
These three datasets were chosen as they represent three distinct fine-tuning tasks (Text simplification, question answering, and text summarization, respectively) which we hypothesize can help to improve the language models suitability for dealing with Japanese data.
|
21 |
These three datasets make up the model name: STX.
|
22 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
23 |
# How to use
|
24 |
|
25 |
```python
|
|
|
10 |
# About
|
11 |
This model is Lightblue's QLoRA finetune of OpenOrca's [Open-Orca/OpenOrcaxOpenChat-Preview2-13B](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) model on Japanese fine-tuning datasets.
|
12 |
|
13 |
+
This model specialises on answering **Closed Question Answering** in Japanese. Input a piece of reference text, ask a question, and see the model answer based on the reference text.
|
14 |
+
|
15 |
We trained on equal samples of the following three datasets:
|
16 |
* [SNOW](https://huggingface.co/datasets/snow_simplified_japanese_corpus)
|
17 |
* [TyDiQA (Ja)](https://huggingface.co/datasets/khalidalt/tydiqa-goldp)
|
|
|
22 |
These three datasets were chosen as they represent three distinct fine-tuning tasks (Text simplification, question answering, and text summarization, respectively) which we hypothesize can help to improve the language models suitability for dealing with Japanese data.
|
23 |
These three datasets make up the model name: STX.
|
24 |
|
25 |
+
With these datasets, we achieve the following scores on the JGLUE benchmark:
|
26 |
+
|
27 |
+
| Model Name | Open-Orca/OpenOrcaxOpenChat-Preview2-13B | lightblue/openorca_stx |
|
28 |
+
|------------------------|------------------------------------------|------------------------|
|
29 |
+
| jsquad-1.1-0.3 | 0.692 | 0.836 |
|
30 |
+
| jcommonsenseqa-1.1-0.3 | 0.831 | 0.782 |
|
31 |
+
| jnli-1.1-0.3 | 0.504 | 0.48 |
|
32 |
+
| marc_ja-1.1-0.3 | 0.936 | 0.959 |
|
33 |
+
|
34 |
+
Our model achieves much better results on the question answering benchmark (JSQuAD) than the base checkpoint without monstrous degradation of performance on multi-choice question benchmarks (JCommonSense, JNLI, MARC-Ja) purely through QLoRA training.
|
35 |
+
This shows the potential for applying strong language models such as [Open-Orca/OpenOrcaxOpenChat-Preview2-13B](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) to minimal QLoRA fine-tuning using Japanese fine-tuning datasets to achieve better results at narrow NLP tasks.
|
36 |
+
|
37 |
# How to use
|
38 |
|
39 |
```python
|