Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ FuxiTranyu-8B is an **open-source** **multilingual large language model** traine
|
|
10 |
|
11 |
FuxiTranyu supports 43 natural languages (Arabic, Bengali, Bulgarian, Burmese, Catalan, Chinese, Czech, Dutch, English, Filipino, Finnish, French, German, Greek, Hebrew, Hindi, Hungarian, Indonesian, Italian, Japanese, Kazakh, Khmer, Korean, Kurdish, Kyrgyz, Lao, Malay, Persian, Polish, Portuguese, Romanian, Russian, Spanish, Swedish, Tamil, Tajik, Thai, Turkish, Turkmen, Ukrainian, Urdu, Uzbek, and Vietnamese) and cover 16 programming languages (Java, JavaScript, Python, PHP, C, C++, C#, TypeScript, Go, SQL, Rust, Ruby, Scala, Lua, Assembly, and Visual Basic).
|
12 |
|
13 |
-
More details on the data collection & processing, pretraining and fine-tuning of FuxiTranyu can be found in the technical report.
|
14 |
|
15 |
In addition to the base model and its checkpoints, we also release two instruction-tuned variants: SFT version at [here](https://huggingface.co/TJUNLP/FuxiTranyu-8B-Instruct), and DPO version at [here](https://huggingface.co/TJUNLP/FuxiTranyu-8B-Chat).
|
16 |
## Usage
|
@@ -42,12 +42,11 @@ model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto", torc
|
|
42 |
## Citation info
|
43 |
|
44 |
```bibtex
|
45 |
-
@
|
46 |
title={FuxiTranyu: A Multilingual Large Language Model Trained with Balanced Data},
|
47 |
author={Haoran Sun, Renren Jin, Shaoyang Xu, Leiyu Pan, Supryadi, Menglong Cui, Jiangcun Du, Yikun Lei, Lei Yang, Ling Shi, Juesi Xiao, Shaolin Zhu, and Deyi Xiong},
|
|
|
48 |
year={2024},
|
49 |
-
|
50 |
-
archivePrefix={arXiv},
|
51 |
-
primaryClass={cs.CL}
|
52 |
}
|
53 |
```
|
|
|
10 |
|
11 |
FuxiTranyu supports 43 natural languages (Arabic, Bengali, Bulgarian, Burmese, Catalan, Chinese, Czech, Dutch, English, Filipino, Finnish, French, German, Greek, Hebrew, Hindi, Hungarian, Indonesian, Italian, Japanese, Kazakh, Khmer, Korean, Kurdish, Kyrgyz, Lao, Malay, Persian, Polish, Portuguese, Romanian, Russian, Spanish, Swedish, Tamil, Tajik, Thai, Turkish, Turkmen, Ukrainian, Urdu, Uzbek, and Vietnamese) and cover 16 programming languages (Java, JavaScript, Python, PHP, C, C++, C#, TypeScript, Go, SQL, Rust, Ruby, Scala, Lua, Assembly, and Visual Basic).
|
12 |
|
13 |
+
More details on the data collection & processing, pretraining and fine-tuning of FuxiTranyu can be found in the [technical report](https://arxiv.org/abs/2408.06273).
|
14 |
|
15 |
In addition to the base model and its checkpoints, we also release two instruction-tuned variants: SFT version at [here](https://huggingface.co/TJUNLP/FuxiTranyu-8B-Instruct), and DPO version at [here](https://huggingface.co/TJUNLP/FuxiTranyu-8B-Chat).
|
16 |
## Usage
|
|
|
42 |
## Citation info
|
43 |
|
44 |
```bibtex
|
45 |
+
@article{FuxiTranyu8B,
|
46 |
title={FuxiTranyu: A Multilingual Large Language Model Trained with Balanced Data},
|
47 |
author={Haoran Sun, Renren Jin, Shaoyang Xu, Leiyu Pan, Supryadi, Menglong Cui, Jiangcun Du, Yikun Lei, Lei Yang, Ling Shi, Juesi Xiao, Shaolin Zhu, and Deyi Xiong},
|
48 |
+
journal={arxiv preprint arXiv:2408.06273},
|
49 |
year={2024},
|
50 |
+
url={https://arxiv.org/abs/2408.06273}
|
|
|
|
|
51 |
}
|
52 |
```
|