hrsun15 commited on
Commit
1d5dd4e
·
verified ·
1 Parent(s): da0abcc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -10,7 +10,7 @@ FuxiTranyu-8B is an **open-source** **multilingual large language model** traine
10
 
11
  FuxiTranyu supports 43 natural languages (Arabic, Bengali, Bulgarian, Burmese, Catalan, Chinese, Czech, Dutch, English, Filipino, Finnish, French, German, Greek, Hebrew, Hindi, Hungarian, Indonesian, Italian, Japanese, Kazakh, Khmer, Korean, Kurdish, Kyrgyz, Lao, Malay, Persian, Polish, Portuguese, Romanian, Russian, Spanish, Swedish, Tamil, Tajik, Thai, Turkish, Turkmen, Ukrainian, Urdu, Uzbek, and Vietnamese) and cover 16 programming languages (Java, JavaScript, Python, PHP, C, C++, C#, TypeScript, Go, SQL, Rust, Ruby, Scala, Lua, Assembly, and Visual Basic).
12
 
13
- More details on the data collection & processing, pretraining and fine-tuning of FuxiTranyu can be found in the technical report.
14
 
15
  In addition to the base model and its checkpoints, we also release two instruction-tuned variants: SFT version at [here](https://huggingface.co/TJUNLP/FuxiTranyu-8B-Instruct), and DPO version at [here](https://huggingface.co/TJUNLP/FuxiTranyu-8B-Chat).
16
  ## Usage
@@ -42,12 +42,11 @@ model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto", torc
42
  ## Citation info
43
 
44
  ```bibtex
45
- @misc{FuxiTranyu8B,
46
  title={FuxiTranyu: A Multilingual Large Language Model Trained with Balanced Data},
47
  author={Haoran Sun, Renren Jin, Shaoyang Xu, Leiyu Pan, Supryadi, Menglong Cui, Jiangcun Du, Yikun Lei, Lei Yang, Ling Shi, Juesi Xiao, Shaolin Zhu, and Deyi Xiong},
 
48
  year={2024},
49
- eprint={2408},
50
- archivePrefix={arXiv},
51
- primaryClass={cs.CL}
52
  }
53
  ```
 
10
 
11
  FuxiTranyu supports 43 natural languages (Arabic, Bengali, Bulgarian, Burmese, Catalan, Chinese, Czech, Dutch, English, Filipino, Finnish, French, German, Greek, Hebrew, Hindi, Hungarian, Indonesian, Italian, Japanese, Kazakh, Khmer, Korean, Kurdish, Kyrgyz, Lao, Malay, Persian, Polish, Portuguese, Romanian, Russian, Spanish, Swedish, Tamil, Tajik, Thai, Turkish, Turkmen, Ukrainian, Urdu, Uzbek, and Vietnamese) and cover 16 programming languages (Java, JavaScript, Python, PHP, C, C++, C#, TypeScript, Go, SQL, Rust, Ruby, Scala, Lua, Assembly, and Visual Basic).
12
 
13
+ More details on the data collection & processing, pretraining and fine-tuning of FuxiTranyu can be found in the [technical report](https://arxiv.org/abs/2408.06273).
14
 
15
  In addition to the base model and its checkpoints, we also release two instruction-tuned variants: SFT version at [here](https://huggingface.co/TJUNLP/FuxiTranyu-8B-Instruct), and DPO version at [here](https://huggingface.co/TJUNLP/FuxiTranyu-8B-Chat).
16
  ## Usage
 
42
  ## Citation info
43
 
44
  ```bibtex
45
+ @article{FuxiTranyu8B,
46
  title={FuxiTranyu: A Multilingual Large Language Model Trained with Balanced Data},
47
  author={Haoran Sun, Renren Jin, Shaoyang Xu, Leiyu Pan, Supryadi, Menglong Cui, Jiangcun Du, Yikun Lei, Lei Yang, Ling Shi, Juesi Xiao, Shaolin Zhu, and Deyi Xiong},
48
+ journal={arxiv preprint arXiv:2408.06273},
49
  year={2024},
50
+ url={https://arxiv.org/abs/2408.06273}
 
 
51
  }
52
  ```