IgnoraZ commited on
Commit
b870ba8
·
verified ·
1 Parent(s): bf67e70

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +76 -3
README.md CHANGED
@@ -1,3 +1,76 @@
1
- ---
2
- license: cc-by-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-4.0
3
+ datasets:
4
+ - IgnoraZ/SynthQuestions
5
+ language:
6
+ - en
7
+ base_model:
8
+ - meta-llama/Meta-Llama-3-8B
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+ This is the model from the paper **From Real to Synthetic: Synthesizing Millions of Diversified and Complicated User Instructions with Attributed Grounding**.
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+ - **Model type:** Chat Model
24
+ - **Language(s) (NLP):** English
25
+ - **License:** CC-BY-4.0
26
+ - **Finetuned from model:** LLaMA-3-8B
27
+ - **Finetuned with data:** 1M dataset from `IgnoraZ/SynthQuestions`
28
+
29
+ For more details like hyper-parameters, please refer to our paper.
30
+
31
+ ### Model Sources
32
+
33
+ <!-- Provide the basic links for the model. -->
34
+
35
+ - **Repository:** https://github.com/Ignoramus0817/SynthQuestions
36
+ - **Paper:** https://www.arxiv.org/abs/2506.03968
37
+
38
+ ## How to Get Started with the Model
39
+
40
+ This is a model in HF format, which can be deployed with common inference frameworks like Transformers, vLLM, SGLang and so on.
41
+ We finetuned it with custom chat template instead of the default one from LLaMA. **Please make sure to use the chat template in the `tokenizer_config.json` when inferring.**
42
+
43
+ ## Evaluation
44
+
45
+ <!-- This section describes the evaluation protocols and provides the results. -->
46
+
47
+ ### Alignment Benchmarks
48
+ | Model | Arena Hard (WR%) | Alpaca Eval 2.0 (LC) |
49
+ | :------------: | :--------------: | :------------------: |
50
+ | SynthQuestions | 15.4 | 18.87 |
51
+
52
+
53
+ ### Closed-form Benchmarks
54
+ | Model | IFEVAL | MMLU | ARC-C | GPQA | GSM8K | MATH |
55
+ | :------------: | :----: | :---: | :---: | :--: | :---: | :---: |
56
+ | SynthQuestions | 57.05 | 65.79 | 63.92 | 30.3 | 70.53 | 22.71 |
57
+
58
+ ## Citation
59
+
60
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
61
+
62
+ ```
63
+ @misc{zhu2025realsyntheticsynthesizingmillions,
64
+ title={From Real to Synthetic: Synthesizing Millions of Diversified and Complicated User Instructions with Attributed Grounding},
65
+ author={Chiwei Zhu and Benfeng Xu and Xiaorui Wang and Zhendong Mao},
66
+ year={2025},
67
+ eprint={2506.03968},
68
+ archivePrefix={arXiv},
69
+ primaryClass={cs.CL},
70
+ url={https://arxiv.org/abs/2506.03968},
71
+ }
72
+ ```
73
+
74
+ ## Model Card Contact
75
+
76
+ Please contact [email protected].