Hwanjun commited on
Commit
4ce822e
·
verified ·
1 Parent(s): 5cd4b71

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -7,12 +7,14 @@ pipeline_tag: summarization
7
  <b style="font-size: 40px;">SummLlama3-8B</b>
8
  </div>
9
 
10
- Are you looking for a summarizer that outperforms the nearly 10x larger Llama3-70B-Instruct while offering much faster inference speed?
11
 
12
  Our **SummLlama3-8B** could be exactly what you need!
13
 
14
  SummLlama3 is initialized from Llama3-8B-Instruct, with additional training using Direct Preference Optimization (DPO) based on human-like summarization feedback.
15
 
 
 
16
  Please refer to [our paper](link) to catch up how to exploit LLM-generated feedback in the context of text summarization.
17
 
18
  We also released a larger model, **SummLlama3-70B**. Please go to the [Huggingface link](https://huggingface.co/DISLab/SummLlama3-70B) for this model.
 
7
  <b style="font-size: 40px;">SummLlama3-8B</b>
8
  </div>
9
 
10
+ Are you looking for a summarizer that can generate more **human-preferred summaries**?
11
 
12
  Our **SummLlama3-8B** could be exactly what you need!
13
 
14
  SummLlama3 is initialized from Llama3-8B-Instruct, with additional training using Direct Preference Optimization (DPO) based on human-like summarization feedback.
15
 
16
+ It outperforms the nearly 10x larger Llama3-70B-Instruct while offering much faster inference speed.
17
+
18
  Please refer to [our paper](link) to catch up how to exploit LLM-generated feedback in the context of text summarization.
19
 
20
  We also released a larger model, **SummLlama3-70B**. Please go to the [Huggingface link](https://huggingface.co/DISLab/SummLlama3-70B) for this model.