Hwanjun commited on
Commit
0d555ec
1 Parent(s): 6645da9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -3
README.md CHANGED
@@ -7,13 +7,18 @@ pipeline_tag: summarization
7
  <b style="font-size: 40px;">SummLlama3-8B</b>
8
  </div>
9
 
10
- Are you looking for a summarizer that can generate more **human-preferred summaries**?
11
 
12
  Our **SummLlama3-8B** could be exactly what you need!
13
 
14
- SummLlama3 is initialized from Llama3-8B-Instruct, with additional training using Direct Preference Optimization (DPO) based on human-like summarization feedback.
15
 
16
- It outperforms the nearly 10x larger Llama3-70B-Instruct while offering much faster inference speed.
 
 
 
 
 
17
 
18
  Please refer to [our paper](link) to catch up how to exploit LLM-generated feedback in the context of text summarization.
19
 
 
7
  <b style="font-size: 40px;">SummLlama3-8B</b>
8
  </div>
9
 
10
+ Are you looking for a summarizer that can generate more **human-preferred summaries** across multiple domains?
11
 
12
  Our **SummLlama3-8B** could be exactly what you need!
13
 
14
+ SummLlama3 is initialized from Llama3-8B-Instruct, with additional training using Direct Preference Optimization (DPO) based on large-scale (over 100K) summarization feedback.
15
 
16
+ The feedback encompasses a wide range of input documents, from short to lengthy texts, including both dialogue and non-dialogue formats, and spans across seven distinct domains:
17
+
18
+ - Four non-dialouge domains: News, Lifestyle, Report, Medical
19
+ - Three dialogue domains: Daily Life, Interview, Meeting
20
+
21
+ Surprisingly, it outperforms the nearly 10x larger Llama3-70B-Instruct while offering much faster inference speed.
22
 
23
  Please refer to [our paper](link) to catch up how to exploit LLM-generated feedback in the context of text summarization.
24