Update README.md
Browse files
README.md
CHANGED
@@ -7,13 +7,18 @@ pipeline_tag: summarization
|
|
7 |
<b style="font-size: 40px;">SummLlama3-8B</b>
|
8 |
</div>
|
9 |
|
10 |
-
Are you looking for a summarizer that can generate more **human-preferred summaries
|
11 |
|
12 |
Our **SummLlama3-8B** could be exactly what you need!
|
13 |
|
14 |
-
SummLlama3 is initialized from Llama3-8B-Instruct, with additional training using Direct Preference Optimization (DPO) based on
|
15 |
|
16 |
-
|
|
|
|
|
|
|
|
|
|
|
17 |
|
18 |
Please refer to [our paper](link) to catch up how to exploit LLM-generated feedback in the context of text summarization.
|
19 |
|
|
|
7 |
<b style="font-size: 40px;">SummLlama3-8B</b>
|
8 |
</div>
|
9 |
|
10 |
+
Are you looking for a summarizer that can generate more **human-preferred summaries** across multiple domains?
|
11 |
|
12 |
Our **SummLlama3-8B** could be exactly what you need!
|
13 |
|
14 |
+
SummLlama3 is initialized from Llama3-8B-Instruct, with additional training using Direct Preference Optimization (DPO) based on large-scale (over 100K) summarization feedback.
|
15 |
|
16 |
+
The feedback encompasses a wide range of input documents, from short to lengthy texts, including both dialogue and non-dialogue formats, and spans across seven distinct domains:
|
17 |
+
|
18 |
+
- Four non-dialouge domains: News, Lifestyle, Report, Medical
|
19 |
+
- Three dialogue domains: Daily Life, Interview, Meeting
|
20 |
+
|
21 |
+
Surprisingly, it outperforms the nearly 10x larger Llama3-70B-Instruct while offering much faster inference speed.
|
22 |
|
23 |
Please refer to [our paper](link) to catch up how to exploit LLM-generated feedback in the context of text summarization.
|
24 |
|