KinglyWayne's picture
Update README.md
bd123e3 verified
|
raw
history blame
14 kB
metadata
language:
  - zh
pipeline_tag: text-generation
license: apache-2.0
task_categories:
  - text-generation
size_categories:
  - 10B<n<100B

Chinese Fineweb Edu Dataset V2 [中文] [English]

OpenCSG

[OpenCSG Community] [github] [wechat] [Twitter]

Chinese Fineweb Edu Dataset V2 is a comprehensive upgrade of the original Chinese Fineweb Edu, designed and optimized for natural language processing (NLP) tasks in the education sector. This high-quality Chinese pretraining dataset has undergone significant improvements and expansions, aimed at providing researchers and developers with more diverse and broadly applicable educational corpus resources. With a dataset size of 188 million entries (approximately 420 billion tokens), Fineweb Edu v2 not only increases the volume but also optimizes the data filtering methods and scoring models to ensure effectiveness and practicality in the educational domain.

Enhanced Scoring Model

In the Chinese Fineweb edu v2 version, the data selection scoring model has undergone a significant upgrade, utilizing the larger and more powerful OpenCSG csg-wukong-enterprise V2 model. The training data for this model has been increased to 1 million entries, covering a variety of text types such as books, news, blogs, and 25% English data. Compared to the previous version, the csg-wukong-enterprise V2 model boasts a larger parameter count and deeper semantic understanding, excelling particularly in Chinese text comprehension and processing. The model not only performs more detailed analysis of text structure and content but also captures deeper semantic and emotional nuances embedded in the language.

This improvement means that during the data selection process, the model can more accurately assess the educational value, writing quality, and practical application of the text. Especially when dealing with high-demand texts in education and technology, the Fineweb2 scoring model ensures high quality and consistency in the selection results. This advancement significantly enhances the reliability of the data selection, providing stronger support for subsequent model training.

Prompt Improvements

During the construction of the Fineweb2 dataset, the data filtering process was particularly crucial. To ensure that only text with real educational value and practicality was selected, we carefully optimized the design of the prompts used for data filtering. The new prompts more accurately evaluate the educational value, writing quality, and practicality of web content, refining the filtering process for better precision.

The new prompts clearly define scoring standards for educational content and also set expectations for writing style, coherence, and thematic depth. The specific scoring criteria are as follows:

Below is an excerpt from a web page. Please use the following 5-point rating system to assess the writing quality, educational value, and practicality of the webpage:

以下是一段网页内容摘录。请使用以下5分制评分系统来评估该网页的写作水平、教育价值和实用性:
0分:如果网页没有提供任何教育价值,完全由无关信息(如广告、宣传材料、少儿不宜内容)组成。
1分:如果网页提供了一些可能有教育价值的基本信息,但包含较多的无关或非学术内容(如广告和宣传材料)。
2分:如果网页涉及某些与教育相关的元素,但与教育标准不太吻合。它可能将教育内容与非教育材料混杂,对潜在的有用的主题进行浅显概述,或以不连贯的写作风格呈现信息。
3分:如果网页适合教育使用,并介绍了与某些学校课程中可能学到的关键概念,或对个人发展有用的实用信息。它的内容连贯但可能不全面,或包含一些无关信息。它可能类似于教科书的一小段节选,可以学习但有明显局限,如涉及过于复杂的概念、过于具体的不重要事件。
4分:如果网页与教育高度相关,对个人学习发展有益,表现出清晰一致的写作风格。它可能类似于教科书的一个章节或教程,提供大量教育内容,极少包含无关信息,且概念对学生来说不会过于深奥。内容连贯、重点突出,对结构化学习有价值。
5分:如果网页摘录在教育价值上表现极好,完全适合小学、中学或大学教学或专业人士学习。它遵循详细的推理过程,写作风格易于理解,对主题提供深刻而全面的见解,不包含任何非教育性或无实用意义内容。

网页内容摘录:
{}

在审查这段网页摘录后:请简要地为您的评分进行合理的解释,最多不超过100字,最后以“教育得分:<分数>”的格式结束。请根据所列出的标准系统地赋予分数。

After reviewing this webpage excerpt, briefly explain the reasoning behind your score in no more than 100 words, ending with the format: "Educational Score: ." Please assign the score systematically based on the listed criteria.

After merging all data, the sample score distribution was as follows: texts with scores of 3 and above were selected, totaling 188 million entries (about 420 billion tokens). These data, which are not only extensive but also carefully filtered and deduplicated, ensure the high quality and uniqueness of the dataset. These scored data will be used to train large-scale language models within the Fineweb2 dataset, helping them achieve superior performance in various tasks.

experiment

Expanded Data Sources

The range of data sources for the Fineweb2 dataset has been further extended. Compared to the original Fineweb, Fineweb2 introduces massive datasets from various fields and sources, including Industry2, CCI3, michao, wanjuan1.0, wudao, and ChineseWebText. These datasets cover a broader range of industries and domains, enhancing the diversity and applicability of the dataset.

experiment

In conclusion, the Fineweb2 dataset not only surpasses its predecessor in scale but also significantly improves the quality of data, content diversity, and precision of filtering. This lays a solid foundation for the further development of Chinese NLP applications and provides researchers with richer resources to explore and optimize various model training methods.

We warmly invite developers and researchers interested in this field to follow and engage with the community, working together to advance the technology. Stay tuned for the open-source release of the dataset!

License Agreement

Usage of the Chinese Fineweb Edu dataset requires adherence to the OpenCSG Community License. The Chinese Fineweb Edu dataset supports commercial use. If you plan to use the OpenCSG model or its derivatives for commercial purposes, you must comply with the terms and conditions outlined in the OpenCSG Community License as well as the Apache 2.0 License. For commercial use, please send an email to lorraineg@opencsg.com and obtain permission.

Chinese Fineweb Edu V2数据集介绍

OpenCSG

[OpenCSG 社区] [github] [微信] [推特]

Chinese Fineweb Edu v2 是Chinese Fineweb Edu的全新升级版,专为教育领域的自然语言处理(NLP)任务设计和优化的高质量中文预训练数据集。该数据集在前一版本的基础上进行了大规模的改进和扩展,致力于为研究人员和开发者提供更加多样化、广泛适用的教育类语料资源。Fineweb Edu v2 不仅数据量达到**188M条数据**,约**420B tokens**,还优化了数据的筛选方式和打分模型,以确保其在教育领域的有效性和实用性。

更强的打分模型

在Chinese Fineweb edu v2版本中,数据筛选的打分模型进行了重大升级,采用了规模更大、性能更强的OpenCSG csg-wukong-enterprise V2模型。该模型的训练数据增加到100万条,涵盖了多种类型的文本,如书籍、新闻、博客,以及25%的英文数据。相比于上一版本的打分模型,csg-wukong-enterprise V2拥有更大的参数量和更深层次的语义理解能力,特别是在中文文本理解和处理方面表现出色。该模型不仅能对文本的结构、内容进行更细致的分析,还能有效捕捉隐藏在语言中的深层次语义和情感信息。 这种提升意味着在数据筛选过程中,模型能够更加精准地评估文本的教育价值、写作质量以及其对实际应用的价值。尤其是在处理教育类、技术类等高要求的文本时,Fineweb2的打分模型确保了筛选结果的高质量和高一致性。这一进步显著提高了数据筛选的可靠性,为后续的模型训练提供了更有力的保障。

Prompt改进

在Fineweb2数据集的构建过程中,数据筛选环节尤为重要。为确保筛选出真正具有教育价值和实用性的文本,我们对数据筛选的Prompt设计进行了细致的优化。新的Prompt能够更加准确地评估网页内容的教育价值、写作水平和实用性,从而使筛选过程更加细化和精确。

新的Prompt不仅明确了对教育内容的评分标准,还对文本的写作风格、连贯性以及主题深度提出了要求。具体评分标准如下:

以下是一段网页内容摘录。请使用以下5分制评分系统来评估该网页的写作水平、教育价值和实用性:
0分:如果网页没有提供任何教育价值,完全由无关信息(如广告、宣传材料、少儿不宜内容)组成。
1分:如果网页提供了一些可能有教育价值的基本信息,但包含较多的无关或非学术内容(如广告和宣传材料)。
2分:如果网页涉及某些与教育相关的元素,但与教育标准不太吻合。它可能将教育内容与非教育材料混杂,对潜在的有用的主题进行浅显概述,或以不连贯的写作风格呈现信息。
3分:如果网页适合教育使用,并介绍了与某些学校课程中可能学到的关键概念,或对个人发展有用的实用信息。它的内容连贯但可能不全面,或包含一些无关信息。它可能类似于教科书的一小段节选,可以学习但有明显局限,如涉及过于复杂的概念、过于具体的不重要事件。
4分:如果网页与教育高度相关,对个人学习发展有益,表现出清晰一致的写作风格。它可能类似于教科书的一个章节或教程,提供大量教育内容,极少包含无关信息,且概念对学生来说不会过于深奥。内容连贯、重点突出,对结构化学习有价值。
5分:如果网页摘录在教育价值上表现极好,完全适合小学、中学或大学教学或专业人士学习。它遵循详细的推理过程,写作风格易于理解,对主题提供深刻而全面的见解,不包含任何非教育性或无实用意义内容。

网页内容摘录:
{}

在审查这段网页摘录后:请简要地为您的评分进行合理的解释,最多不超过100字,最后以“教育得分:<分数>”的格式结束。请根据所列出的标准系统地赋予分数。

所有数据集合并后,样本的得分分布如下,通过csg-wukong-enterprise V2模型对这些数据进行评分后,最终选取了3分以上的文本,总计达到188M条数据,约420B tokens。这些数据不仅数量庞大,且经过了严格的筛选和去重处理,确保了数据集的高质量和高独特性。这些经过打分的数据将在Fineweb2的数据集中用于训练大规模语言模型,帮助其在各类任务中实现更高的性能表现。

experiment

数据筛选范围扩大

Fineweb2数据集的数据来源进一步扩展。相较于初代Fineweb,Fineweb2引入了来自多个不同领域和来源的海量数据,新增了Industry2、CCI3、michao、wanjuan1.0、wudao和ChineseWebText等高质量数据集。这些数据集覆盖了更广泛的行业和领域,增加了数据集的多样性和广泛适用性。

experiment

最终,Fineweb2的数据集不仅在规模上远超前作,还在数据的质量、内容的多样性、筛选的精确度等方面有了显著提升。这为未来中文NLP应用的进一步发展打下了坚实的基础,同时也为研究人员提供了更加丰富的资源去探索和优化各种模型训练方法。

我们诚邀对这一领域感兴趣的开发者和研究者关注和联系社区,共同推动技术的进步。敬请期待数据集的开源发布!

许可协议

使用 Chinese Fineweb Edu V2数据集需要遵循 OpenCSG 社区许可证。Chinese Fineweb Edu V2数据集支持商业用途。如果您计划将 OpenCSG 模型或其衍生产品用于商业目的,您必须遵守 OpenCSG 社区许可证以及 Apache 2.0 许可证中的条款和条件。如用于商业用途,需发送邮件至 lorraineg@opencsg.com,并获得许可。