developers update
Browse files
README.md
CHANGED
@@ -208,7 +208,7 @@ model-index:
|
|
208 |
**Model Summary:**
|
209 |
Granite-3.0-8B-Base is a decoder-only language model to support a variety of text-to-text generation tasks. It is trained from scratch following a two-stage training strategy. In the first stage, it is trained on 10 trillion tokens sourced from diverse domains. During the second stage, it is further trained on 2 trillion tokens using a carefully curated mix of high-quality data, aiming to enhance its performance on specific tasks.
|
210 |
|
211 |
-
- **Developers:** IBM
|
212 |
- **GitHub Repository:** [ibm-granite/granite-3.0-language-models](https://github.com/ibm-granite/granite-3.0-language-models)
|
213 |
- **Website**: [Granite Docs](https://www.ibm.com/granite/docs/)
|
214 |
- **Paper:** [Granite 3.0 Language Models](https://github.com/ibm-granite/granite-3.0-language-models/blob/main/paper.pdf)
|
@@ -217,6 +217,7 @@ Granite-3.0-8B-Base is a decoder-only language model to support a variety of tex
|
|
217 |
|
218 |
**Supported Languages:**
|
219 |
English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, and Chinese. Users may finetune Granite 3.0 models for languages beyond these 12 languages.
|
|
|
220 |
**Intended use:**
|
221 |
Prominent use cases of LLMs in text-to-text generation include summarization, text classification, extraction, question-answering, and more. All Granite Base models are able to handle these tasks as they were trained on a large amount of data from various domains. Moreover, they can serve as baseline to create specialized models for specific application scenarios.
|
222 |
|
|
|
208 |
**Model Summary:**
|
209 |
Granite-3.0-8B-Base is a decoder-only language model to support a variety of text-to-text generation tasks. It is trained from scratch following a two-stage training strategy. In the first stage, it is trained on 10 trillion tokens sourced from diverse domains. During the second stage, it is further trained on 2 trillion tokens using a carefully curated mix of high-quality data, aiming to enhance its performance on specific tasks.
|
210 |
|
211 |
+
- **Developers:** IBM Granite Team
|
212 |
- **GitHub Repository:** [ibm-granite/granite-3.0-language-models](https://github.com/ibm-granite/granite-3.0-language-models)
|
213 |
- **Website**: [Granite Docs](https://www.ibm.com/granite/docs/)
|
214 |
- **Paper:** [Granite 3.0 Language Models](https://github.com/ibm-granite/granite-3.0-language-models/blob/main/paper.pdf)
|
|
|
217 |
|
218 |
**Supported Languages:**
|
219 |
English, German, Spanish, French, Japanese, Portuguese, Arabic, Czech, Italian, Korean, Dutch, and Chinese. Users may finetune Granite 3.0 models for languages beyond these 12 languages.
|
220 |
+
|
221 |
**Intended use:**
|
222 |
Prominent use cases of LLMs in text-to-text generation include summarization, text classification, extraction, question-answering, and more. All Granite Base models are able to handle these tasks as they were trained on a large amount of data from various domains. Moreover, they can serve as baseline to create specialized models for specific application scenarios.
|
223 |
|