riotu-lab commited on
Commit
97d86d8
1 Parent(s): 45d3872

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +74 -0
README.md CHANGED
@@ -1,3 +1,77 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - ar
5
+ tags:
6
+ - ArabianGPT
7
+ widget:
8
+ - text: "أعلنت وزارة الحج في المملكة العربية السعودية"
9
+ example_title: "مثال ١"
10
+ - text: "يبدو اليوم جميلا، سأقوم بتحضير"
11
+ example_title: "مثال ٢"
12
+ - text: "إن التقنيات الحديثة"
13
+ example_title: "مثال ٣"
14
  ---
15
+
16
+ # ArabianGPT Model Overview
17
+
18
+ ## Disclaimer for the Use of Large Language Models (LLMs) for Text Generation
19
+
20
+ <p style="color: red;">We disclaim all responsibility for any harm, inaccuracies, or inappropriate content generated by ArabianGPT-0.8B, and users engage with and apply the model's outputs at their own risk.</p>
21
+
22
+ > **Important Note:** Currently, we offer a raw pre-trained model. Our team is actively working on releasing instruction-based LLMs that are fine-tuned and augmented with LRHF. The first set of pre-trained models has been made available for community exploration. While we do have models fine-tuned for specific tasks such as summarization and sentiment analysis, they are still in the development phase.
23
+
24
+ ## How you can use this Pre-Trained?
25
+ You are invited to utilize this pre-trained, native Arabic language model as an experimental tool to assess its capabilities, aid in its fine-tuning, and evaluate its performance across a variety of downstream tasks. We encourage you to review our technical report for a comprehensive understanding of the model's performance metrics and the specific downstream tasks it has been tested on. This will provide valuable insights into its applicability and effectiveness in diverse applications.
26
+
27
+
28
+ ## Introduction
29
+ ArabianGPT-0.8B, part of the ArabianLLM initiatives, is a specialized GPT model optimized for the Arabic language. Developed at Prince Sultan University's Robotics and Internet of Things Lab, this model is a leap forward in natural language modeling and generation for Arabic, tackling the language's unique challenges.
30
+
31
+ ## Key Features
32
+ - **Architecture**: GPT-2
33
+ - **Model Size**: 0.8 billion parameters
34
+ - **Layers**: 36
35
+ - **Model Attention Layers (MAL)**: 20
36
+ - **Context Window Size**: 1024 tokens
37
+
38
+ ## Training
39
+ - **Dataset**: Scraped texts contains scientific articles, and general texts
40
+ - **Data Size**: 117 GB
41
+ - **Tokenizer**: Aranizer 64K
42
+ - **Tokens**: Over 14 billion
43
+ - **Hardware**: 5 NDIVIA A100 GPUs
44
+ - **Performance**: loss of 3.6
45
+
46
+
47
+ ## Role in ArabianLLM Initiatives
48
+ ArabianGPT-0.8B is crucial for advancing Arabic language processing, addressing challenges unique to Arabic morphology and dialects.
49
+
50
+ ## Usage
51
+ Suitable for Arabic text generation tasks. Example usage with Transformers Pipeline:
52
+ ```python
53
+ from transformers import pipeline
54
+
55
+ pipe = pipeline("text-generation", model="riotu-lab/ArabianGPT-08B", max_new_tokens=1024)
56
+ text = ''
57
+ pipe(text)
58
+ ```
59
+
60
+ ## Limitations and Ethical Considerations
61
+
62
+ - The model may have context understanding or text generation limitations in certain scenarios.
63
+ - Emphasis on ethical use to prevent misinformation or harmful content propagation.
64
+
65
+ ## Acknowledgments
66
+
67
+ Special thanks to Prince Sultan University, particularly the Robotics and Internet of Things Lab.
68
+
69
+ ## Contact Information
70
+
71
+ For inquiries: [riotu@psu.edu.sa](mailto:riotu@psu.edu.sa).
72
+
73
+ ## Disclaimer for the Use of Large Language Models (LLMs) for Text Generation
74
+
75
+ <p style="color: red;">We disclaim all responsibility for any harm, inaccuracies, or inappropriate content generated by ArabianGPT-0.3B, and users engage with and apply the model's outputs at their own risk.</p>
76
+
77
+