Daemontatox commited on
Commit
48d8144
·
verified ·
1 Parent(s): 4b784cb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -6
README.md CHANGED
@@ -9,14 +9,28 @@ tags:
9
  license: apache-2.0
10
  language:
11
  - en
 
 
12
  ---
13
 
14
- # Uploaded model
15
 
16
- - **Developed by:** Daemontatox
17
- - **License:** apache-2.0
18
- - **Finetuned from model :** tiiuae/Falcon3-10B-Instruct
19
 
20
- This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  license: apache-2.0
10
  language:
11
  - en
12
+ pipeline_tag: text-generation
13
+ library_name: transformers
14
  ---
15
 
16
+ # Uploaded Model
17
 
18
+ **Developed by:** Daemontatox
 
 
19
 
20
+ **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)
21
 
22
+ **Finetuned from model:** [tiiuae/Falcon3-10B-Instruct](https://huggingface.co/tiiuae/Falcon3-10B-Instruct)
23
+
24
+ This model was fine-tuned from the Falcon-10B-Instruct model. It was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Hugging Face's TRL library.
25
+
26
+ This model is intended for text generation tasks, with a focus on reasoning capabilities and instruction following, similar to capabilities demonstrated by the ChatGPT-O1-Mini model.
27
+
28
+ ## Training Details
29
+
30
+ This model was fine-tuned with Unsloth and TRL, resulting in significant speed improvements during the training process. Details on specific fine-tuning data, parameters and methods will be added soon. The fine-tuning process has prioritized improving the model's reasoning abilities on various benchmarks.
31
+
32
+ ## Intended Use
33
+
34
+ This model is intended for research and development purposes related to text generation, instruction following, and complex reasoning tasks. It is suitable for applications that require a model capable of handling multi-step logical problems and understanding nuanced instructions.
35
+
36
+ **Focus on Reasoning:** The fine-tuning has been geared towards enhancing the model's ability to tackle reasoning challenges and logic-based tasks.