|
--- |
|
license: creativeml-openrail-m |
|
datasets: |
|
- GAIR/o1-journey |
|
language: |
|
- en |
|
base_model: |
|
- Qwen/Qwen2.5-0.5B-Instruct |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
tags: |
|
- Qwen2.5 |
|
- Llama-Cpp |
|
- CoT |
|
- o1-journey |
|
- text-generation-inference |
|
- safetensors |
|
- Ollama |
|
--- |
|
### Acrux-500M-o1-Journey Model Files |
|
|
|
The **Acrux-500M-o1-Journey** is a lightweight, instruction-tuned language model fine-tuned from the **Qwen2.5-0.5B-Instruct** base model. With a size of 500 million parameters, it is designed for **cost-effective deployment** and **fast text generation** while maintaining quality performance for instruction-following tasks. |
|
|
|
| **File Name** | **Size** | **Description** | **Upload Status** | |
|
|----------------------------|----------------|-------------------------------------------|--------------------| |
|
| `.gitattributes` | 1.57 kB | Git attributes for managing LFS files. | Uploaded | |
|
| `README.md` | 195 Bytes | Model overview or documentation. | Updated | |
|
| `added_tokens.json` | 657 Bytes | Custom tokens for the tokenizer. | Uploaded | |
|
| `config.json` | 859 Bytes | Model configuration file. | Uploaded | |
|
| `generation_config.json` | 280 Bytes | Configuration for text generation. | Uploaded | |
|
| `merges.txt` | 1.82 MB | Merge rules for byte-pair encoding (BPE). | Uploaded | |
|
| `pytorch_model.bin` | 988 MB | Model weights (PyTorch format). | Uploaded (LFS) | |
|
| `special_tokens_map.json` | 644 Bytes | Mapping for special tokens. | Uploaded | |
|
| `tokenizer.json` | 11.4 MB | Full tokenizer configuration. | Uploaded (LFS) | |
|
| `tokenizer_config.json` | 7.73 kB | Additional tokenizer settings. | Uploaded | |
|
| `vocab.json` | 2.78 MB | Vocabulary for the tokenizer. | Uploaded | |
|
### **Key Features:** |
|
|
|
1. **Compact Size with Efficient Performance:** |
|
The smaller parameter count (500M) ensures faster inference and reduced hardware requirements. |
|
|
|
2. **Instruction Optimization:** |
|
Fine-tuned to follow prompts effectively, making it suitable for interactive applications and prompt-based tasks. |
|
|
|
3. **Domain-Specific Training:** |
|
Trained on the **GAIR/o1-journey** dataset, providing tailored capabilities for specific use cases. |
|
|
|
--- |
|
|
|
### **Training Details:** |
|
- **Base Model:** [Qwen2.5-0.5B-Instruct](#) |
|
- **Dataset Used for Fine-Tuning:** [GAIR/o1-journey](#) |
|
- A compact dataset focusing on instruction-driven generation with 1.42k samples. |
|
|
|
--- |
|
### **Capabilities:** |
|
|
|
1. **Instruction Following:** |
|
- Generates accurate and coherent responses to user instructions. |
|
- Handles summarization, question-answering, and conversational tasks. |
|
|
|
2. **Fast Inference:** |
|
- Ideal for real-time applications due to reduced latency from its smaller size. |
|
|
|
3. **Interactive AI Development:** |
|
- Suitable for chatbots, virtual assistants, and instructional interfaces. |
|
|
|
--- |
|
### **Usage Instructions:** |
|
|
|
1. **Setup:** |
|
Download all model files, ensuring compatibility with the Hugging Face Transformers library. |
|
|
|
2. **Loading the Model:** |
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
model_name = "prithivMLmods/Acrux-500M-o1-Journey" |
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
model = AutoModelForCausalLM.from_pretrained(model_name) |
|
``` |
|
3. **Sample Generate Text:** |
|
```python |
|
input_text = "Explain the concept of machine learning in simple terms." |
|
inputs = tokenizer(input_text, return_tensors="pt") |
|
outputs = model.generate(**inputs, max_length=100, temperature=0.7) |
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
``` |
|
4. **Optimize Generation:** |
|
Adjust parameters in `generation_config.json` for better control of output, such as: |
|
- `temperature` for randomness. |
|
- `top_p` for sampling diversity. |
|
- `max_length` for output size. |
|
--- |