File size: 5,239 Bytes
8da45b6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111

---

license: creativeml-openrail-m
datasets:
- AI-MO/NuminaMath-CoT
language:
- en
base_model:
- Qwen/Qwen2.5-7B-Instruct
pipeline_tag: text-generation
library_name: transformers
tags:
- Qwen2.5
- Ollama
- Neumind
- Math
- Instruct
- safetensors
- pytorch
- trl

---

[![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)


# QuantFactory/Neumind-Math-7B-Instruct-GGUF
This is quantized version of [prithivMLmods/Neumind-Math-7B-Instruct](https://huggingface.co/prithivMLmods/Neumind-Math-7B-Instruct) created using llama.cpp

# Original Model Card


### Neumind-Math-7B-Instruct Model Files 

The **Neumind-Math-7B-Instruct** is a fine-tuned model based on **Qwen2.5-7B-Instruct**, optimized for mathematical reasoning, step-by-step problem-solving, and instruction-based tasks in the mathematics domain. The model is designed for applications requiring structured reasoning, numerical computations, and mathematical proof generation.

| File Name                          | Size       | Description                              | Upload Status  |
|------------------------------------|------------|------------------------------------------|----------------|
| `.gitattributes`                   | 1.57 kB    | Git attributes configuration file        | Uploaded       |
| `README.md`                        | 265 Bytes  | ReadMe file with basic information       | Updated        |
| `added_tokens.json`                | 657 Bytes  | Additional token definitions             | Uploaded       |
| `config.json`                      | 860 Bytes  | Model configuration settings             | Uploaded       |
| `generation_config.json`           | 281 Bytes  | Generation settings                      | Uploaded       |
| `merges.txt`                       | 1.82 MB    | Tokenizer merge rules                    | Uploaded       |
| `pytorch_model-00001-of-00004.bin` | 4.88 GB    | Model shard 1 of 4                       | Uploaded (LFS) |
| `pytorch_model-00002-of-00004.bin` | 4.93 GB    | Model shard 2 of 4                       | Uploaded (LFS) |
| `pytorch_model-00003-of-00004.bin` | 4.33 GB    | Model shard 3 of 4                       | Uploaded (LFS) |
| `pytorch_model-00004-of-00004.bin` | 1.09 GB    | Model shard 4 of 4                       | Uploaded (LFS) |
| `pytorch_model.bin.index.json`     | 28.1 kB    | Model index JSON                         | Uploaded       |
| `special_tokens_map.json`          | 644 Bytes  | Mapping of special tokens                | Uploaded       |
| `tokenizer.json`                   | 11.4 MB    | Tokenizer configuration                  | Uploaded (LFS) |
| `tokenizer_config.json`            | 7.73 kB    | Additional tokenizer settings            | Uploaded       |
| `vocab.json`                       | 2.78 MB    | Vocabulary for tokenization              | Uploaded       |

---

### **Key Features:**

1. **Mathematical Reasoning:**  
   Specifically fine-tuned for solving mathematical problems, including arithmetic, algebra, calculus, and geometry.

2. **Step-by-Step Problem Solving:**  
   Provides detailed, logical solutions for complex mathematical tasks and demonstrates problem-solving methodologies.

3. **Instructional Applications:**  
   Tailored for use in educational settings, such as tutoring systems, math content creation, and interactive learning tools.

---

### **Training Details:**
- **Base Model:** [Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B)  
- **Dataset:** Trained on **AI-MO/NuminaMath-CoT**, a large dataset of mathematical problems and chain-of-thought (CoT) reasoning. The dataset contains **860k problems** across various difficulty levels, enabling the model to tackle a wide spectrum of mathematical tasks.

---

### **Capabilities:**

- **Complex Problem Solving:**  
   Solves a wide range of mathematical problems, from basic arithmetic to advanced calculus and algebraic equations.

- **Chain-of-Thought Reasoning:**  
   Excels in step-by-step logical reasoning, making it suitable for tasks requiring detailed explanations.

- **Instruction-Based Generation:**  
   Ideal for generating educational content, such as worked examples, quizzes, and tutorials.

---

### **Usage Instructions:**

1. **Model Setup:**  
   Download all model shards and the associated configuration files. Ensure the files are correctly placed for seamless loading.

2. **Inference:**  
   Load the model using frameworks like PyTorch and Hugging Face Transformers. Ensure the `pytorch_model.bin.index.json` file is in the same directory for shard-based loading.

3. **Customization:**  
   Adjust generation parameters using `generation_config.json` to optimize outputs for your specific application.
---

### **Applications:**

- **Education:**  
   Interactive math tutoring, content creation, and step-by-step problem-solving tools.
- **Research:**  
   Automated theorem proving and symbolic mathematics.
- **General Use:**  
   Solving everyday mathematical queries and generating numerical datasets.
---