|
|
|
--- |
|
|
|
license: creativeml-openrail-m |
|
datasets: |
|
- AI-MO/NuminaMath-CoT |
|
language: |
|
- en |
|
base_model: |
|
- Qwen/Qwen2.5-7B-Instruct |
|
pipeline_tag: text-generation |
|
library_name: transformers |
|
tags: |
|
- Qwen2.5 |
|
- Ollama |
|
- Neumind |
|
- Math |
|
- Instruct |
|
- safetensors |
|
- pytorch |
|
- trl |
|
|
|
--- |
|
|
|
[![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory) |
|
|
|
|
|
# QuantFactory/Neumind-Math-7B-Instruct-GGUF |
|
This is quantized version of [prithivMLmods/Neumind-Math-7B-Instruct](https://huggingface.co/prithivMLmods/Neumind-Math-7B-Instruct) created using llama.cpp |
|
|
|
# Original Model Card |
|
|
|
|
|
### Neumind-Math-7B-Instruct Model Files |
|
|
|
The **Neumind-Math-7B-Instruct** is a fine-tuned model based on **Qwen2.5-7B-Instruct**, optimized for mathematical reasoning, step-by-step problem-solving, and instruction-based tasks in the mathematics domain. The model is designed for applications requiring structured reasoning, numerical computations, and mathematical proof generation. |
|
|
|
| File Name | Size | Description | Upload Status | |
|
|------------------------------------|------------|------------------------------------------|----------------| |
|
| `.gitattributes` | 1.57 kB | Git attributes configuration file | Uploaded | |
|
| `README.md` | 265 Bytes | ReadMe file with basic information | Updated | |
|
| `added_tokens.json` | 657 Bytes | Additional token definitions | Uploaded | |
|
| `config.json` | 860 Bytes | Model configuration settings | Uploaded | |
|
| `generation_config.json` | 281 Bytes | Generation settings | Uploaded | |
|
| `merges.txt` | 1.82 MB | Tokenizer merge rules | Uploaded | |
|
| `pytorch_model-00001-of-00004.bin` | 4.88 GB | Model shard 1 of 4 | Uploaded (LFS) | |
|
| `pytorch_model-00002-of-00004.bin` | 4.93 GB | Model shard 2 of 4 | Uploaded (LFS) | |
|
| `pytorch_model-00003-of-00004.bin` | 4.33 GB | Model shard 3 of 4 | Uploaded (LFS) | |
|
| `pytorch_model-00004-of-00004.bin` | 1.09 GB | Model shard 4 of 4 | Uploaded (LFS) | |
|
| `pytorch_model.bin.index.json` | 28.1 kB | Model index JSON | Uploaded | |
|
| `special_tokens_map.json` | 644 Bytes | Mapping of special tokens | Uploaded | |
|
| `tokenizer.json` | 11.4 MB | Tokenizer configuration | Uploaded (LFS) | |
|
| `tokenizer_config.json` | 7.73 kB | Additional tokenizer settings | Uploaded | |
|
| `vocab.json` | 2.78 MB | Vocabulary for tokenization | Uploaded | |
|
|
|
--- |
|
|
|
### **Key Features:** |
|
|
|
1. **Mathematical Reasoning:** |
|
Specifically fine-tuned for solving mathematical problems, including arithmetic, algebra, calculus, and geometry. |
|
|
|
2. **Step-by-Step Problem Solving:** |
|
Provides detailed, logical solutions for complex mathematical tasks and demonstrates problem-solving methodologies. |
|
|
|
3. **Instructional Applications:** |
|
Tailored for use in educational settings, such as tutoring systems, math content creation, and interactive learning tools. |
|
|
|
--- |
|
|
|
### **Training Details:** |
|
- **Base Model:** [Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) |
|
- **Dataset:** Trained on **AI-MO/NuminaMath-CoT**, a large dataset of mathematical problems and chain-of-thought (CoT) reasoning. The dataset contains **860k problems** across various difficulty levels, enabling the model to tackle a wide spectrum of mathematical tasks. |
|
|
|
--- |
|
|
|
### **Capabilities:** |
|
|
|
- **Complex Problem Solving:** |
|
Solves a wide range of mathematical problems, from basic arithmetic to advanced calculus and algebraic equations. |
|
|
|
- **Chain-of-Thought Reasoning:** |
|
Excels in step-by-step logical reasoning, making it suitable for tasks requiring detailed explanations. |
|
|
|
- **Instruction-Based Generation:** |
|
Ideal for generating educational content, such as worked examples, quizzes, and tutorials. |
|
|
|
--- |
|
|
|
### **Usage Instructions:** |
|
|
|
1. **Model Setup:** |
|
Download all model shards and the associated configuration files. Ensure the files are correctly placed for seamless loading. |
|
|
|
2. **Inference:** |
|
Load the model using frameworks like PyTorch and Hugging Face Transformers. Ensure the `pytorch_model.bin.index.json` file is in the same directory for shard-based loading. |
|
|
|
3. **Customization:** |
|
Adjust generation parameters using `generation_config.json` to optimize outputs for your specific application. |
|
--- |
|
|
|
### **Applications:** |
|
|
|
- **Education:** |
|
Interactive math tutoring, content creation, and step-by-step problem-solving tools. |
|
- **Research:** |
|
Automated theorem proving and symbolic mathematics. |
|
- **General Use:** |
|
Solving everyday mathematical queries and generating numerical datasets. |
|
--- |
|
|