aashish1904's picture
Upload README.md with huggingface_hub
8da45b6 verified
---
license: creativeml-openrail-m
datasets:
- AI-MO/NuminaMath-CoT
language:
- en
base_model:
- Qwen/Qwen2.5-7B-Instruct
pipeline_tag: text-generation
library_name: transformers
tags:
- Qwen2.5
- Ollama
- Neumind
- Math
- Instruct
- safetensors
- pytorch
- trl
---
[![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
# QuantFactory/Neumind-Math-7B-Instruct-GGUF
This is quantized version of [prithivMLmods/Neumind-Math-7B-Instruct](https://huggingface.co/prithivMLmods/Neumind-Math-7B-Instruct) created using llama.cpp
# Original Model Card
### Neumind-Math-7B-Instruct Model Files
The **Neumind-Math-7B-Instruct** is a fine-tuned model based on **Qwen2.5-7B-Instruct**, optimized for mathematical reasoning, step-by-step problem-solving, and instruction-based tasks in the mathematics domain. The model is designed for applications requiring structured reasoning, numerical computations, and mathematical proof generation.
| File Name | Size | Description | Upload Status |
|------------------------------------|------------|------------------------------------------|----------------|
| `.gitattributes` | 1.57 kB | Git attributes configuration file | Uploaded |
| `README.md` | 265 Bytes | ReadMe file with basic information | Updated |
| `added_tokens.json` | 657 Bytes | Additional token definitions | Uploaded |
| `config.json` | 860 Bytes | Model configuration settings | Uploaded |
| `generation_config.json` | 281 Bytes | Generation settings | Uploaded |
| `merges.txt` | 1.82 MB | Tokenizer merge rules | Uploaded |
| `pytorch_model-00001-of-00004.bin` | 4.88 GB | Model shard 1 of 4 | Uploaded (LFS) |
| `pytorch_model-00002-of-00004.bin` | 4.93 GB | Model shard 2 of 4 | Uploaded (LFS) |
| `pytorch_model-00003-of-00004.bin` | 4.33 GB | Model shard 3 of 4 | Uploaded (LFS) |
| `pytorch_model-00004-of-00004.bin` | 1.09 GB | Model shard 4 of 4 | Uploaded (LFS) |
| `pytorch_model.bin.index.json` | 28.1 kB | Model index JSON | Uploaded |
| `special_tokens_map.json` | 644 Bytes | Mapping of special tokens | Uploaded |
| `tokenizer.json` | 11.4 MB | Tokenizer configuration | Uploaded (LFS) |
| `tokenizer_config.json` | 7.73 kB | Additional tokenizer settings | Uploaded |
| `vocab.json` | 2.78 MB | Vocabulary for tokenization | Uploaded |
---
### **Key Features:**
1. **Mathematical Reasoning:**
Specifically fine-tuned for solving mathematical problems, including arithmetic, algebra, calculus, and geometry.
2. **Step-by-Step Problem Solving:**
Provides detailed, logical solutions for complex mathematical tasks and demonstrates problem-solving methodologies.
3. **Instructional Applications:**
Tailored for use in educational settings, such as tutoring systems, math content creation, and interactive learning tools.
---
### **Training Details:**
- **Base Model:** [Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B)
- **Dataset:** Trained on **AI-MO/NuminaMath-CoT**, a large dataset of mathematical problems and chain-of-thought (CoT) reasoning. The dataset contains **860k problems** across various difficulty levels, enabling the model to tackle a wide spectrum of mathematical tasks.
---
### **Capabilities:**
- **Complex Problem Solving:**
Solves a wide range of mathematical problems, from basic arithmetic to advanced calculus and algebraic equations.
- **Chain-of-Thought Reasoning:**
Excels in step-by-step logical reasoning, making it suitable for tasks requiring detailed explanations.
- **Instruction-Based Generation:**
Ideal for generating educational content, such as worked examples, quizzes, and tutorials.
---
### **Usage Instructions:**
1. **Model Setup:**
Download all model shards and the associated configuration files. Ensure the files are correctly placed for seamless loading.
2. **Inference:**
Load the model using frameworks like PyTorch and Hugging Face Transformers. Ensure the `pytorch_model.bin.index.json` file is in the same directory for shard-based loading.
3. **Customization:**
Adjust generation parameters using `generation_config.json` to optimize outputs for your specific application.
---
### **Applications:**
- **Education:**
Interactive math tutoring, content creation, and step-by-step problem-solving tools.
- **Research:**
Automated theorem proving and symbolic mathematics.
- **General Use:**
Solving everyday mathematical queries and generating numerical datasets.
---