--- base_model: unsloth/mistral-7b-instruct-v0.1-bnb-4bit tags: - text-generation-inference - transformers - unsloth - mistral - trl license: apache-2.0 language: - en datasets: - Laurent1/MedQuad-MedicalQnADataset_128tokens_max --- # Model Card for Mistral-7B-Instruct-v0.1-Unsloth-MedicalQA drawing This is a medical question-answering model fine-tuned for healthcare domain
Foundation Model: https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1
Dataset: https://huggingface.co/datasets/Laurent1/MedQuad-MedicalQnADataset_128tokens_max
The model has been fine-tuned using CUDA-enabled GPU hardware with optimized training through [Unsloth](https://github.com/unslothai/unsloth). [](https://github.com/unslothai/unsloth) ## Model Details The model is based upon the foundation model: Mistral-7B-Instruct-v0.1.
It has been tuned with Supervised Fine-tuning Trainer using the Unsloth optimization framework for faster and more efficient training. ### Libraries - unsloth - transformers - torch - trl - peft - einops - bitsandbytes - datasets ## Training Configuration ### Model Parameters - max_sequence_length = 2048 - load_in_4bit = True - LoRA rank (r) = 32 - lora_alpha = 16 - lora_dropout = 0 ### Target Modules for LoRA - q_proj - k_proj - v_proj - o_proj - gate_proj - up_proj - down_proj ### Training Hyperparameters - per_device_train_batch_size = 2 - gradient_accumulation_steps = 16 - warmup_steps = 5 - warmup_ratio = 0.03 - max_steps = 1600 - learning_rate = 1e-4 - weight_decay = 0.01 - lr_scheduler_type = "linear" - optimizer = "paged_adamw_32bit" ## Training Statistics ### Hardware Utilization - Training duration: 10,561.28 seconds (approximately 176.02 minutes) - Peak reserved memory: 5.416 GB - Peak reserved memory for training: 0.748 GB - Peak reserved memory % of max memory: 13.689% - Peak reserved memory for training % of max memory: 1.891% ### Dataset The model was trained on the MedQuad dataset, which contains medical questions and answers. The training data was processed using a chat template format for instruction-tuning. ## Bias, Risks, and Limitations Users (both direct and downstream) should be aware of the following: 1. This model is intended for medical question-answering but should not be used as a substitute for professional medical advice. 2. The model's responses should be verified by healthcare professionals before making any medical decisions. 3. Generation of plausible yet incorrect medical information remains a possibility. 4. The model's knowledge is limited to its training data and may not cover all medical conditions or recent medical developments. ## Usage The model can be loaded and used with the Unsloth library: ```python from unsloth import FastLanguageModel max_seq_length = 2048 # Choose any! We auto support RoPE Scaling internally! dtype = ( None # None for auto detection. Float16 for Tesla T4, V100, Bfloat16 for Ampere+ ) model, tokenizer = FastLanguageModel.from_pretrained( "bouthros/Mistral-7B-Instruct-v0.1-Unsloth-MedicalQA", max_seq_length=2048, load_in_4bit=True, ) ``` Example usage: ```python messages = [ {"from": "human", "value": "What are the types of liver cancer?"}, ] inputs = tokenizer.apply_chat_template( messages, tokenize=True, add_generation_prompt=True, return_tensors="pt" ).to("cuda") ``` ## Model Access The model is available on Hugging Face Hub at: bouthros/Mistral-7B-Instruct-v0.1-Unsloth-MedicalQA ## Citation If you use this model, please cite the original Mistral-7B-Instruct-v0.1 model and the MedQuad dataset.