--- library_name: transformers tags: [] --- # Model Card for Model ID ## Model Details ### Model Description This is a quantized model of the original version mohammed/whisper-small-arabic-cv-11 - **Developed by:** Mohammed Bakheet - **Funded by [optional]:** Kalam Technology - **Language(s) (NLP):** Arabic, English ## Uses This a quantized model that reads arabic voice and transcribes/translate it into english ### Direct Use First, install the following packages using the following commands: pip install -U optimum[exporters,onnxruntime] transformers pip install huggingface_hub ```python # uncomment the following installation if you are using a notebook: #!pip install -U optimum[exporters,onnxruntime] transformers #!pip install huggingface_hub # import the required packages from optimum.onnxruntime import ORTModelForSpeechSeq2Seq from transformers import WhisperTokenizerFast, WhisperFeatureExtractor, pipeline # set model name/id model_name = 'mohammed/quantized-whisper-small' # folder name model = ORTModelForSpeechSeq2Seq.from_pretrained(model_name, export=False) tokenizer = WhisperTokenizerFast.from_pretrained(model_name) feature_extractor = WhisperFeatureExtractor.from_pretrained(model_name) forced_decoder_ids = tokenizer.get_decoder_prompt_ids(language="ar", task="transcribe") pipe = pipeline('automatic-speech-recognition', model=model, tokenizer=tokenizer, feature_extractor=feature_extractor, model_kwargs={"forced_decoder_ids": forced_decoder_ids}) # the file to be transcribed pipe('Recording.mp3') ``` ### Out-of-Scope Use The model does a direct translation of Arabic speech, and doesn't do a direct transcription, we are still working on that. ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ```python First, install the following packages using the following commands: pip install -U optimum[exporters,onnxruntime] transformers pip install huggingface_hub from optimum.onnxruntime import ORTModelForSpeechSeq2Seq from transformers import WhisperTokenizerFast, WhisperFeatureExtractor, pipeline model_name = 'mohammed/quantized-whisper-small' # folder name model = ORTModelForSpeechSeq2Seq.from_pretrained(model_name, export=False) tokenizer = WhisperTokenizerFast.from_pretrained(model_name) feature_extractor = WhisperFeatureExtractor.from_pretrained(model_name) forced_decoder_ids = tokenizer.get_decoder_prompt_ids(language="ar", task="transcribe") pipe = pipeline('automatic-speech-recognition', model=model, tokenizer=tokenizer, feature_extractor=feature_extractor, model_kwargs={"forced_decoder_ids": forced_decoder_ids}) # the file to be transcribed pipe('Recording.mp3') ``` ### Training Data Please refer to the original model at "mohammed/whisper-small-arabic-cv-11" ### Training Procedure Please refer to the original model at "mohammed/whisper-small-arabic-cv-11" #### Preprocessing [optional] Please refer to the original model at "mohammed/whisper-small-arabic-cv-11" #### Training Hyperparameters - **Training regime:** Please refer to the original model at "mohammed/whisper-small-arabic-cv-11"