--- license: mit task_categories: - automatic-speech-recognition tags: - collator --- ## Dynamic Audio Data Augmentation ## Key Benefits Enhanced Robustness: By varying spectrogram parameters and injecting realistic noise, our models learn to handle a wide range of audio conditions. Low Overhead: The augmentation is integrated into the existing pipeline, ensuring minimal additional computational cost. Data collator (low overhead) versus Dataset (higher overhead) ### On-the-Fly Spectrogram Parameter Adjustment: n_fft and hop_length: Values for n_fft and hop_length are randomly selected from predefined ranges for each audio sample, providing varied spectrogram representations. ### Log-Mel Modulation: Augmentation process integrates with the existing log-Mel spectrogram calculation. This means we modulate the parameters of the log-Mel spectrogram dynamically, ensuring no additional overhead is introduced while providing effective data augmentation. ### Efficiency and Performance Log-Mel Spectrogram Manipulation: Augmentation process seamlessly integrates into the existing log-Mel spectrogram calculation, adding no extra overhead. This efficient design ensures that our preprocessing remains computationally lightweight and fast. #### Adaptive Context-Aware Noise Injection Preprocessing pipeline that includes adaptive context-aware noise injection to enhance model robustness. This method dynamically adjusts noise intensity based on the amplitude of the audio signal, ensuring realistic and effective augmentation. - **Types of Noise**: White, pink, and environmental noise. - **Dynamic Adjustment**: Noise intensity is scaled based on the amplitude of the audio signal. - **Integration**: The noise injection process is seamlessly integrated into our existing log-Mel spectrogram calculation pipeline, adding minimal overhead. ##### Key Benefits - **Improved Generalization**: Models become more resilient to noise and diverse audio conditions. - **Low Overhead**: The augmentation process leverages the existing pipeline, ensuring efficient computation without significant additional cost. ##### Example Usage ```python ## HF transformers or pure pytorch data_collator = DataCollatorSpeechSeq2SeqWithPadding( processor=processor, decoder_start_token_id=model.config.decoder_start_token_id, apply_augmentation=True, apply_noise_injection=True # Enable adaptive noise injection ) dataloader = torch.utils.data.DataLoader(dataset, batch_size=2, shuffle=True, collate_fn=data_collator) for batch in dataloader: outputs = model(batch)