license: mit
language: fr
datasets:
- Cnam-LMSSC/vibravox
tags:
- audio
- audio-to-audio
- speech
Master Model Card: Vibravox Audio Bandwidth extension Models
Overview
This master model card serves as an entry point for exploring multiple audio bandwidth extension (BWE) models trained on different sensor data from the Vibravox dataset.
These models are designed to to enhance the audio quality of body-conducted captured speech, by denoising and regenerating mid and high frequencies from low frequency content only.
The models are trained on specific sensors to address various audio capture scenarios using body conducted sound and vibration sensors.
Disclaimer
Each of these models has been trained for specific non-conventional speech sensors and is intended to be used with in-domain data.
Please be advised that using these models outside their intended sensor data may result in suboptimal performance.
Usage
All models are trained using Configurable EBEN (see publication in IEEE TASLP - arXiv link) and adapted to different sensor inputs. They are intended to be used at a sample rate of 16kHz.
Training Procedure
Detailed instructions for reproducing the experiments are available on the jhauret/vibravox Github repository and in the VibraVox paper on arXiV.
Available Models
The following models are available, each trained on a different sensor on the speech_clean
subset of (https://huggingface.co/datasets/Cnam-LMSSC/vibravox):
Transducer | Huggingface model link | EBEN configuration |
---|---|---|
In-ear comply foam-embedded microphone | EBEN_soft_in_ear_microphone | M=4,P=2,Q=4 |
In-ear rigid earpiece-embedded microphone | EBEN_rigid_in_ear_microphone | M=4,P=2,Q=4 |
Forehead miniature vibration sensor | EBEN_forehead_accelerometer | M=4,P=4,Q=4 |
Temple vibration pickup | EBEN_temple_vibration_pickup | M=4,P=1,Q=4 |
Laryngophone | EBEN_throat_microphone | M=4,P=2,Q=4 |