Skip to content

Include some core functions and model to handle speech separation

License

Notifications You must be signed in to change notification settings

jasdasdf/speech_separation

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

29 Commits
 
 
 
 
 
 
 
 

Repository files navigation

speech_separation

This is a repository for speech separation tasks.

This project is highly inspired by the paper[1], and is still working to improve the performance.

Data

AVspeech dataset : contains 4700 hours of video segments, from a total of 290k YouTube videos.

Preprocessing

There are several preprocess functions in the lib. Including STFT, iSTFT, power-law compression etc.

The visual frames are transfered to 512 face embeddings with facenet pre-trained model[2].

Model

Audio part : Dilated CNN + Bidirectional LSTM.

Video part : Still working.

Loss function : modified discriminative loss function inspired from paper[3].

Prediction

Apply complex ratio mask (cRM) to enhance phase spectrum. Maintain the quality during transformation by hyperbolic tangent fucntion.[4]

The model will be evaluated by signal-to-distortion ratio.

Reference

[1] Lookng to Listen at the Cocktail Party:A Speaker-Independent Audio-Visual Model for Speech Separation, A. Ephrat et al., arXiv:1804.03619v2 [cs.SD] 9 Aug 2018

[2] FaceNet Pretrained model

[3] Joint Optimization of Masks and Deep Recurrent Neural Networks for Monaural Source Separation, P. Hunag et al,arXiv:1502.04149v4 [cs.SD] 1 Oct 2015

[4] Complex Ratio Masking for Monaural Speech Separation

About

Include some core functions and model to handle speech separation

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%