Paper Implementation for :
[1] Deep Neural Network Baseline For Dcase Challenge 2016 [Paper]
This code runs on the DCASE 2016 Audio Dataset.
wav_dev_fd development audio folder
wav_eva_fd evaluation audio folder
dev_fd development features folder
eva_fd evaluation features folder
label_csv development meta file
txt_eva_path evaluation test file
new_p evaluation evaluate file
Go ahead and clone this repository using
$ git clone https://github.com/DeepLearn-lab/audio_CHIME.git
If you are looking for a quick running version go inside single_file folder and run
$ python mainfile.py
The process involves three steps:
- Feature Extraction
- Training on Development Dataset
- Testing on Evaluation Dataset
We are going to extract mel frequencies on raw audio waveforms. Go ahead and uncomment
feature_extraction function which would extract these features and save it in the .f pickle.
We train our model on these extracted featuers. We use a convolution neural network for training and testing purpose. Alteration in model can be done in model.py file.
All hyper-parameters can be set in util.py. Once you have made all the required changes or want to run on the pre-set ones, run
$ python mainfile.py
This will run the model which we test and use EER for rating our model.