Skip to content

Classification of marmoset calls using a hybrid convolutional-recurrent neural network.

License

Notifications You must be signed in to change notification settings

pedro-varela1/Marmoset-Call-Classification

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Marmoset Call Classification

Classification of marmoset calls using tweetynet. Check out more at poster.

Development of the MarmosetCallClassification work.

The pipeline converts audio to spectrograms, generates predictions with the trained neural network and displays spectrograms with the classifications (optional).

Performance

Performance of the TweetyNet model trained on marmoset vocalizations*:

Accuracy Loss Segment Error Rate
0.9584 0.1472 0.2821

Training metrics for validation data

Accuracy of the model as a function of the time of trained audios. Accuracy

Average error of the model as a function of the time of trained audios. Loss

Normalized confusion matrix of the model on validation data calls. Confusion Matrix

Installation

Install vak version >= 0.8.

Use

python3 predict.py <data_dir> <out_dir> --plot_spec

  • You don't need to modify predict.toml!

data_dir: Directory containing the data to predict on (wav audio files).

out_dir: Directory to save the outputs (preprocessed spectrograms and predictions). Two folders will be created: results to store the predictions (predictions.csv and spect_plots for the spectrograms with the classifications if --plot_spec) and data to store the preprocessed data that will serve as input for tweetynet.

--plot_spec: Add to plot the spectrograms.

About

Classification of marmoset calls using a hybrid convolutional-recurrent neural network.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages