Re-Implementation of Google Research's VGGish model used for extracting audio features using Pytorch with GPU support.
-
Updated
Sep 28, 2022 - Python
Re-Implementation of Google Research's VGGish model used for extracting audio features using Pytorch with GPU support.
Audio encoding authentication and validation library for verifying audio as being from a trusted source
SERVER: Multi-modal Speech Emotion Recognition using Transformer-based and Vision-based Embeddings
Extract audio embeddings from an audio file using Python
Audio Embeddings using VGGish
Generate realistic, synthetic call center conversations
Visualizations of music semantics calculus using Spotify and deep embeddings.
Audio Deep Learning Project in Java
Audio search using Azure Cognitive Search
Directly from voice, recognise speaker emotion, intensity, & sentiment in speaker utterances.
Generate audio embedding out of pruned L3
Pytorch port of Google Research's VGGish model used for extracting audio features.
Add a description, image, and links to the audio-embedding topic page so that developers can more easily learn about it.
To associate your repository with the audio-embedding topic, visit your repo's landing page and select "manage topics."