Look, Listen, and Learn (L3) [4] Audio subnetwork produces generic audio representations that can be used for myriad downstream tasks. However, L3-Net Audio requires 18 MB and 12 MB of static and dynamic memory respectively, making it infeasible for small edge devices with a single microcontroller. EdgeL3 [2] is competetive with L3 Audio while being 95.45% sparse. However, it still has a high activation memory requirement.
To jointly handle both static and dynamic memory, we introduce Specialized Embedding Approximation[1], a teacher-student learning paradigm where the student audio embedding model is trained to approximate only the part of the teacher's embedding manifold which is relevant to the target data-domain. Notice the difference between data-domain and dataset. Restricting the specialization on a particular downstream dataset would compromise intra-domain generalizability.
edgel3
is an open-source Python library for downloading the smaller versions of L3 models and computing deep audio embeddings from such models.
- The
sea
models are specialized for SONYC-UST [5] data domain. Training pipelines can be found [here]. - The
sparse
models provided have been re-trained using two different mechanisms: fine-tuningft
and knowledge distillationkd
. Training pipelines can be found [here].
For non-compressed L3-Net, please refer to OpenL3 [3]
edgel3
has been tested with Tensorflow 2.0 and Keras 2.3.1.
pip install tensorflow==2.0.0
edgel3 depends on the pysoundfile
module to load audio files, which depends on the non-Python library libsndfile
. On Windows and macOS, these will be installed via pip
and you can therefore skip this step.
However, on Linux this must be installed manually via your platform's package manager.
For Debian-based distributions (such as Ubuntu), this can be done by simply running
apt-get install libsndfile1
For more detailed information, please consult the pysoundfile
installation documentation.
The simplest way to install edgel3 is by using pip
, which will also install the additional required dependencies
if needed. To install edgel3 using pip
, simply run
pip install edgel3
NOTE: Use v0.2.1
for both sparse
and sea
models. v0.1.0
has only sparse
models.
To install the latest version of edgel3 from source:
-
Clone or pull the lastest version:
git clone https://github.com/ksangeeta2429/edgel3.git
-
Install using pip to handle python dependencies:
cd edgel3 pip install -e .
Load a SONYC-UST specialized L3 audio (reduced input represenation and reduced architecture) that outputs an embedding of length 128
model = edgel3.models.load_embedding_model(model_type='sea', emb_dim=128)
Load a 95.45% sparse L3 audio re-trained with fine-tuning
model = edgel3.models.load_embedding_model(model_type='sparse', retrain_type='ft', sparsity=95.45)
Load a 87.0% sparse L3 audio re-trained with knowledge distillation
model = edgel3.models.load_embedding_model(model_type='sparse', retrain_type='kd', sparsity=87.0)
For more examples, please see the tutorial and module usage.
If you use the SEA/EdgeL3 Github repos or the pre-trained models, please cite the relevant work:
[1] Specialized Embedding Approximation for Edge Intelligence: A case study in Urban Sound Classification
Sangeeta Srivastava, Dhrubojyoti Roy, Mark Cartwright, Juan Pablo Bello, and Anish Arora.
To be published in IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP), Toronto, Canada, June 2021.
[2] EdgeL3: Compressing L3-Net for Mote-Scale Urban Noise Monitoring
Sangeeta Kumari, Dhrubojyoti Roy, Mark Cartwright, Juan Pablo Bello, and Anish Arora.
Parallel AI and Systems for the Edge (PAISE), Rio de Janeiro, Brazil, May 2019.
[3] Look, Listen and Learn More: Design Choices for Deep Audio Embeddings
Jason Cramer, Ho-Hsiang Wu, Justin Salamon, and Juan Pablo Bello.
IEEE Int. Conf. on Acoustics, Speech and Signal Processing (ICASSP), pages 3852–3856, Brighton, UK, May 2019.
[4] Look, Listen and Learn
Relja Arandjelović and Andrew Zisserman
IEEE International Conference on Computer Vision (ICCV), Venice, Italy, Oct. 2017.
[5] SONYC Urban Sound Tagging (SONYC-UST): a multilabel dataset from an urban acoustic sensor network
Mark Cartwright, Ana Elisa Mendez Mendez, Graham Dove, Jason Cramer et al. 2019.