Interpretable Convolutional Filters with SincNet

11/23/2018
by   Mirco Ravanelli, et al.
0

Deep learning is currently playing a crucial role toward higher levels of artificial intelligence. This paradigm allows neural networks to learn complex and abstract representations, that are progressively obtained by combining simpler ones. Nevertheless, the internal "black-box" representations automatically discovered by current neural architectures often suffer from a lack of interpretability, making of primary interest the study of explainable machine learning techniques. This paper summarizes our recent efforts to develop a more interpretable neural model for directly processing speech from the raw waveform. In particular, we propose SincNet, a novel Convolutional Neural Network (CNN) that encourages the first layer to discover more meaningful filters by exploiting parametrized sinc functions. In contrast to standard CNNs, which learn all the elements of each filter, only low and high cutoff frequencies of band-pass filters are directly learned from data. This inductive bias offers a very compact way to derive a customized filter-bank front-end, that only depends on some parameters with a clear physical meaning. Our experiments, conducted on both speaker and speech recognition, show that the proposed architecture converges faster, performs better, and is more interpretable than standard CNNs.

READ FULL TEXT
research
12/13/2018

Speech and Speaker Recognition from Raw Waveform with SincNet

Deep neural networks can learn complex and abstract representations, tha...
research
03/12/2021

Learning spectro-temporal representations of complex sounds with parameterized neural networks

Deep Learning models have become potential candidates for auditory neuro...
research
11/12/2019

WaveletKernelNet: An Interpretable Deep Neural Network for Industrial Intelligent Diagnosis

Convolutional neural network (CNN), with ability of feature learning and...
research
07/29/2018

Speaker Recognition from raw waveform with SincNet

Deep learning is progressively gaining popularity as a viable alternativ...
research
07/16/2020

Training Interpretable Convolutional Neural Networks by Differentiating Class-specific Filters

Convolutional neural networks (CNNs) have been successfully used in a ra...
research
08/22/2022

Low-Level Physiological Implications of End-to-End Learning of Speech Recognition

Current speech recognition architectures perform very well from the poin...
research
03/15/2021

Lightweight and interpretable neural modeling of an audio distortion effect using hyperconditioned differentiable biquads

In this work, we propose using differentiable cascaded biquads to model ...

Please sign up or login with your details

Forgot password? Click here to reset