Parametric Spectral Filters For Fast Converging, Scalable Convolutional Neural Networks
Luke Wood, Eric Larson
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:05:00
Using spectral multiplication to compute convolution in neural networks has been investigated by a number of researchers because of its potential in speeding up computations for large images. However, previous methods require the learning of arbitrarily large convolution filters in the spectral domain, causing two untenable problems: an explosion in the number of trainable parameters per filter and an inability to reuse filters across images of differing sizes. To address this, we propose the usage of spectral approximation functions to approximate the massive Spectral domain filters with only a few trainable parameters. Our empirical analysis suggests that the proposed approximation maintains the benefits of arbitrarily large filters (such as improved rate of convergence in training, accuracy, and stability) while relying on significantly fewer trainable parameters.
Chairs:
Jinyu Li