Weight sharing for LMS algorithms: Convolutional neural networks inspired multichannel adaptive filtering

Weight sharing has been a key to the success of convolutional neural networks, as it forces a neural network to detect common ‘local’ features across an image by applying the same weights across all input samples (pixels). This has been shown to resolve both the computational and performance issues...

Full description

Saved in:
Bibliographic Details
Published inDigital signal processing Vol. 127; p. 103580
Main Authors Cheong Took, Clive, Mandic, Danilo
Format Journal Article
LanguageEnglish
Published Elsevier Inc 01.07.2022
Subjects
Online AccessGet full text
ISSN1051-2004
1095-4333
1095-4333
DOI10.1016/j.dsp.2022.103580

Cover

More Information
Summary:Weight sharing has been a key to the success of convolutional neural networks, as it forces a neural network to detect common ‘local’ features across an image by applying the same weights across all input samples (pixels). This has been shown to resolve both the computational and performance issues as less data are required for training and there is a lower risk of overfitting the data. While such an approach has been instrumental in machine learning problems, it has not yet been adopted in large-scale signal processing paradigms. To this end, we study weight sharing for LMS type algorithms in a multi-channel setting and analyse its effect on the existence, uniqueness, and convergence of the solution. As a result, the proposed Weight Sharing Multichannel Least Mean Squares (WS-MLMS) algorithm minimises the sum of error squares across “channels” rather than the traditional across “time” minimisation. Rigorous analysis of our proposed WS-MLMS algorithm demonstrates that weight sharing leads to better convergence properties and enhanced capability to cope with a large number of channels in terms of both computational complexity and stability. Simulation studies on weight sharing, in scenarios as massive as 256×256 MIMO systems, support the approach. •Weight sharing of adaptive coefficients.•Invariance against large eigenvalue spread of input correlation.•Enhanced stability for processing a large number of channels.
ISSN:1051-2004
1095-4333
1095-4333
DOI:10.1016/j.dsp.2022.103580