Find in Library
Search millions of books, articles, and more
Indexed Open Access Databases
A Class of Diffusion Zero Attracting Stochastic Gradient Algorithms With Exponentiated Error Cost Functions
oleh: Zhengyan Luo, Haiquan Zhao, Xiangping Zeng
Format: | Article |
---|---|
Diterbitkan: | IEEE 2020-01-01 |
Deskripsi
In this paper, a class of diffusion zero-attracting stochastic gradient algorithms with exponentiated error cost functions is put forward due to its good performance for sparse system identification. Distributed estimation algorithms based on the popular mean-square error criterion have poor behavior for sparse system identification with color noise. To overcome this drawback, a class of stochastic gradient least exponentiated (LE) algorithms with exponentiated error cost functions were proposed, which achieved a low steady-state compared with the least mean square (LMS) algorithm. However, those LE algorithms may suffer from performance deterioration in the spare system. For sparse system identification in the adaptive network, a polynomial variable scaling factor improved diffusion least sum of exponentials (PZA-VSIDLSE) algorithm and an l<sub>p</sub>-norm constraint diffusion least exponentiated square (LP-DLE2) algorithm are proposed in this work. Instead of using the l<sub>1</sub>-norm penalty, an l<sub>p</sub>-norm penalty and a polynomial zero-attractor are employed as a substitution in the cost functions of the LE algorithms. Then, we perform mean behavior model and mean square behavior modal of the LP-DLE2 algorithm with several common assumptions. Moreover, simulations in the context of distributed network sparse system identification show that the proposed algorithms have a low steady-state compared with the existing algorithms.