Abstract
Distributed k-winners-takes-all (k-WTA) neural network (k-WTANN) models have better scalability than centralized ones. In this work, a distributed k-WTANN model with a simple structure is designed for the efficient selection of k winners among a group of more than k agents via competition based on their inputs. Unlike an existing distributed k-WTANN model, the proposed model does not rely on consensus filters, and only has one state variable. We prove that under mild conditions, the proposed distributed k-WTANN model has global asymptotic convergence. The theoretical conclusions are validated via numerical examples, which also show that our model is of better convergence speed than the existing distributed k-WTANN model.
Original language | English |
---|---|
Article number | 119528 |
Journal | Information Sciences |
Volume | 647 |
DOIs | |
Publication status | Published - Nov 2023 |
MoE publication type | A1 Journal article-refereed |
Funding
This work is supported in part by the National Natural Science Foundation of China under Grant 62206109 , the Guangdong Key Laboratory of Data Security and Privacy Preserving under Grant 2017B030301004 , the Guangdong Basic and Applied Basic Research Foundation under Grant 2022A1515010976 , the Young Scholar Program of Pazhou Lab under Grant PZL2021KF0022 , the Science and Technology Program of Guangzhou under Grant 202201010457 . G. Geng is supported by the Pearl River Talents Plan . We would like to thank the reviewers for their valuable comments.
Keywords
- Distributed competition
- k-winners-take-all
- Recurrent neural network