Abstract
In order to realize the full potential of wireless edge artificial intelligence (AI), very large and diverse datasets will often be required for energy-demanding model training on resource-constrained edge devices. This paper proposes a lead federated neuromorphic learning (LFNL) technique, which is a decentralized energy-efficient brain-inspired computing method based on spiking neural networks. The proposed technique will enable edge devices to exploit brain-like biophysiological structure to collaboratively train a global model while helping preserve privacy. Experimental results show that, under the situation of uneven dataset distribution among edge devices, LFNL achieves a comparable recognition accuracy to existing edge AI techniques, while substantially reducing data traffic by >3.5× and computational latency by >2.0×. Furthermore, LFNL significantly reduces energy consumption by >4.5× compared to standard federated learning with a slight accuracy loss up to 1.5%. Therefore, the proposed LFNL can facilitate the development of brain-inspired computing and edge AI.
| Original language | English (US) |
|---|---|
| Article number | 4269 |
| Journal | Nature communications |
| Volume | 13 |
| Issue number | 1 |
| DOIs | |
| State | Published - Dec 2022 |
| Externally published | Yes |
All Science Journal Classification (ASJC) codes
- General Chemistry
- General Biochemistry, Genetics and Molecular Biology
- General Physics and Astronomy