The final, formatted version of the article will be published soon.
ORIGINAL RESEARCH article
Front. Neurosci.
Sec. Neuromorphic Engineering
Volume 18 - 2024 |
doi: 10.3389/fnins.2024.1453419
SiamEFT: Adaptive-Time Feature Extraction Hybrid Network for RGBE Multi-Domain Object Tracking
Provisionally accepted- 1 School of Optics and Photonics, Beijing Institute of Technology, Beijing, China
- 2 Center of Brain Sciences, Beijing Institute of Basic Medical Sciences, Beijing, China
Integrating RGB and Event (RGBE) multi-domain information obtained by high-dynamic-range and temporal-resolution event cameras has been considered an effective scheme for robust object tracking. However, existing RGBE tracking methods have overlooked the unique spatio-temporal features over different domains, leading to object tracking failure and inefficiency, especally for objects against complex backgrounds. To address this problem, we propose a novel tracker based on adaptive-time feature extraction hybrid networks, namely Siamese Event Frame Tracker (SiamEFT), which focuses on the effective representation and utilization of the diverse spatiotemporal features of RGBE. We first design an adaptive-time attention module to aggregate event data into frames based on adaptive-time weights to enhance information representation.Subsequently, the SiamEF module and cross-network fusion module combining artificial neural networks and spiking neural networks hybrid network are designed to effectively extract and fuse the spatio-temporal features of RGBE. Extensive experiments on two RGBE datasets (VisEvent and COESOT) show that the SiamEFT achieves a success rate of 0.456 and 0.574, outperforming the state-of-the-art competing methods and exhibiting a 2.3-fold enhancement in efficiency. These results validate the superior accuracy and efficiency of SiamEFT in diverse and challenging scenes.
Keywords: RGB and Event, Spatio-temporal, Hybrid network, spiking neural networks, neuromorphic computing, object tracking RGB and Event, Object Tracking
Received: 23 Jun 2024; Accepted: 24 Jul 2024.
Copyright: © 2024 Liu, Wang, Song, Huang, Huang, Zhou and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Gang Wang, Center of Brain Sciences, Beijing Institute of Basic Medical Sciences, Beijing, China
Yong Song, School of Optics and Photonics, Beijing Institute of Technology, Beijing, China
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.