期刊
NEUROCOMPUTING
卷 562, 期 -, 页码 -出版社
ELSEVIER
DOI: 10.1016/j.neucom.2023.126885
关键词
Spiking Neural Networks; Neuromorphic computing; Neuromorphic hardware; Reinforcement learning; DQN; UAV
This paper presents an energy-efficient implementation of a Reinforcement Learning algorithm using Spiking Neural Networks for obstacle avoidance task on an Unmanned Aerial Vehicle. The SNN algorithm achieves better results than a Convolutional Neural Network, with 6x less energy consumption.
Spiking Neural Networks (SNN) promise extremely low-power and low-latency inference on neuromorphic hardware. Recent studies demonstrate the competitive performance of SNNs compared with Artificial Neural Networks (ANN) in conventional classification tasks. In this work, we present an energy-efficient implementation of a Reinforcement Learning (RL) algorithm using SNNs to solve an obstacle avoidance task performed by an Unmanned Aerial Vehicle (UAV), taking a Dynamic Vision Sensor (DVS) as event-based input. We train the SNN directly, improving upon state-of-art implementations based on hybrid (not directly trained) SNNs. For this purpose, we devise an adaptation of the Spatio-Temporal Backpropagation algorithm (STBP) for RL. We then compare the SNN with a state-of-art Convolutional Neural Network (CNN) designed to solve the same task. To this aim, we train both networks by exploiting a photorealistic training pipeline based on AirSim. To achieve a realistic latency and throughput assessment for embedded deployment, we designed and trained three different embedded SNN versions to be executed on state-of-art neuromorphic hardware, targeting state-of-the-art. We compared SNN and CNN in terms of obstacle avoidance performance showing that the SNN algorithm achieves better results than the CNN with a factor of 6x less energy. We also characterize the different SNN hardware implementations in terms of energy and spiking activity.
作者
我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。
推荐
暂无数据