Classification of drones and birds using convolutional neural networks applied to radar micro-Doppler spectrogram images
Date
26/03/2020Grant ID
ST/N006569/1
Keywords
Metadata
Show full item recordAltmetrics Handle Statistics
Altmetrics DOI Statistics
Abstract
This study presents a convolutional neural network (CNN) based drone classification method. The primary criterion for a high-fidelity neural network based classification is a real dataset of large size and diversity for training. The first goal of the study was to create a large database of micro-Doppler spectrogram images of in-flight drones and birds. Two separate datasets with the same images have been created, one with RGB images and other with grayscale images. The RGB dataset was used for GoogLeNet architecture-based training. The grayscale dataset was used for training with a series architecture developed during this study. Each dataset was further divided into two categories, one with four classes (drone, bird, clutter and noise) and the other with two classes (drone and non-drone). During training, 20% of the dataset has been used as a validation set. After the completion of training, the models were tested with previously unseen and unlabelled sets of data. The validation and testing accuracy for the developed series network have been found to be 99.6% and 94.4% respectively for four classes and 99.3% and 98.3% respectively for two classes. The GoogLenet based model showed both validation and testing accuracies to be around 99% for all the cases.
Citation
Rahman , S & Robertson , D A 2020 , ' Classification of drones and birds using convolutional neural networks applied to radar micro-Doppler spectrogram images ' , IET Radar Sonar and Navigation , vol. 14 , no. 5 , pp. 653-661 . https://doi.org/10.1049/iet-rsn.2019.0493
Publication
IET Radar Sonar and Navigation
Status
Peer reviewed
ISSN
1751-8784Type
Journal article
Rights
Copyright © 2019 Institution of Engineering and Technology. This work has been made available online in accordance with publisher policies or with permission. Permission for further reuse of this content should be sought from the publisher or the rights holder. This is the author created accepted manuscript following peer review and may differ slightly from the final published version. The final published version of this work is available at https://doi.org/10.1049/iet-rsn.2019.0493
Description
Funding: UK Science and Technology Facilities Council ST/N006569/1 (DR).Collections
Items in the St Andrews Research Repository are protected by copyright, with all rights reserved, unless otherwise indicated.