remotesensing-logo

Journal Browser

Journal Browser

Advances in Object and Activity Detection in Remote Sensing Imagery

A special issue of Remote Sensing (ISSN 2072-4292).

Deadline for manuscript submissions: closed (28 February 2022) | Viewed by 25334

Printed Edition Available!
A printed edition of this Special Issue is available here.

Special Issue Editors


E-Mail Website
Guest Editor
School of Computing and Mathematics, Charles Sturt University, Port Macquarie, NSW 2444, Australia
Interests: signal and image processing; machine learning; deep convolutional neural nets; data analytics; computer vision; thermal imaging
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Research fellow, Charles Sturt University, Charles Sturt University, Port Macquarie, NSW 2444, Australia
Interests: signal processing; machine learning; computer vision; pattern recognition

Special Issue Information

Dear Colleagues,

The widespread availability of drone and satellite imaging data has considerably enhanced air and space-borne monitoring of objects and their activities. Remote monitoring is crucial in various applications areas such as surveillance; border control; rescue operations for disaster management; precision agriculture; environment monitoring; weed detection; land, pest animal, wildlife, and sea-life surveys; and individual or group activity detection. These applications involve automated object and activity recognition.

The recent revolution in deep learning has enabled considerable development in the fields of object and activity detection. Visual object detection tries to find objects of target classes with precise localisation in an image and assign each object instance a corresponding class label. At the same time, activity recognition aims to determine the actions or activities of an agent or group of agents based on sensor or video observation data. It is a very important and challenging problem to detect, identify, track, and understand the behaviour of objects through images/videos taken by various cameras. Together, objects and their activity recognition in imaging data captured by remote sensing platforms is a highly dynamic and challenging research topic. During the last decade, there was significant growth in the number of publications in the field of object and activity recognition. In particular, many researchers have proposed application domains to identify objects and their specific behaviours from the air and spaceborne imagery.

This Special Issue welcomes papers that explore novel and challenging topics for object and activity detection in remote sensing images/videos acquired with diverse platforms.

We want to invite you to submit articles on object and activity detection in remote sensing imagery. We look forward to receiving your submissions which will be carefully reviewed within a much shorter turnaround time than most current journals in this domain.

Dr. Anwaar Ulhaq
Dr. Douglas Pinto Sampaio Gomes
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Remote Sensing is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2700 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • Object detection in aerial imagery
  • Action recognition in UAV/satellite imagery
  • Activity detection in remote sensing
  • Weed detection in remote sensing images
  • Border control and surveillance from UAV
  • Group activity detection
  • Sea- and wildlife monitoring
  • Deep learning-based object and activity detection

Published Papers (8 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Editorial

Jump to: Research

2 pages, 152 KiB  
Editorial
Editorial for the Special Issue “Advances in Object and Activity Detection in Remote Sensing Imagery”
by Anwaar Ulhaq and Douglas Pinto Sampaio Gomes
Remote Sens. 2022, 14(8), 1844; https://0-doi-org.brum.beds.ac.uk/10.3390/rs14081844 - 12 Apr 2022
Viewed by 1086
Abstract
Advances in data collection and accessibility, such as unmanned aerial vehicle (UAV) technology, the availability of satellite imagery, and the increasing performance of deep learning models, have had significant impacts on solving various remote sensing problems and proposing new applications ranging from vegetation [...] Read more.
Advances in data collection and accessibility, such as unmanned aerial vehicle (UAV) technology, the availability of satellite imagery, and the increasing performance of deep learning models, have had significant impacts on solving various remote sensing problems and proposing new applications ranging from vegetation and wildlife monitoring to crowd monitoring [...] Full article
(This article belongs to the Special Issue Advances in Object and Activity Detection in Remote Sensing Imagery)

Research

Jump to: Editorial

17 pages, 23169 KiB  
Article
A Multiview Semantic Vegetation Index for Robust Estimation of Urban Vegetation Cover
by Asim Khan, Warda Asim, Anwaar Ulhaq and Randall W. Robinson
Remote Sens. 2022, 14(1), 228; https://0-doi-org.brum.beds.ac.uk/10.3390/rs14010228 - 05 Jan 2022
Cited by 4 | Viewed by 2787
Abstract
Urban vegetation growth is vital for developing sustainable and liveable cities in the contemporary era since it directly helps people’s health and well-being. Estimating vegetation cover and biomass is commonly done by calculating various vegetation indices for automated urban vegetation management and monitoring. [...] Read more.
Urban vegetation growth is vital for developing sustainable and liveable cities in the contemporary era since it directly helps people’s health and well-being. Estimating vegetation cover and biomass is commonly done by calculating various vegetation indices for automated urban vegetation management and monitoring. However, most of these indices fail to capture robust estimation of vegetation cover due to their inherent focus on colour attributes with limited viewpoint and ignore seasonal changes. To solve this limitation, this article proposed a novel vegetation index called the Multiview Semantic Vegetation Index (MSVI), which is robust to color, viewpoint, and seasonal variations. Moreover, it can be applied directly to RGB images. This Multiview Semantic Vegetation Index (MSVI) is based on deep semantic segmentation and multiview field coverage and can be integrated into any vegetation management platform. This index has been tested on Google Street View (GSV) imagery of Wyndham City Council, Melbourne, Australia. The experiments and training achieved an overall pixel accuracy of 89.4% and 92.4% for FCN and U-Net, respectively. Thus, the MSVI can be a helpful instrument for analysing urban forestry and vegetation biomass since it provides an accurate and reliable objective method for assessing the plant cover at street level. Full article
(This article belongs to the Special Issue Advances in Object and Activity Detection in Remote Sensing Imagery)
Show Figures

Graphical abstract

14 pages, 10029 KiB  
Article
Automated Detection of Animals in Low-Resolution Airborne Thermal Imagery
by Anwaar Ulhaq, Peter Adams, Tarnya E. Cox, Asim Khan, Tom Low and Manoranjan Paul
Remote Sens. 2021, 13(16), 3276; https://0-doi-org.brum.beds.ac.uk/10.3390/rs13163276 - 19 Aug 2021
Cited by 10 | Viewed by 3192
Abstract
Detecting animals to estimate abundance can be difficult, particularly when the habitat is dense or the target animals are fossorial. The recent surge in the use of thermal imagers in ecology and their use in animal detections can increase the accuracy of population [...] Read more.
Detecting animals to estimate abundance can be difficult, particularly when the habitat is dense or the target animals are fossorial. The recent surge in the use of thermal imagers in ecology and their use in animal detections can increase the accuracy of population estimates and improve the subsequent implementation of management programs. However, the use of thermal imagers results in many hours of captured flight videos which require manual review for confirmation of species detection and identification. Therefore, the perceived cost and efficiency trade-off often restricts the use of these systems. Additionally, for many off-the-shelf systems, the exported imagery can be quite low resolution (<9 Hz), increasing the difficulty of using automated detections algorithms to streamline the review process. This paper presents an animal species detection system that utilises the cost-effectiveness of these lower resolution thermal imagers while harnessing the power of transfer learning and an enhanced small object detection algorithm. We have proposed a distant object detection algorithm named Distant-YOLO (D-YOLO) that utilises YOLO (You Only Look Once) and improves its training and structure for the automated detection of target objects in thermal imagery. We trained our system on thermal imaging data of rabbits, their active warrens, feral pigs, and kangaroos collected by thermal imaging researchers in New South Wales and Western Australia. This work will enhance the visual analysis of animal species while performing well on low, medium and high-resolution thermal imagery. Full article
(This article belongs to the Special Issue Advances in Object and Activity Detection in Remote Sensing Imagery)
Show Figures

Figure 1

33 pages, 10883 KiB  
Article
A Wide Area Multiview Static Crowd Estimation System Using UAV and 3D Training Simulator
by Shivang Shukla, Bernard Tiddeman and Helen C. Miles
Remote Sens. 2021, 13(14), 2780; https://0-doi-org.brum.beds.ac.uk/10.3390/rs13142780 - 15 Jul 2021
Cited by 2 | Viewed by 2859
Abstract
Crowd size estimation is a challenging problem, especially when the crowd is spread over a significant geographical area. It has applications in monitoring of rallies and demonstrations and in calculating the assistance requirements in humanitarian disasters. Therefore, accomplishing a crowd surveillance system for [...] Read more.
Crowd size estimation is a challenging problem, especially when the crowd is spread over a significant geographical area. It has applications in monitoring of rallies and demonstrations and in calculating the assistance requirements in humanitarian disasters. Therefore, accomplishing a crowd surveillance system for large crowds constitutes a significant issue. UAV-based techniques are an appealing choice for crowd estimation over a large region, but they present a variety of interesting challenges, such as integrating per-frame estimates through a video without counting individuals twice. Large quantities of annotated training data are required to design, train, and test such a system. In this paper, we have first reviewed several crowd estimation techniques, existing crowd simulators and data sets available for crowd analysis. Later, we have described a simulation system to provide such data, avoiding the need for tedious and error-prone manual annotation. Then, we have evaluated synthetic video from the simulator using various existing single-frame crowd estimation techniques. Our findings show that the simulated data can be used to train and test crowd estimation, thereby providing a suitable platform to develop such techniques. We also propose an automated UAV-based 3D crowd estimation system that can be used for approximately static or slow-moving crowds, such as public events, political rallies, and natural or man-made disasters. We evaluate the results by applying our new framework to a variety of scenarios with varying crowd sizes. The proposed system gives promising results using widely accepted metrics including MAE, RMSE, Precision, Recall, and F1 score to validate the results. Full article
(This article belongs to the Special Issue Advances in Object and Activity Detection in Remote Sensing Imagery)
Show Figures

Figure 1

30 pages, 15260 KiB  
Article
Quad-FPN: A Novel Quad Feature Pyramid Network for SAR Ship Detection
by Tianwen Zhang, Xiaoling Zhang and Xiao Ke
Remote Sens. 2021, 13(14), 2771; https://0-doi-org.brum.beds.ac.uk/10.3390/rs13142771 - 14 Jul 2021
Cited by 121 | Viewed by 5688
Abstract
Ship detection from synthetic aperture radar (SAR) imagery is a fundamental and significant marine mission. It plays an important role in marine traffic control, marine fishery management, and marine rescue. Nevertheless, there are still some challenges hindering accuracy improvements of SAR ship detection, [...] Read more.
Ship detection from synthetic aperture radar (SAR) imagery is a fundamental and significant marine mission. It plays an important role in marine traffic control, marine fishery management, and marine rescue. Nevertheless, there are still some challenges hindering accuracy improvements of SAR ship detection, e.g., complex background interferences, multi-scale ship feature differences, and indistinctive small ship features. Therefore, to address these problems, a novel quad feature pyramid network (Quad-FPN) is proposed for SAR ship detection in this paper. Quad-FPN consists of four unique FPNs, i.e., a DEformable COnvolutional FPN (DE-CO-FPN), a Content-Aware Feature Reassembly FPN (CA-FR-FPN), a Path Aggregation Space Attention FPN (PA-SA-FPN), and a Balance Scale Global Attention FPN (BS-GA-FPN). To confirm the effectiveness of each FPN, extensive ablation studies are conducted. We conduct experiments on five open SAR ship detection datasets, i.e., SAR ship detection dataset (SSDD), Gaofen-SSDD, Sentinel-SSDD, SAR-Ship-Dataset, and high-resolution SAR images dataset (HRSID). Qualitative and quantitative experimental results jointly reveal Quad-FPN’s optimal SAR ship detection performance compared with the other 12 competitive state-of-the-art convolutional neural network (CNN)-based SAR ship detectors. To confirm the excellent migration application capability of Quad-FPN, the actual ship detection in another two large-scene Sentinel-1 SAR images is conducted. Their satisfactory detection results indicate the practical application value of Quad-FPN in marine surveillance. Full article
(This article belongs to the Special Issue Advances in Object and Activity Detection in Remote Sensing Imagery)
Show Figures

Graphical abstract

18 pages, 12840 KiB  
Article
ADT-Det: Adaptive Dynamic Refined Single-Stage Transformer Detector for Arbitrary-Oriented Object Detection in Satellite Optical Imagery
by Yongbin Zheng, Peng Sun, Zongtan Zhou, Wanying Xu and Qiang Ren
Remote Sens. 2021, 13(13), 2623; https://0-doi-org.brum.beds.ac.uk/10.3390/rs13132623 - 04 Jul 2021
Cited by 25 | Viewed by 3418
Abstract
The detection of arbitrary-oriented and multi-scale objects in satellite optical imagery is an important task in remote sensing and computer vision. Despite significant research efforts, such detection remains largely unsolved due to the diversity of patterns in orientation, scale, aspect ratio, and visual [...] Read more.
The detection of arbitrary-oriented and multi-scale objects in satellite optical imagery is an important task in remote sensing and computer vision. Despite significant research efforts, such detection remains largely unsolved due to the diversity of patterns in orientation, scale, aspect ratio, and visual appearance; the dense distribution of objects; and extreme imbalances in categories. In this paper, we propose an adaptive dynamic refined single-stage transformer detector to address the aforementioned challenges, aiming to achieve high recall and speed. Our detector realizes rotated object detection with RetinaNet as the baseline. Firstly, we propose a feature pyramid transformer (FPT) to enhance feature extraction of the rotated object detection framework through a feature interaction mechanism. This is beneficial for the detection of objects with diverse patterns in terms of scale, aspect ratio, visual appearance, and dense distributions. Secondly, we design two special post-processing steps for rotated objects with arbitrary orientations, large aspect ratios and dense distributions. The output features of FPT are fed into post-processing steps. In the first step, it performs the preliminary regression of locations and angle anchors for the refinement step. In the refinement step, it performs adaptive feature refinement first and then gives the final object detection result precisely. The main architecture of the refinement step is dynamic feature refinement (DFR), which is proposed to adaptively adjust the feature map and reconstruct a new feature map for arbitrary-oriented object detection to alleviate the mismatches between rotated bounding boxes and axis-aligned receptive fields. Thirdly, the focus loss is adopted to deal with the category imbalance problem. Experiments on two challenging satellite optical imagery public datasets, DOTA and HRSC2016, demonstrate that the proposed ADT-Det detector achieves a state-of-the-art detection accuracy (79.95% mAP for DOTA and 93.47% mAP for HRSC2016) while running very fast (14.6 fps with a 600 × 600 input image size). Full article
(This article belongs to the Special Issue Advances in Object and Activity Detection in Remote Sensing Imagery)
Show Figures

Figure 1

19 pages, 5468 KiB  
Article
Fine-Grained Tidal Flat Waterbody Extraction Method (FYOLOv3) for High-Resolution Remote Sensing Images
by Lili Zhang, Yu Fan, Ruijie Yan, Yehong Shao, Gaoxu Wang and Jisen Wu
Remote Sens. 2021, 13(13), 2594; https://0-doi-org.brum.beds.ac.uk/10.3390/rs13132594 - 02 Jul 2021
Cited by 8 | Viewed by 2126
Abstract
The tidal flat is long and narrow area along rivers and coasts with high sediment content, so there is little feature difference between the waterbody and the background, and the boundary of the waterbody is blurry. The existing waterbody extraction methods are mostly [...] Read more.
The tidal flat is long and narrow area along rivers and coasts with high sediment content, so there is little feature difference between the waterbody and the background, and the boundary of the waterbody is blurry. The existing waterbody extraction methods are mostly used for the extraction of large water bodies like rivers and lakes, whereas less attention has been paid to tidal flat waterbody extraction. Extracting tidal flat waterbody accurately from high-resolution remote sensing imagery is a great challenge. In order to solve the low accuracy problem of tidal flat waterbody extraction, we propose a fine-grained tidal flat waterbody extraction method, named FYOLOv3, which can extract tidal flat water with high accuracy. The FYOLOv3 mainly includes three parts: an improved object detection network based on YOLOv3 (Seattle, WA, USA), a fully convolutional network (FCN) without pooling layers, and a similarity algorithm for water extraction. The improved object detection network uses 13 convolutional layers instead of Darknet-53 as the model backbone network, which guarantees the water detection accuracy while reducing the time cost and alleviating the overfitting phenomenon; secondly, the FCN without pooling layers is proposed to obtain the accurate pixel value of the tidal flat waterbody by learning the semantic information; finally, a similarity algorithm for water extraction is proposed to distinguish the waterbody from non-water pixel by pixel to improve the extraction accuracy of tidal flat water bodies. Compared to the other convolutional neural network (CNN) models, the experiments show that our method has higher accuracy on the waterbody extraction of tidal flats from remote sensing images, and the IoU of our method is 2.43% higher than YOLOv3 and 3.7% higher than U-Net (Freiburg, Germany). Full article
(This article belongs to the Special Issue Advances in Object and Activity Detection in Remote Sensing Imagery)
Show Figures

Graphical abstract

26 pages, 4312 KiB  
Article
UAV-Assisted Wide Area Multi-Camera Space Alignment Based on Spatiotemporal Feature Map
by Jing Li, Yuguang Xie, Congcong Li, Yanran Dai, Jiaxin Ma, Zheng Dong and Tao Yang
Remote Sens. 2021, 13(6), 1117; https://0-doi-org.brum.beds.ac.uk/10.3390/rs13061117 - 15 Mar 2021
Cited by 3 | Viewed by 2664
Abstract
In this paper, we investigate the problem of aligning multiple deployed camera into one united coordinate system for cross-camera information sharing and intercommunication. However, the difficulty is greatly increased when faced with large-scale scene under chaotic camera deployment. To address this problem, we [...] Read more.
In this paper, we investigate the problem of aligning multiple deployed camera into one united coordinate system for cross-camera information sharing and intercommunication. However, the difficulty is greatly increased when faced with large-scale scene under chaotic camera deployment. To address this problem, we propose a UAV-assisted wide area multi-camera space alignment approach based on spatiotemporal feature map. It employs the great global perception of Unmanned Aerial Vehicles (UAVs) to meet the challenge from wide-range environment. Concretely, we first present a novel spatiotemporal feature map construction approach to represent the input aerial and ground monitoring data. In this way, the motion consistency across view is well mined to overcome the great perspective gap between the UAV and ground cameras. To obtain the corresponding relationship between their pixels, we propose a cross-view spatiotemporal matching strategy. Through solving relative relationship with the above air-to-ground point correspondences, all ground cameras can be aligned into one surveillance space. The proposed approach was evaluated in both simulation and real environments qualitatively and quantitatively. Extensive experimental results demonstrate that our system can successfully align all ground cameras with very small pixel error. Additionally, the comparisons with other works on different test situations also verify its superior performance. Full article
(This article belongs to the Special Issue Advances in Object and Activity Detection in Remote Sensing Imagery)
Show Figures

Graphical abstract

Back to TopTop