Machine Learning Techniques Applied to Geospatial Big Data

A special issue of Applied Sciences (ISSN 2076-3417). This special issue belongs to the section "Environmental Sciences".

Deadline for manuscript submissions: closed (30 September 2021) | Viewed by 69135

Special Issue Editors


E-Mail Website
Guest Editor
1. Geological Research Division, Korea Institute of Geoscience and Mineral Resources (KIGAM), 124, Gwahak-ro Yuseong-gu, Daejeon 34132, Republic of Korea
2. Department of Geophysical Exploration, Korea University of Science and Technology, 217 Gajeong-ro Yuseong-gu, Daejeon 34113, Republic of Korea
Interests: GIS application; geological hazard; geological resources
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor

Special Issue Information

Dear Colleagues,

As computer and space technologies have been developed, geoscience information systems (GIS) and remote sensing (RS), which deal with spatial information, have been maturing rapidly. In addition, over the last few decades, machine learning techniques on geospatial big data have been successfully applied to science and engineering research fields. Machine learning techniques have been widely applied to GIS and RS research fields and have recently produced valuable results in the areas of geoscience, environment, natural hazards, and natural resources.

This Special Issue of Applied Sciences, “Machine Learning Techniques Applied to Geospatial Big Data”, aims to attract novel contributions covering machine learning techniques applied to geospatial big data in the field of GIS and RS.

Topics of interest include but are not limited to the following:

  • The application of machine learning techniques combined with GIS;
  • The application of machine learning techniques to remote sensing;
  • The application of machine learning techniques to global positioning systems (GPS);
  • Spatial analysis and geocomputation based on machine learning techniques;
  • Spatial prediction using machine learning techniques;
  • Geospatial big data processing of geoinformation using machine learning techniques;
  • Comparison analysis among several machine learning techniques applied to GIS and RS;
  • The application of machine learning techniques to geosciences, environments, natural hazards, and natural resources as case studies.

Prof. Dr. Saro Lee
Prof. Dr. Hyung-Sup Jung
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Applied Sciences is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • Machine learning
  • Deep learning
  • Data mining
  • Geospatial big data
  • Geoinformatics
  • Geoscience information systems (GIS)
  • Remote sensing
  • Global positioning systems (GPS)
  • Spatial analysis

Published Papers (15 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

19 pages, 3476 KiB  
Article
A Method for Identifying Geospatial Data Sharing Websites by Combining Multi-Source Semantic Information and Machine Learning
by Quanying Cheng, Yunqiang Zhu, Hongyun Zeng, Jia Song, Shu Wang, Jinqu Zhang, Lang Qian and Yanmin Qi
Appl. Sci. 2021, 11(18), 8705; https://0-doi-org.brum.beds.ac.uk/10.3390/app11188705 - 18 Sep 2021
Cited by 3 | Viewed by 1885
Abstract
Geospatial data sharing is an inevitable requirement for scientific and technological innovation and economic and social development decisions in the era of big data. With the development of modern information technology, especially Web 2.0, a large number of geospatial data sharing websites (GDSW) [...] Read more.
Geospatial data sharing is an inevitable requirement for scientific and technological innovation and economic and social development decisions in the era of big data. With the development of modern information technology, especially Web 2.0, a large number of geospatial data sharing websites (GDSW) have been developed on the Internet. GDSW is a point of access to geospatial data, which is able to provide a geospatial data inventory. How to precisely identify these data websites is the foundation and prerequisite of sharing and utilizing web geospatial data and is also the main challenge of data sharing at this stage. GDSW identification can be regarded as a binary website classification problem, which can be solved by the current popular machine learning method. However, the websites obtained from the Internet contain a large number of blogs, companies, institutions, etc. If GDSW is directly used as the sample data of machine learning, it will greatly affect the classification precision. For this reason, this paper proposes a method to precisely identify GDSW by combining multi-source semantic information and machine learning. Firstly, based on the keyword set, we used the Baidu search engine to find the websites that may be related to geospatial data in the open web environment. Then, we used the multi-source semantic information of geospatial data content, morphology, sources, and shared websites to filter out a large number of websites that contained geospatial keywords but were not related to geospatial data in the search results through the calculation of comprehensive similarity. Finally, the filtered geospatial data websites were used as the sample data of machine learning, and the GDSWs were identified and evaluated. In this paper, training sets are extracted from the original search data and the data filtered by multi-source semantics, the two datasets are trained by machine learning classification algorithms (KNN, LR, RF, and SVM), and the same test datasets are predicted. The results show that: (1) compared with the four classification algorithms, the classification precision of RF and SVM on the original data is higher than that of the other two algorithms. (2) Taking the data filtered by multi-source semantic information as the sample data for machine learning, the precision of all classification algorithms has been greatly improved. The SVM algorithm has the highest precision among the four classification algorithms. (3) In order to verify the robustness of this method, different initial sample data mentioned above are selected for classification using the same method. The results show that, among the four classification algorithms, the classification precision of SVM is still the highest, which shows that the proposed method is robust and scalable. Therefore, taking the data filtered by multi-source semantic information as the sample data to train through machine learning can effectively improve the classification precision of GDSW, and comparing the four classification algorithms, SVM has the best classification effect. In addition, this method has good robustness, which is of great significance to promote and facilitate the sharing and utilization of open geospatial data. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

15 pages, 7329 KiB  
Article
Automatic Detection of Cracks in Asphalt Pavement Using Deep Learning to Overcome Weaknesses in Images and GIS Visualization
by Pang-jo Chun, Tatsuro Yamane and Yukino Tsuzuki
Appl. Sci. 2021, 11(3), 892; https://0-doi-org.brum.beds.ac.uk/10.3390/app11030892 - 20 Jan 2021
Cited by 33 | Viewed by 4582
Abstract
The crack ratio is one of the indices used to quantitatively evaluate the soundness of asphalt pavement. However, since the inspection of pavement requires much labor and cost, automatic inspection of pavement damage by image analysis is required in order to reduce the [...] Read more.
The crack ratio is one of the indices used to quantitatively evaluate the soundness of asphalt pavement. However, since the inspection of pavement requires much labor and cost, automatic inspection of pavement damage by image analysis is required in order to reduce the burden of such work. In this study, a system was constructed that automatically detects and evaluates cracks from images of pavement using a convolutional neural network, a kind of deep learning. The most novel aspect of this study is that the accuracy was recursively improved through retraining the convolutional neural network (CNN) by collecting images which had previously been incorrectly analyzed. Then, study and implementation were conducted of a system for plotting the results in a GIS. In addition, an experiment was carried out applying this system to images actually taken from an MMS (mobile mapping system), and this confirmed that the system had high crack evaluation performance. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

22 pages, 1955 KiB  
Article
The Capacitated Location-Allocation Problem Using the VAOMP (Vector Assignment Ordered Median Problem) Unified Approach in GIS (Geospatial Information Systam)
by Alireza Vafaeinejad, Samira Bolouri, Ali Asghar Alesheikh, Mahdi Panahi and Chang-Wook Lee
Appl. Sci. 2020, 10(23), 8505; https://0-doi-org.brum.beds.ac.uk/10.3390/app10238505 - 28 Nov 2020
Cited by 8 | Viewed by 2526
Abstract
The Vector Assignment Ordered Median Problem (VAOMP) is a new unified approach for location-allocation problems, which are one of the most important forms of applied analysis in GIS (Geospatial Information System). Solving location-allocation problems with exact methods is difficult and time-consuming, especially when [...] Read more.
The Vector Assignment Ordered Median Problem (VAOMP) is a new unified approach for location-allocation problems, which are one of the most important forms of applied analysis in GIS (Geospatial Information System). Solving location-allocation problems with exact methods is difficult and time-consuming, especially when the number of objectives and criteria increases. One of the most important criteria in location-allocation problems is the capacity of facilities. Firstly, this study develops a new VAOMP approach by including capacity as a criterion, resulting in a new model known as VAOCMP (Vector Assignment Ordered Capacitated Median Problem). Then secondly, the results of applying VAOMP, in scenario 1, and VAOCMP, in scenario 2, for the location-allocation of fire stations in Tehran, with the objective of minimizing the arrival time of fire engines to an incident site to no more than 5 min, are examined using both the Tabu Search and Simulated Annealing algorithms in GIS. The results of scenario 1 show that 52,840 demands were unable to be served with 10 existing stations. In scenario 2, given that each facility could not accept demand above its capacity, the number of demands without service increased to 59,080, revealing that the number of stations in the study area is insufficient. Adding 35 candidate stations and performing relocation-reallocation revealed that at least three other stations are needed for optimal service. Thirdly, and finally, the VAOMP and VAOCMP were implemented in a modest size problem. The implementation results for both algorithms showed that the Tabu Search algorithm performed more effectively. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

18 pages, 6290 KiB  
Article
Susceptibility Mapping on Urban Landslides Using Deep Learning Approaches in Mt. Umyeon
by Sunmin Lee, Won-Kyung Baek, Hyung-Sup Jung and Saro Lee
Appl. Sci. 2020, 10(22), 8189; https://0-doi-org.brum.beds.ac.uk/10.3390/app10228189 - 19 Nov 2020
Cited by 31 | Viewed by 3054
Abstract
In recent years, the incidence of localized heavy rainfall has increased as abnormal weather events occur more frequently. In densely populated urban areas, this type of heavy rain can cause extreme landslide damage, so that it is necessary to estimate and analyze the [...] Read more.
In recent years, the incidence of localized heavy rainfall has increased as abnormal weather events occur more frequently. In densely populated urban areas, this type of heavy rain can cause extreme landslide damage, so that it is necessary to estimate and analyze the susceptibility of future landslides. In this regard, deep learning (DL) methodologies have been used to identify areas prone to landslides recently. Therefore, in this study, DL methodologies, including a deep neural network (DNN), kernel-based DNN, and convolutional neural network (CNN) were used to identify areas where landslides could occur. As a detailed step for this purpose, landslide occurrence was first determined as landslide inventory through aerial photographs with comparative analysis using field survey data; a training set was built for model training through oversampling based on the landslide inventory. A total of 17 landslide influencing variables that influence the frequency of landslides by topography and geomorphology, as well as soil and forest variables, were selected to establish a landslide inventory. Then models were built using DNN, kernel-based DNN, and CNN models, and the susceptibility of landslides in the study area was determined. Model performance was evaluated through the average precision (AP) score and root mean square error (RMSE) for each of the three models. Finally, DNN, kernel-based DNN, and CNN models showed performances of 99.45%, 99.44%, and 99.41%, and RMSE values of 0.1694, 0.1806, and 0.1747, respectively. As a result, all three models showed similar performance, indicating excellent predictive ability of the models developed in this study. The information of landslides occurring in urban areas, which cause a great damage even with a small number of occurrences, can provide a basis for reference to the government and local authorities for urban landslide management. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

16 pages, 1750 KiB  
Article
Scenario-Based Marine Oil Spill Emergency Response Using Hybrid Deep Reinforcement Learning and Case-Based Reasoning
by Kui Huang, Wen Nie and Nianxue Luo
Appl. Sci. 2020, 10(15), 5269; https://0-doi-org.brum.beds.ac.uk/10.3390/app10155269 - 30 Jul 2020
Cited by 13 | Viewed by 3554
Abstract
Case-based reasoning (CBR) systems often provide a basis for decision makers to make management decisions in disaster prevention and emergency response. For decades, many CBR systems have been implemented by using expert knowledge schemes to build indexes for case identification from a case [...] Read more.
Case-based reasoning (CBR) systems often provide a basis for decision makers to make management decisions in disaster prevention and emergency response. For decades, many CBR systems have been implemented by using expert knowledge schemes to build indexes for case identification from a case library of situations and to explore the relations among cases. However, a knowledge elicitation bottleneck occurs for many knowledge-based CBR applications because expert reasoning is difficult to precisely explain. To solve these problems, this paper proposes a method using only knowledge to recognize marine oil spill cases. The proposed method combines deep reinforcement learning (DRL) with strategy selection to determine emergency responses for marine oil spill accidents by quantification of the marine oil spill scenario as the reward for the DRL agent. These accidents are described by scenarios and are considered the state inputs in the hybrid DRL/CBR framework. The challenges and opportunities of the proposed method are discussed considering different scenarios and the intentions of decision makers. This approach may be helpful in terms of developing hybrid DRL/CBR-based tools for marine oil spill emergency response. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

26 pages, 4996 KiB  
Article
Comparison of Support Vector Machine, Bayesian Logistic Regression, and Alternating Decision Tree Algorithms for Shallow Landslide Susceptibility Mapping along a Mountainous Road in the West of Iran
by Viet-Ha Nhu, Danesh Zandi, Himan Shahabi, Kamran Chapi, Ataollah Shirzadi, Nadhir Al-Ansari, Sushant K. Singh, Jie Dou and Hoang Nguyen
Appl. Sci. 2020, 10(15), 5047; https://0-doi-org.brum.beds.ac.uk/10.3390/app10155047 - 22 Jul 2020
Cited by 52 | Viewed by 5278
Abstract
This paper aims to apply and compare the performance of the three machine learning algorithms–support vector machine (SVM), bayesian logistic regression (BLR), and alternating decision tree (ADTree)–to map landslide susceptibility along the mountainous road of the Salavat Abad saddle, Kurdistan province, Iran. We [...] Read more.
This paper aims to apply and compare the performance of the three machine learning algorithms–support vector machine (SVM), bayesian logistic regression (BLR), and alternating decision tree (ADTree)–to map landslide susceptibility along the mountainous road of the Salavat Abad saddle, Kurdistan province, Iran. We identified 66 shallow landslide locations, based on field surveys, by recording the locations of the landslides by a global position System (GPS), Google Earth imagery and black-and-white aerial photographs (scale 1: 20,000) and 19 landslide conditioning factors, then tested these factors using the information gain ratio (IGR) technique. We checked the validity of the models using statistical metrics, including sensitivity, specificity, accuracy, kappa, root mean square error (RMSE), and area under the receiver operating characteristic curve (AUC). We found that, although all three machine learning algorithms yielded excellent performance, the SVM algorithm (AUC = 0.984) slightly outperformed the BLR (AUC = 0.980), and ADTree (AUC = 0.977) algorithms. We observed that not only all three algorithms are useful and effective tools for identifying shallow landslide-prone areas but also the BLR algorithm can be used such as the SVM algorithm as a soft computing benchmark algorithm to check the performance of the models in future. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

17 pages, 3781 KiB  
Article
Short-Term Spatio-Temporal Drought Forecasting Using Random Forests Model at New South Wales, Australia
by Abhirup Dikshit, Biswajeet Pradhan and Abdullah M. Alamri
Appl. Sci. 2020, 10(12), 4254; https://0-doi-org.brum.beds.ac.uk/10.3390/app10124254 - 21 Jun 2020
Cited by 46 | Viewed by 5188
Abstract
Droughts can cause significant damage to agriculture and water resources, leading to severe economic losses and loss of life. One of the most important aspect is to develop effective tools to forecast drought events that could be helpful in mitigation strategies. The understanding [...] Read more.
Droughts can cause significant damage to agriculture and water resources, leading to severe economic losses and loss of life. One of the most important aspect is to develop effective tools to forecast drought events that could be helpful in mitigation strategies. The understanding of droughts has become more challenging because of the effect of climate change, urbanization and water management; therefore, the present study aims to forecast droughts by determining an appropriate index and analyzing its changes, using climate variables. The work was conducted in three different phases, first being the determination of Standard Precipitation Evaporation Index (SPEI), using global climatic dataset of Climate Research Unit (CRU) from 1901–2018. The indices are calculated at different monthly intervals which could depict short-term or long-term changes, and the index value represents different drought classes, ranging from extremely dry to extremely wet. However, the present study was focused only on forecasting at short-term scales for New South Wales (NSW) region of Australia and was conducted at two different time scales, one month and three months. The second phase involved dividing the data into three sample sizes, training (1901–2010), testing (2011–2015) and validation (2016–2018). Finally, a machine learning approach, Random Forest (RF), was used to train and test the data, using various climatic variables, e.g., rainfall, potential evapotranspiration, cloud cover, vapor pressure and temperature (maximum, minimum and mean). The final phase was to analyze the performance of the model based on statistical metrics and drought classes. Regarding this, the performance of the testing period was conducted by using statistical metrics, Coefficient of Determination (R2) and Root-Mean-Square-Error (RMSE) method. The performance of the model showed a considerably higher value of R2 for both the time scales. However, statistical metrics analyzes the variation between the predicted and observed index values, and it does not consider the drought classes. Therefore, the variation in predicted and observed SPEI values were analyzed based on different drought classes, which were validated by using the Receiver Operating Characteristic (ROC)-based Area under the Curve (AUC) approach. The results reveal that the classification of drought classes during the validation period had an AUC of 0.82 for SPEI 1 case and 0.84 for SPEI 3 case. The study depicts that the Random Forest model can perform both regression and classification analysis for drought studies in NSW. The work also suggests that the performance of any model for drought forecasting should not be limited only through statistical metrics, but also by examining the variation in terms of drought characteristics. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Graphical abstract

21 pages, 6366 KiB  
Article
An Artificial Intelligence Approach to Prediction of Corn Yields under Extreme Weather Conditions Using Satellite and Meteorological Data
by Nari Kim, Sang-Il Na, Chan-Won Park, Morang Huh, Jaiho Oh, Kyung-Ja Ha, Jaeil Cho and Yang-Won Lee
Appl. Sci. 2020, 10(11), 3785; https://0-doi-org.brum.beds.ac.uk/10.3390/app10113785 - 29 May 2020
Cited by 23 | Viewed by 4584
Abstract
This paper describes the development of an optimized corn yield prediction model under extreme weather conditions for the Midwestern United States (US). We tested six different artificial intelligence (AI) models using satellite images and meteorological data for the dominant growth period. To examine [...] Read more.
This paper describes the development of an optimized corn yield prediction model under extreme weather conditions for the Midwestern United States (US). We tested six different artificial intelligence (AI) models using satellite images and meteorological data for the dominant growth period. To examine the effects of extreme weather events, we defined the drought and heatwave by considering the characteristics of corn growth and selected the cases for sensitivity tests from a historical database. In particular, we conducted an optimization for the hyperparameters of the deep neural network (DNN) model to ensure the best configuration for accuracy improvement. The result for drought cases showed that our DNN model was approximately 51–98% more accurate than the other five AI models in terms of root mean square error (RMSE). For the heatwave cases, our DNN model showed approximately 30–77% better accuracy in terms of RMSE. The correlation coefficient was 0.954 for drought cases and 0.887–0.914 for heatwave cases. Moreover, the accuracy of our DNN model was very stable, despite the increases in the duration of the heatwave. It indicates that the optimized DNN model can provide robust predictions for corn yield under conditions of extreme weather and can be extended to other prediction models for various crops in future work. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

31 pages, 10536 KiB  
Article
Evaluating the Performance of Individual and Novel Ensemble of Machine Learning and Statistical Models for Landslide Susceptibility Assessment at Rudraprayag District of Garhwal Himalaya
by Sunil Saha, Anik Saha, Tusar Kanti Hembram, Biswajeet Pradhan and Abdullah M. Alamri
Appl. Sci. 2020, 10(11), 3772; https://0-doi-org.brum.beds.ac.uk/10.3390/app10113772 - 29 May 2020
Cited by 43 | Viewed by 3731
Abstract
Landslides are known as the world’s most dangerous threat in mountainous regions and pose a critical obstacle for both economic and infrastructural progress. It is, therefore, quite relevant to discuss the pattern of spatial incidence of this phenomenon. The current research manifests a [...] Read more.
Landslides are known as the world’s most dangerous threat in mountainous regions and pose a critical obstacle for both economic and infrastructural progress. It is, therefore, quite relevant to discuss the pattern of spatial incidence of this phenomenon. The current research manifests a set of individual and ensemble of machine learning and probabilistic approaches like an artificial neural network (ANN), support vector machine (SVM), random forest (RF), logistic regression (LR), and their ensembles such as ANN-RF, ANN-SVM, SVM-RF, SVM-LR, LR-RF, LR-ANN, ANN-LR-RF, ANN-RF-SVM, ANN-SVM-LR, RF-SVM-LR, and ANN-RF-SVM-LR for mapping landslide susceptibility in Rudraprayag district of Garhwal Himalaya, India. A landslide inventory map along with sixteen landslide conditioning factors (LCFs) was used. Randomly partitioned sets of 70%:30% were used to ascertain the goodness of fit and predictive ability of the models. The contribution of LCFs was analyzed using the RF model. The altitude and drainage density were found to be the responsible factors in causing the landslide in the study area according to the RF model. The robustness of models was assessed through three threshold dependent measures, i.e., receiver operating characteristic (ROC), precision and accuracy, and two threshold independent measures, i.e., mean-absolute-error (MAE) and root-mean-square-error (RMSE). Finally, using the compound factor (CF) method, the models were prioritized based on the results of the validation methods to choose best model. Results show that ANN-RF-LR indicated a realistic finding, concentrating only on 17.74% of the study area as highly susceptible to landslide. The ANN-RF-LR ensemble demonstrated the highest goodness of fit and predictive capacity with respective values of 87.83% (area under the success rate curve) and 93.98% (area under prediction rate curve), and the highest robustness correspondingly. These attempts will play a significant role in ensemble modeling, in building reliable and comprehensive models. The proposed ANN-RF-LR ensemble model may be used in the other geographic areas having similar geo-environmental conditions. It may also be used in other types of geo-hazard modeling. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Graphical abstract

18 pages, 8175 KiB  
Article
Multilayer Perceptron-Based Phenological and Radiometric Normalization for High-Resolution Satellite Imagery
by Dae Kyo Seo and Yang Dam Eo
Appl. Sci. 2019, 9(21), 4543; https://0-doi-org.brum.beds.ac.uk/10.3390/app9214543 - 25 Oct 2019
Cited by 12 | Viewed by 2354
Abstract
Radiometric normalization is an essential preprocessing step that must be performed to detect changes in multi-temporal satellite images and, in general, relative radiometric normalization is utilized. However, most relative radiometric normalization methods assume a linear relationship and they cannot take into account nonlinear [...] Read more.
Radiometric normalization is an essential preprocessing step that must be performed to detect changes in multi-temporal satellite images and, in general, relative radiometric normalization is utilized. However, most relative radiometric normalization methods assume a linear relationship and they cannot take into account nonlinear properties, such as the distribution of the earth’s surface or phenological differences that are caused by the growth of vegetation. Thus, this paper proposes a novel method that assumes a nonlinear relationship and it uses a representative nonlinear regression model—multilayer perceptron (MLP). The proposed method performs radiometric resolution compression while considering both the complexity and time cost, and radiometric control set samples are extracted based on a no-change set method. Subsequently, the spectral index is selected for each band to compensate for the phenological properties, phenological normalization is performed based on MLP, and the global radiometric properties are adjusted through postprocessing. Finally, a performance evaluation is conducted by comparing the results herein with those from conventional relative radiometric normalization algorithms. The experimental results show that the proposed method outperforms conventional methods in terms of both visual inspection and quantitative evaluation. In other words, the applicability of the proposed method to the normalization of multi-temporal images with nonlinear properties is confirmed. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

16 pages, 5754 KiB  
Article
A High-Accuracy Model Average Ensemble of Convolutional Neural Networks for Classification of Cloud Image Patches on Small Datasets
by Van Hiep Phung and Eun Joo Rhee
Appl. Sci. 2019, 9(21), 4500; https://0-doi-org.brum.beds.ac.uk/10.3390/app9214500 - 23 Oct 2019
Cited by 120 | Viewed by 15738
Abstract
Research on clouds has an enormous influence on sky sciences and related applications, and cloud classification plays an essential role in it. Much research has been conducted which includes both traditional machine learning approaches and deep learning approaches. Compared with traditional machine learning [...] Read more.
Research on clouds has an enormous influence on sky sciences and related applications, and cloud classification plays an essential role in it. Much research has been conducted which includes both traditional machine learning approaches and deep learning approaches. Compared with traditional machine learning approaches, deep learning approaches achieved better results. However, most deep learning models need large data to train due to the large number of parameters. Therefore, they cannot get high accuracy in case of small datasets. In this paper, we propose a complete solution for high accuracy of classification of cloud image patches on small datasets. Firstly, we designed a suitable convolutional neural network (CNN) model for small datasets. Secondly, we applied regularization techniques to increase generalization and avoid overfitting of the model. Finally, we introduce a model average ensemble to reduce the variance of prediction and increase the classification accuracy. We experiment the proposed solution on the Singapore whole-sky imaging categories (SWIMCAT) dataset, which demonstrates perfect classification accuracy for most classes and confirms the robustness of the proposed model. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

20 pages, 4943 KiB  
Article
SEVUCAS: A Novel GIS-Based Machine Learning Software for Seismic Vulnerability Assessment
by Saro Lee, Mahdi Panahi, Hamid Reza Pourghasemi, Himan Shahabi, Mohsen Alizadeh, Ataollah Shirzadi, Khabat Khosravi, Assefa M. Melesse, Mohamad Yekrangnia, Fatemeh Rezaie, Hamidreza Moeini, Binh Thai Pham and Baharin Bin Ahmad
Appl. Sci. 2019, 9(17), 3495; https://0-doi-org.brum.beds.ac.uk/10.3390/app9173495 - 24 Aug 2019
Cited by 45 | Viewed by 5669
Abstract
Since it is not possible to determine the exact time of a natural disaster’s occurrence and the amount of physical and financial damage on humans or the environment resulting from their event, decision-makers need to identify areas with potential vulnerability in order to [...] Read more.
Since it is not possible to determine the exact time of a natural disaster’s occurrence and the amount of physical and financial damage on humans or the environment resulting from their event, decision-makers need to identify areas with potential vulnerability in order to reduce future losses. In this paper, a GIS-based open source software entitled Seismic-Related Vulnerability Calculation Software (SEVUCAS), based on the Step-wise Weight Assessment Ratio Analysis (SWARA) method and geographic information system, has been developed to assess seismic vulnerability by considering four groups of criteria (i.e., geotechnical, structural, socio-economic, and physical distance to needed facilities and away from dangerous facilities). The software was developed in C# language using ArcGIS Engine functions, which provide enhanced visualization as well as user-friendly and automatic software for the seismic vulnerability assessment of buildings. Weighting of the criteria (indicators) and alternatives (sub-indicators) was done using SWARA. Also, two interpolation methods based on a radial basis function (RBF) and teaching–learning-based optimization (TLBO) were used to optimize the weights of the criteria and the classes of each alternative as well. After weighing the criteria and alternatives, the weighted overlay analysis was used to determine the final vulnerability map in the form of contours and statistical data. The difference between this software and similar ones is that people with a low level of knowledge in the area of earthquake crisis management can use it to determine and estimate the seismic vulnerabilities of their houses. This visualized operational forecasting software provides an applicable tool for both government and people to make quick and correct decisions to determine higher priority structures for seismic retrofitting implementation. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

20 pages, 8234 KiB  
Article
An Improved Perceptual Hash Algorithm Based on U-Net for the Authentication of High-Resolution Remote Sensing Image
by Kaimeng Ding, Zedong Yang, Yingying Wang and Yueming Liu
Appl. Sci. 2019, 9(15), 2972; https://0-doi-org.brum.beds.ac.uk/10.3390/app9152972 - 25 Jul 2019
Cited by 15 | Viewed by 3340
Abstract
Data security technology is of great significance for the effective use of high-resolution remote sensing (HRRS) images in GIS field. Integrity authentication technology is an important technology to ensure the security of HRRS images. Traditional authentication technologies perform binary level authentication of the [...] Read more.
Data security technology is of great significance for the effective use of high-resolution remote sensing (HRRS) images in GIS field. Integrity authentication technology is an important technology to ensure the security of HRRS images. Traditional authentication technologies perform binary level authentication of the data and cannot meet the authentication requirements for HRRS images, while perceptual hashing can achieve perceptual content-based authentication. Compared with traditional algorithms, the existing edge-feature-based perceptual hash algorithms have already achieved high tampering authentication accuracy for the authentication of HRRS images. However, because of the traditional feature extraction methods they adopt, they lack autonomous learning ability, and their robustness still exists and needs to be improved. In this paper, we propose an improved perceptual hash scheme based on deep learning (DL) for the authentication of HRRS images. The proposed method consists of a modified U-net model to extract robust feature and a principal component analysis (PCA)-based encoder to generate perceptual hash values for HRRS images. In the training stage, a training sample generation method combining artificial processing and Canny operator is proposed to generate robust edge features samples. Moreover, to improve the performance of the network, exponential linear unit (ELU) and batch normalization (BN) are applied to extract more robust and accurate edge feature. The experiments have shown that the proposed algorithm has almost 100% robustness to format conversion between TIFF and BMP, LSB watermark embedding and lossless compression. Compared with the existing algorithms, the robustness of the proposed algorithm to lossy compression has been improved, with an average increase of 10%. What is more, the algorithm has good sensitivity to detect local subtle tampering to meet the high-accuracy requirements of authentication for HRRS images. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

19 pages, 6691 KiB  
Article
Extracting Crop Spatial Distribution from Gaofen 2 Imagery Using a Convolutional Neural Network
by Yan Chen, Chengming Zhang, Shouyi Wang, Jianping Li, Feng Li, Xiaoxia Yang, Yuanyuan Wang and Leikun Yin
Appl. Sci. 2019, 9(14), 2917; https://0-doi-org.brum.beds.ac.uk/10.3390/app9142917 - 22 Jul 2019
Cited by 14 | Viewed by 3040
Abstract
Using satellite remote sensing has become a mainstream approach for extracting crop spatial distribution. Making edges finer is a challenge, while simultaneously extracting crop spatial distribution information from high-resolution remote sensing images using a convolutional neural network (CNN). Based on the characteristics of [...] Read more.
Using satellite remote sensing has become a mainstream approach for extracting crop spatial distribution. Making edges finer is a challenge, while simultaneously extracting crop spatial distribution information from high-resolution remote sensing images using a convolutional neural network (CNN). Based on the characteristics of the crop area in the Gaofen 2 (GF-2) images, this paper proposes an improved CNN to extract fine crop areas. The CNN comprises a feature extractor and a classifier. The feature extractor employs a spectral feature extraction unit to generate spectral features, and five coding-decoding-pair units to generate five level features. A linear model is used to fuse features of different levels, and the fusion results are up-sampled to obtain a feature map consistent with the structure of the input image. This feature map is used by the classifier to perform pixel-by-pixel classification. In this study, the SegNet and RefineNet models and 21 GF-2 images of Feicheng County, Shandong Province, China, were chosen for comparison experiment. Our approach had an accuracy of 93.26%, which is higher than those of the existing SegNet (78.12%) and RefineNet (86.54%) models. This demonstrates the superiority of the proposed method in extracting crop spatial distribution information from GF-2 remote sensing images. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Graphical abstract

17 pages, 7014 KiB  
Article
A Process-Oriented Method for Tracking Rainstorms with a Time-Series of Raster Datasets
by Cunjin Xue, Jingyi Liu, Guanghui Yang and Chengbin Wu
Appl. Sci. 2019, 9(12), 2468; https://0-doi-org.brum.beds.ac.uk/10.3390/app9122468 - 17 Jun 2019
Cited by 7 | Viewed by 2536
Abstract
Extreme rainstorms have important socioeconomic consequences, but understanding their fine spatial structures and temporal evolution still remains challenging. In order to achieve this, in view of an evolutionary property of rainstorms, this paper designs a process-oriented algorithm for identifying and tracking rainstorms, named [...] Read more.
Extreme rainstorms have important socioeconomic consequences, but understanding their fine spatial structures and temporal evolution still remains challenging. In order to achieve this, in view of an evolutionary property of rainstorms, this paper designs a process-oriented algorithm for identifying and tracking rainstorms, named PoAIR. PoAIR uses time-series of raster datasets and consists of three steps. The first step combines an accumulated rainfall time-series and spatial connectivity to identify rainstorm objects at each time snapshot. Secondly, PoAIR adopts the geometrical features of eccentricity, rectangularity, roundness, and shape index, as well as the thematic feature of the mean rainstorm intensity, to match the same rainstorm objects in successive snapshots, and then tracks the same rainstorm objects during a rainstorm evolution sequence. In the third step, an evolutionary property of a rainstorm sequence is used to extrapolate its spatial location and geometrical features at the next time snapshot and reconstructs a rainstorm process by linking rainstorm sequences with an area-overlapping threshold. Experiments on simulated datasets demonstrate that PoAIR performs better than the Thunderstorm Identification, Tracking, Analysis and Nowcasting algorithm (TITAN) in both rainfall tracking and identifying the splitting, merging, and merging-splitting of rainstorm objects. Additionally, applications of PoAIR to Integrated Multi-satellitE Retrievals for Global Precipitation Mission (GPM/IMERG) final products covering mainland China show that PoAIR can effectively track rainstorm objects. Full article
(This article belongs to the Special Issue Machine Learning Techniques Applied to Geospatial Big Data)
Show Figures

Figure 1

Back to TopTop