Genetic Programming for High-Level Feature Learning in Crop Classification
<p>An example GP tree for NDVI.</p> "> Figure 2
<p>The workflow of GP for feature learning in crop classification.</p> "> Figure 3
<p>An example tree of the proposed GP method with the <span class="html-italic">COMB1</span> function. This tree can produce five features, and each feature is constructed by using arithmetic operators and input bands (terminals).</p> "> Figure 4
<p>The new GP tree representation with <span class="html-italic">COMB1</span> as a root node produces a predefined number of features by extending the width of the tree: using <span class="html-italic">COMB1</span> to produce five features (<b>a</b>); and using <span class="html-italic">COMB1</span> to produce <span class="html-italic">m</span> features (<b>b</b>). F<sub>i</sub> denotes a subtree that produces a feature using functions (e.g., arithmetic operators) and terminals.</p> "> Figure 5
<p>The new representation with the <span class="html-italic">COMB2</span> and <span class="html-italic">COMB3</span> functions produces a flexible/dynamic number of features by extending the depth of the tree. <span class="html-italic">COMB2</span> and <span class="html-italic">COMB3</span> are only used as the root node to produce two and three features (<b>a</b>); while they are used as the root and internal nodes to produce more features (<b>b</b>). F<sub>i</sub> denotes a subtree that produces a feature using functions (e.g., arithmetic operators) and terminals.</p> "> Figure 6
<p>Illustration of the five-fold cross-validation with a classification algorithm.</p> "> Figure 7
<p>Schematic of crossover (<b>a</b>); and mutation operations (<b>b</b>).</p> "> Figure 8
<p>The study area location (<b>a</b>) and crop samples (<b>b</b>).</p> "> Figure 9
<p>Classification results of KNN, NB, DT, and SVM by using original images (<b>a</b>–<b>d</b>); and GP learned features with the highest accuracies (<b>e</b>–<b>h</b>).</p> "> Figure 10
<p>Example of 10 features learned by GP.</p> "> Figure 11
<p>The frequency of the top 10 bands used in the GP methods learning the fixed and flexible numbers of features.</p> "> Figure 12
<p>Comparisons of classification accuracies of (<b>a</b>) KNN, (<b>b</b>) NB, (<b>c</b>) DT, and (<b>d</b>) SVM using features learned by GP and selected from VIs.</p> "> Figure 13
<p>The average accuracies and SDs of GP and VIs using different classifiers and feature numbers.</p> "> Figure 14
<p>The overall accuracies of the GP wrapped four classifiers (i.e., KNN, NB, DT and SVM), and MLP for various training subsets.</p> ">
Abstract
:1. Introduction
- Manual feature extraction is tedious and time-consuming. Features are usually extracted by trial and error, and the classification results tend to be subjective [6]. Features tend to vary for the same crop classification in the same study area for different studies.
- Most of the existing methods can only extract low-level features from the original image, such as spectral features, VIs, and phenological features. Low-level features might ignore useful time series information while including redundant information for crop classification. Complex factors, such as intra-class variability, inter-class similarity, light-scattering mechanisms, and atmospheric conditions, are difficult to be accounted for with only domain knowledge and human experience.
- Manual feature extraction approaches depend on the inter-regional and inter-annual variations of the crop calendar. Phenological characteristics of the same crops in different regions may differ because of various climate and geographic factors. Therefore, manually extracted features may be merely adequate for the specific study area and temporal pattern.
- (1)
- The new approach for feature learning based on GP represents, to the best of our knowledge, the first time that GP is used to learn multiple features for crop classification.
- (2)
- The new GP representation with COMB functions learns multiple high-level features based on a single tree for the classification of crops. The new GP representation automatically selects the discriminable low-level features from the raw images and then constructs a fixed or flexible number of high-level features based on the low-level ones.
2. Background on GP
3. GP-Based Feature Learning for Crop Classification
3.1. The New Tree Representation of GP
3.2. Function Set and Terminal Set
3.3. Fitness Function
3.4. Genetic Operators
4. Experiment Design
4.1. Study Area
4.2. Data Collection and Preprocessing
4.3. Crop Classification Experiments and Parameter Settings
4.4. Comparison Experiments
5. Results and Analysis
5.1. Crop Classification Results
5.2. Feature Visualization and Analysis
5.3. Comparisons with VIs
5.4. Comparisons with MLP
6. Discussion
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Gao, L.; Bryan, B.A. Finding pathways to national-scale land-sector sustainability. Nature 2017, 544, 217–222. [Google Scholar] [CrossRef] [PubMed]
- Cui, Z.; Zhang, H.; Chen, X.; Zhang, C.; Ma, W.; Huang, C.; Zhang, W.; Mi, G.; Miao, Y.; Li, X.; et al. Pursuing sustainable productivity with millions of smallholder farmers. Nature 2018, 555, 363–366. [Google Scholar] [CrossRef] [PubMed]
- Renard, D.; Tilman, D. National food production stabilized by crop diversity. Nature 2019, 571, 257–260. [Google Scholar] [CrossRef] [PubMed]
- Veloso, A.; Mermoz, S.; Bouvet, A.; Le Toan, T.; Planells, M.; Dejoux, J.-F.; Ceschia, E. Understanding the temporal behavior of crops using Sentinel-1 and Sentinel-2-like data for agricultural applications. Remote Sens. Environ. 2017, 199, 415–426. [Google Scholar] [CrossRef]
- Johnson, D.M. Using the Landsat archive to map crop cover history across the United States. Remote Sens. Environ. 2019, 232, 111286. [Google Scholar] [CrossRef]
- Zhong, L.; Hu, L.; Zhou, H. Deep learning based multi-temporal crop classification. Remote Sens. Environ. 2019, 221, 430–443. [Google Scholar] [CrossRef]
- Pôças, I.; Calera, A.; Campos, I.; Cunha, M. Remote sensing for estimating and mapping single and basal crop coefficientes: A review on spectral vegetation indices approaches. Agric. Water Manag. 2020, 233, 106081. [Google Scholar] [CrossRef]
- Xue, J.; Su, B. Significant Remote Sensing Vegetation Indices: A Review of Developments and Applications. J. Sens. 2017, 2017, 1353691. [Google Scholar] [CrossRef]
- Huete, A.; Didan, K.; Miura, T.; Rodriguez, E.P.; Gao, X.; Ferreira, L.G. Overview of the radiometric and biophysical performance of the MODIS vegetation indices. Remote Sens. Environ. 2002, 83, 195–213. [Google Scholar] [CrossRef]
- Zhang, G.; Xiao, X.; Biradar, C.M.; Dong, J.; Qin, Y.; Menarguez, M.A.; Zhou, Y.; Zhang, Y.; Jin, C.; Wang, J.; et al. Spatiotemporal patterns of paddy rice croplands in China and India from 2000 to 2015. Sci. Total Environ. 2017, 579, 82–92. [Google Scholar] [CrossRef]
- Zhang, G.; Xiao, X.; Dong, J.; Xin, F.; Zhang, Y.; Qin, Y.; Doughty, R.B.; Moore, B. Fingerprint of rice paddies in spatial–temporal dynamics of atmospheric methane concentration in monsoon Asia. Nat. Commun. 2020, 11, 554. [Google Scholar] [CrossRef]
- Chang, J.; Hansen, M.C.; Pittman, K.; Carroll, M.; DiMiceli, C. Corn and Soybean Mapping in the United States Using MODIS Time-Series Data Sets. Agron. J. 2007, 99, 1654–1664. [Google Scholar] [CrossRef]
- Song, X.-P.; Potapov, P.V.; Krylov, A.; King, L.; Di Bella, C.M.; Hudson, A.; Khan, A.; Adusei, B.; Stehman, S.V.; Hansen, M.C. National-scale soybean mapping and area estimation in the United States using medium resolution satellite imagery and field survey. Remote Sens. Environ. 2017, 190, 383–395. [Google Scholar] [CrossRef]
- Bargiel, D. A new method for crop classification combining time series of radar images and crop phenology information. Remote Sens. Environ. 2017, 198, 369–383. [Google Scholar] [CrossRef]
- Wang, S.; Azzari, G.; Lobell, D.B. Crop type mapping without field-level labels: Random forest transfer and unsupervised clustering techniques. Remote Sens. Environ. 2019, 222, 303–317. [Google Scholar] [CrossRef]
- Zhong, L.; Gong, P.; Biging, G.S. Efficient corn and soybean mapping with temporal extendability: A multi-year experiment using Landsat imagery. Remote Sens. Environ. 2014, 140, 1–13. [Google Scholar] [CrossRef]
- Bi, Y.; Xue, B.; Zhang, M. An Effective Feature Learning Approach Using Genetic Programming With Image Descriptors for Image Classification [Research Frontier]. IEEE Comput. Intell. Mag. 2020, 15, 65–77. [Google Scholar] [CrossRef]
- Zhang, L.; Zhang, L.; Du, B. Deep Learning for Remote Sensing Data: A Technical Tutorial on the State of the Art. IEEE Geosci. Remote Sens. Mag. 2016, 4, 22–40. [Google Scholar] [CrossRef]
- Zhang, C.; Sargent, I.; Pan, X.; Li, H.; Gardiner, A.; Hare, J.; Atkinson, P.M. Joint Deep Learning for land cover and land use classification. Remote Sens. Environ. 2019, 221, 173–187. [Google Scholar] [CrossRef]
- Eiben, A.E.; Smith, J. From evolutionary computation to the evolution of things. Nature 2015, 521, 476–482. [Google Scholar] [CrossRef]
- Koza, J.R. Genetic Programming: On the Programming of Computers by Means of Natural Selection; MIT Press: Cambridge, MA, USA, 1992. [Google Scholar]
- Tran, B.; Xue, B.; Zhang, M. Genetic programming for multiple-feature construction on high-dimensional classification. Pattern Recognit. 2019, 93, 404–417. [Google Scholar] [CrossRef]
- Ain, Q.U.; Al-Sahaf, H.; Xue, B.; Zhang, M. Generating Knowledge-Guided Discriminative Features Using Genetic Programming for Melanoma Detection. IEEE Trans. Emerg. Top. Comput. Intell. 2021, 5, 554–569. [Google Scholar] [CrossRef]
- Liang, Y.; Zhang, M.; Browne, W.N. Figure-ground image segmentation using feature-based multi-objective genetic programming techniques. Neural Comput. Appl. 2019, 31, 3075–3094. [Google Scholar] [CrossRef]
- Puente, C.; Olague, G.; Trabucchi, M.; Arjona-Villicaña, P.D.; Soubervielle-Montalvo, C. Synthesis of Vegetation Indices Using Genetic Programming for Soil Erosion Estimation. Remote Sens. 2019, 11, 156. [Google Scholar] [CrossRef]
- Cabral, A.I.R.; Silva, S.; Silva, P.C.; Vanneschi, L.; Vasconcelos, M.J. Burned area estimations derived from Landsat ETM+ and OLI data: Comparing Genetic Programming with Maximum Likelihood and Classification and Regression Trees. ISPRS J. Photogramm. Remote Sens. 2018, 142, 94–105. [Google Scholar] [CrossRef]
- Batista, J.E.; Cabral, A.I.; Vasconcelos, M.J.; Vanneschi, L.; Silva, S. Improving Land Cover Classification Using Genetic Programming for Feature Construction. Remote Sens. 2021, 13, 1623. [Google Scholar] [CrossRef]
- Tan, Y.; He, J.; Yue, W.; Zhang, L.; Wang, Q.R. Spatial pattern change of the cultivated land before and after the second national land survey in China. J. Nat. Resour. 2017, 32, 186–197. [Google Scholar] [CrossRef]
- Hu, Q.; Wu, W.; Song, Q.; Yu, Q.; Lu, M.; Yang, P.; Tang, H.; Long, Y. Extending the Pairwise Separability Index for Multicrop Identification Using Time-Series MODIS Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6349–6361. [Google Scholar] [CrossRef]
- Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
- Löw, F.; Michel, U.; Dech, S.; Conrad, C. Impact of feature selection on the accuracy and spatial uncertainty of per-field crop classification using Support Vector Machines. ISPRS J. Photogramm. Remote Sens. 2013, 85, 102–119. [Google Scholar] [CrossRef]
- Gitelson, A.A.; Viña, A.; Ciganda, V.; Rundquist, D.C.; Arkebauer, T.J. Remote estimation of canopy chlorophyll content in crops. Geophys. Res. Lett. 2005, 32, L08403. [Google Scholar] [CrossRef]
- Espejo, P.G.; Ventura, S.; Herrera, F. A Survey on the Application of Genetic Programming to Classification. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2010, 40, 121–144. [Google Scholar] [CrossRef]
Function | Type | No. Arguments | Descriptions |
---|---|---|---|
COMB1 | Root | m (Predefined) | Concatenate multiple inputs into a feature vector |
COMB2 | Root/Internal | Two | Concatenate two inputs into a feature vector |
COMB3 | Root/Internal | Three | Concatenate three inputs into a feature vector |
Add (+) | Internal | Two | Add two arguments |
Sub (−) | Internal | Two | Subtract two arguments |
Mul (×) | Internal | Two | Multiply two arguments |
ProD (/) | Internal | Two | Protected division. Divide the second argument using the first argument. Return one if the divisor is zero. |
IF | Internal | Three | If the first argument is greater than 0, the second argument is returned. Otherwise, the third argument is returned. |
KNN (%) | NB (%) | DT (%) | SVM (%) | ||
---|---|---|---|---|---|
Original image | 89.72 | 92.50 | 94.03 | 97.22 | |
GP for fixed number of features | 5 | 96.81↑ | 96.81↑ | 96.11↑ | 96.67↓ |
10 | 97.36↑ | 96.67↑ | 94.86↑ | 97.64↑ | |
15 | 96.81↑ | 96.94↑ | 95.69↑ | 97.36↑ | |
20 | 96.94↑ | 96.53↑ | 95.28↑ | 97.50↑ | |
25 | 96.11↑ | 96.81↑ | 95.42↑ | 97.64↑ | |
30 | 97.22↑ | 97.08↑ | 95.00 ↑ | 97.78↑ | |
GP for flexible number of features | 97.78↑ (56) | 96.67↑ (23) | 95.14↑ (32) | 97.22 = (65) |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lu, M.; Bi, Y.; Xue, B.; Hu, Q.; Zhang, M.; Wei, Y.; Yang, P.; Wu, W. Genetic Programming for High-Level Feature Learning in Crop Classification. Remote Sens. 2022, 14, 3982. https://doi.org/10.3390/rs14163982
Lu M, Bi Y, Xue B, Hu Q, Zhang M, Wei Y, Yang P, Wu W. Genetic Programming for High-Level Feature Learning in Crop Classification. Remote Sensing. 2022; 14(16):3982. https://doi.org/10.3390/rs14163982
Chicago/Turabian StyleLu, Miao, Ying Bi, Bing Xue, Qiong Hu, Mengjie Zhang, Yanbing Wei, Peng Yang, and Wenbin Wu. 2022. "Genetic Programming for High-Level Feature Learning in Crop Classification" Remote Sensing 14, no. 16: 3982. https://doi.org/10.3390/rs14163982