Open Access
6 January 2022 Image-based neural architecture automatic search method for hyperspectral image classification
Zhonggang Hu, Wenxing Bao, Kewen Qu, Hongbo Liang
Author Affiliations +
Fig. 1
Feature extraction of HSI: (a) a 3×3 neighboring block, (b) a 5×5 neighboring block, and (c) a masked image.

Fig. 2
Abstract description of NAS method. The search strategy selects an architecture from a defined search space. The architecture is passed to the performance evaluation strategy, which returns the estimated performance of the architecture to the search strategy. Architecture D is an optional architecture in the search space.

Fig. 3
HSI data preprocessing process, where X1, X2, and X3 are the training, validation, and test datasets, respectively. I1, I2, and I3 are the training, validation, and test index sets, respectively.

Fig. 4
I-NAS for the HSIC instances. The model can take an input of any size and produce an output with a corresponding size. In training, the predicted label is selected in the output according to the position of the training pixel. The cross-entropy represents the gradient of the cross-entropy objective function [Eq. (1)]. Architecture D is an optional architecture in the search space.

Fig. 5
CNN framework of cells stacking for HSI classification.

Fig. 6
Cell structure search process diagram.

Algorithm 1
I-NAS for HSI classification.
Require: Initialize number of training group X1, validation group X2, and operation set O.
 1: for each pixel do
 2:  Create the index according to each class of label.
 3: end for
 4: split the sample set carrying position information into training, validation and test dataset according to X1 and X2.
 5: Mask the training, validation and test dataset respectively.
 6: Architecture search phase:
 7: initialize search epochs, learning rate ξ and ε, the architecture variable B, and the CNN weight w.
 8: for every search epoch do
 9:  wwξwLtrain(w,B);
10:  BBεBLval(wξwLtrain(w,B),B).
11: end for
12: choose the best B* according to the performance on validation dataset.
13: according to o(n,m)=argmaxoOBo*(n,m), acquire the best I-NAS architecture Ar.
14: Train and test the optimal I-NAS:
15: initialize training epochs, the weight w* of Ar and learning rate ξ.
16: for every training epoch do
17:  w*w*ξw*Ltrain(w*).
18: end for
19: for every test epoch do
20:  according to test dataset carrying position information, acquire the predict by I-NAS.
21: end for
22: obtain OA, AA, Kappa by evaluating the predict and test labels.

Fig. 7
IN dataset. (a) False-color image. (b) Ground truth map. (c) Class name and color code and the corresponding number of classes.

Fig. 8
UP dataset. (a) False-color image. (b) Ground truth map. (c) Class name and color code and the corresponding number of classes.

Fig. 9
The optimal structure of a cell based on the I-NAS method on IN.

Fig. 10
Optimized structure of another cell based on the I-NAS method on IN.

Fig. 11
The optimal structure of a cell based on I-NAS method on UP.

Fig. 12
Optimized structure of another cell based on I-NAS method on UP.

Table 1
Classification results of different methods for IN dataset.
ClassSVM3DCNNSSRNADGANCapsNetP-NASI-NAS
167.10 ± 1.6789.80 ± 1.2371.08 ± 2.4898.04 ± 2.9968.50 ± 9.25100 ± 0.00100 ± 0.00
270.11 ± 2.5890.00 ± 2.2294.67 ± 4.8197.72 ± 2.1893.94 ± 2.7886.52 ± 3.5989.76 ± 4.42
363.09 ± 5.8186.82 ± 1.7188.17 ± 6.1895.55 ± 3.5788.77 ± 5.1196.97 ± 1.5896.85 ± 1.81
444.02 ± 4.2794.18 ± 1.9399.13 ± 9.4996.02 ± 2.6771.03 ± 9.3399.87 ± 0.2598.82 ± 2.07
585.97 ± 2.3792.20 ± 1.0692.61 ± 7.2596.43 ± 3.1387.05 ± 3.2996.41 ± 2.3595.78 ± 1.72
692.25 ± 1.3599.50 ± 1.3297.03 ± 5.5697.76 ± 2.8998.83 ± 1.2698.47 ± 1.0199.32 ± 0.55
781.81 ± 1.6682.75 ± 2.3050.68 ± 4.3969.23 ± 9.6572.64 ± 9.61100 ± 0.00100 ± 0.00
898.47 ± 1.0399.61 ± 0.9798.55 ± 4.1099.54 ± 0.7899.85 ± 0.42100 ± 0.0099.95 ± 0.15
950.11 ± 5.7499.20 ± 0.9768.49 ± 3.7071.78 ± 9.7156.66 ± 9.62100 ± 0.00100 ± 0.00
1068.38 ± 1.9787.71 ± 5.1294.58 ± 1.0595.73 ± 1.6494.71 ± 3.0192.57 ± 1.1196.19 ± 1.62
1182.22 ± 2.5991.19 ± 0.9289.55 ± 5.2696.90 ± 1.8497.25 ± 1.3491.83 ± 3.2090.49 ± 4.09
1268.26 ± 2.2993.10 ± 1.1574.24 ± 9.7297.14 ± 2.3387.01 ± 5.1594.24 ± 1.5595.32 ± 1.97
1394.03 ± 2.9299.58 ± 1.3299.69 ± 9.7196.93 ± 5.9996.94 ± 4.58100 ± 0.0099.84 ± 0.30
1494.95 ± 0.8196.22 ± 0.2798.02 ± 7.3699.76 ± 0.3198.25 ± 1.5899.42 ± 0.3498.24 ± 2.32
1557.62 ± 5.2975.21 ± 2.1191.75 ± 9.4097.66 ± 1.9588.12 ± 8.8699.74 ± 0.5299.12 ± 1.07
1683.68 ± 6.2299.00 ± 2.6794.36 ± 9.3984.02 ± 8.0591.13 ± 7.5797.77 ± 2.72100 ± 0.00
OA (%)76.83 ± 1.1991.68 ± 1.4490.74 ± 3.4987.97 ± 6.5694.39 ± 0.6394.25 ± 0.3994.64 ± 1.23
AA (%)75.08 ± 1.4992.31 ± 0.5787.66 ± 5.6893.19 ± 2.7285.91 ± 2.9797.11 ± 0.3297.49 ± 0.47
K×10073.69 ± 1.3189.40 ± 0.9589.39 ± 4.0296.76 ± 0.9293.47 ± 0.7493.39 ± 0.4593.83 ± 1.41

Table 2
Classification results of different methods for UP dataset.
ClassSVM3DCNNSSRNADGANCapsNetP-NASI-NAS
195.38 ± 1.3786.20 ± 3.2388.88 ± 7.3689.59 ± 8.6095.80 ± 0.8195.96 ± 3.1096.66 ± 1.88
294.60 ± 0.8893.11 ± 2.1192.21 ± 1.7391.44 ± 9.6599.83 ± 0.2495.91 ± 1.6896.86 ± 1.46
365.34 ± 4.2763.09 ± 6.5492.79 ± 5.9995.95 ± 3.4768.74 ± 7.5999.22 ± 0.7398.28 ± 1.54
477.73 ± 7.3695.81 ± 1.7796.19 ± 0.8288.65 ± 6.2596.75 ± 1.4994.41 ± 2.0697.44 ± 0.54
594.57 ± 2.8394.14 ± 4.7897.76 ± 3.0597.73 ± 2.2999.90 ± 0.1899.49 ± 0.43100 ± 0.00
667.55 ± 4.4293.06 ± 1.9997.03 ± 0.3691.07 ± 9.4999.49 ± 0.6199.32 ± 0.5597.34 ± 1.09
761.03 ± 6.4257.81 ± 5.4468.23 ± 9.5593.56 ± 4.8777.27 ± 4.7399.97 ± 0.0799.91 ± 0.13
878.50 ± 5.9676.10 ± 2.8690.16 ± 3.7772.77 ± 9.6687.03 ± 3.2397.45 ± 1.1099.25 ± 0.40
999.89 ± 0.0683.20 ± 4.92100 ± 0.0073.05 ± 9.6795.88 ± 4.1195.02 ± 3.6999.92 ± 0.05
OA (%)84.45 ± 1.8887.32 ± 1.8991.20 ± 1.5881.25 ± 9.3395.44 ± 0.8196.71 ± 0.7197.45 ± 0.75
AA (%)81.62 ± 1.9882.51 ± 2.1791.58 ± 1.6088.20 ± 6.0691.31 ± 1.6897.42 ± 0.8398.41 ± 0.36
K×10079.87 ± 2.2583.56 ± 2.7188.17 ± 2.1787.53 ± 9.1494.16 ± 1.0495.65 ± 0.9396.62 ± 0.98

Fig. 13
Classification maps obtained by all considered algorithms on IN dataset. (a) Ground truth, (b) SVM, (c) 3DCNN, (d) SSRN, (e) ADGAN, (f) CapsNet, (g) P-NAS, and (h) I-NAS.

Fig. 14
Classification maps obtained by all considered algorithms on UP dataset. (a) Ground truth, (b) SVM, (c) 3DCNN, (d) SSRN, (e) ADGAN, (f) CapsNet, (g) P-NAS, and (h) I-NAS.

Fig. 15
Impact of different number of training samples on OA results for training. OA results were obtained by all algorithms on (a) IN dataset and (b) UP dataset.

Fig. 16
CA results for each class with different number of total labeled samples for training over the I-NAS on (a) IN dataset and (b) UP dataset.

Table 3
Time consumption of different algorithms on the IN dataset.
MethodTraining time (s)Testing time (s)
SVM0.15 ± 0.013.18 ± 0.08
DCNN1502.29 ± 6.0357.68 ± 3.98
SSRN3368.62 ± 51.0547.58 ± 5.97
ADGAN752.49 ± 4.923.24 ± 0.10
CapsNet1293.09 ± 43.2614.77 ± 0.96
P-NAS1670.81 ± 7.94102.53 ± 2.34
I-NAS153.74 ± 4.380.33 ± 0.04

Table 4
Time consumption of different algorithms on the UP dataset.
MethodTraining time (s)Testing time (s)
SVM0.09 ± 0.015.96 ± 0.48
DCNN458.29 ± 3.11246.78 ± 57.58
SSRN1358.58 ± 70.6982.25 ± 1.39
ADGAN1227.62 ± 4.7714.61 ± 0.05
CapsNet871.47 ± 38.7478.99 ± 1.57
P-NAS1145.18 ± 15.68319.97 ± 2.79
I-NAS427.60 ± 6.681.20 ± 0.18

Table 5
Total number of trainable parameters in different algorithms (MB).
MethodDatasets
Indian PinesPavia University
SSRN0.3301080.189316
CapsNet0.2623290.174164
P-NAS0.0670120.073683
I-NAS0.0827240.089795

Table 6
Time consumption and number of parameters for P-NAS and I-NAS on the IN dataset within the architecture search phase.
MethodArchitecture search time (s)Number of parameters (MB)
P-NAS3090.350.288820
I-NAS453.850.188084

Table 7
Time consumption and number of parameters for P-NAS and I-NAS on the UP dataset within the architecture search phase.
MethodArchitecture search time (s)Number of parameters (MB)
P-NAS2453.180.286947
I-NAS1262.300.185987

Lens.org Logo
CITATIONS
Cited by 3 scholarly publications and 1 patent.
Advertisement
Advertisement
KEYWORDS
Image classification

Hyperspectral imaging

Convolution

Image processing

Neural networks

Feature extraction

Statistical modeling

Back to Top