Hyperspectral Image Classification Based on 3D Asymmetric Inception Network with Data Fusion Transfer Learning
Saved in:
| Title: | Hyperspectral Image Classification Based on 3D Asymmetric Inception Network with Data Fusion Transfer Learning |
|---|---|
| Authors: | Bei Fang, Yu Liu, Haokui Zhang, Juhou He |
| Source: | Remote Sensing, Vol 14, Iss 1711, p 1711 (2022) |
| Publisher Information: | MDPI AG |
| Publication Year: | 2022 |
| Collection: | Directory of Open Access Journals: DOAJ Articles |
| Subject Terms: | hyperspectral image classification, convolutional neural network, light-weight network, 3D asymmetric inception network, transfer learning, Science |
| Description: | Hyperspectral image (HSI) classification has been marked by exceptional progress in recent years. Much of this progess has come from advances in convolutional neural networks (CNNs). Different from the RGB images, HSI images are captured by various remote sensors with different spectral configurations. Moreover, each HSI dataset only contains very limited training samples and thus the model is prone to overfitting when using deep CNNs. In this paper, we first propose a 3D asymmetric inception network, AINet, to overcome the overfitting problem. With the emphasis on spectral signatures over spatial contexts of HSI data, the 3D convolution layer of AINet is replaced with two asymmetric inception units, i.e., a space inception unit and spectrum inception unit, to convey and classify the features effectively. In addition, we exploited a data-fusion transfer learning strategy to improve model initialization and classification performance. Extensive experiments show that the proposed approach outperforms all of the state-of-the-art methods via several HSI benchmarks, including Pavia University, Indian Pines and Kennedy Space Center (KSC). |
| Document Type: | article in journal/newspaper |
| Language: | English |
| Relation: | https://www.mdpi.com/2072-4292/14/7/1711; https://doaj.org/toc/2072-4292; https://doaj.org/article/43f2fbfd6beb4926b762d83ceb587720 |
| DOI: | 10.3390/rs14071711 |
| Availability: | https://doi.org/10.3390/rs14071711 https://doaj.org/article/43f2fbfd6beb4926b762d83ceb587720 |
| Accession Number: | edsbas.6A72A7F4 |
| Database: | BASE |
| Abstract: | Hyperspectral image (HSI) classification has been marked by exceptional progress in recent years. Much of this progess has come from advances in convolutional neural networks (CNNs). Different from the RGB images, HSI images are captured by various remote sensors with different spectral configurations. Moreover, each HSI dataset only contains very limited training samples and thus the model is prone to overfitting when using deep CNNs. In this paper, we first propose a 3D asymmetric inception network, AINet, to overcome the overfitting problem. With the emphasis on spectral signatures over spatial contexts of HSI data, the 3D convolution layer of AINet is replaced with two asymmetric inception units, i.e., a space inception unit and spectrum inception unit, to convey and classify the features effectively. In addition, we exploited a data-fusion transfer learning strategy to improve model initialization and classification performance. Extensive experiments show that the proposed approach outperforms all of the state-of-the-art methods via several HSI benchmarks, including Pavia University, Indian Pines and Kennedy Space Center (KSC). |
|---|---|
| DOI: | 10.3390/rs14071711 |
Nájsť tento článok vo Web of Science