File Download
Supplementary

postgraduate thesis: Deep learning for pansharpening and classification of hyperspectral images

TitleDeep learning for pansharpening and classification of hyperspectral images
Authors
Advisors
Advisor(s):Lam, EYMLee, W
Issue Date2024
PublisherThe University of Hong Kong (Pokfulam, Hong Kong)
Citation
Guan, P. [管培岩]. (2024). Deep learning for pansharpening and classification of hyperspectral images. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.
AbstractHyperspectral images (HSIs) are well known for containing abundant spectral information and have found numerous applications in various areas, such as environmental monitoring and remote sensing. Therefore, developing effective HSI processing and analysis techniques is crucial for these applications. Recently, deep learning has achieved great success in image processing and computer vision areas. Compared to traditional models based on hand-crafted features, deep networks rely on more relaxed priors and have much stronger learning capability for different levels of vision tasks, ranging from regression to segmentation, and to classification. The tremendous development of deep learning offers a promising direction to advance the field of hyperspectral vision. This thesis is in pursuit of exploring and proposing deep learning methods for HSI processing and analysis. The research areas range from low-level hyperspectral pansharpening task to high-level HSI classification task. In general, hyperspectral pansharpening involves the enhancement of spatial resolution of HSIs where local patterns matter, while HSI classification aims to predict the category of each pixel where high-level semantics are important. In the first part of this thesis, we focus on the problem of hyperspectral pansharpening. How to achieve complete and accurate fusion of high-resolution panchromatic images and low-resolution hyperspectral images remains a critical challenge for models. To address this issue, we propose a multistage dual-attention guided fusion network (MDA-Net), which employs three streams to explore the intra-image characteristics and inter-image correlation simultaneously and merges them in multiple stages, where dual-attention is used to guide the fusion by identifying the useful spatial and spectral contents. Experimental results on both real and simulated datasets demonstrate the superiority of our network. Second, we consider the problem of lacking labels in HSI classification. Deep networks require a large amount of labeled data for optimization. However, collecting sufficient labeled HSI samples is very time-consuming and expensive. To overcome this, we propose a self-supervised learning method, termed cross-domain contrastive learning (XDCL), which performs contrast between the spectral and spatial domains to learn semantic representations. We construct effective signals that represent the two domains, respectively, and demonstrate that semantics are shared across them while other contents tend not to be. Experimental results show that our XDCL learns powerful representations that boost the downstream classification task. Third, we further enhance self-supervised learning in HSI classification. Despite great success in feature extraction without labels, self-supervised pre-training needs to be performed from scratch for each HSI, which is laborious and time-consuming. Moreover, its performance is sensitive to HSI quality. We propose to address these issues through transfer learning. We build a progressive self-supervised pre-training framework that performs sequential training on datasets that are increasingly similar to the target HSI. Specifically, it starts with a large general vision dataset and then moves to a related HSI dataset, acquiring a strong initialization for the final pre-training. Self-supervised elastic weight consolidation is also developed to mitigate catastrophic forgetting in continual pre-training. Apart from improving the convergence speed and representation quality significantly, our framework is applicable to various self-supervised learning algorithms.
DegreeDoctor of Philosophy
SubjectHyperspectral imaging
Deep learning (Machine learning)
Dept/ProgramElectrical and Electronic Engineering
Persistent Identifierhttp://hdl.handle.net/10722/343777

 

DC FieldValueLanguage
dc.contributor.advisorLam, EYM-
dc.contributor.advisorLee, W-
dc.contributor.authorGuan, Peiyan-
dc.contributor.author管培岩-
dc.date.accessioned2024-06-06T01:04:55Z-
dc.date.available2024-06-06T01:04:55Z-
dc.date.issued2024-
dc.identifier.citationGuan, P. [管培岩]. (2024). Deep learning for pansharpening and classification of hyperspectral images. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.-
dc.identifier.urihttp://hdl.handle.net/10722/343777-
dc.description.abstractHyperspectral images (HSIs) are well known for containing abundant spectral information and have found numerous applications in various areas, such as environmental monitoring and remote sensing. Therefore, developing effective HSI processing and analysis techniques is crucial for these applications. Recently, deep learning has achieved great success in image processing and computer vision areas. Compared to traditional models based on hand-crafted features, deep networks rely on more relaxed priors and have much stronger learning capability for different levels of vision tasks, ranging from regression to segmentation, and to classification. The tremendous development of deep learning offers a promising direction to advance the field of hyperspectral vision. This thesis is in pursuit of exploring and proposing deep learning methods for HSI processing and analysis. The research areas range from low-level hyperspectral pansharpening task to high-level HSI classification task. In general, hyperspectral pansharpening involves the enhancement of spatial resolution of HSIs where local patterns matter, while HSI classification aims to predict the category of each pixel where high-level semantics are important. In the first part of this thesis, we focus on the problem of hyperspectral pansharpening. How to achieve complete and accurate fusion of high-resolution panchromatic images and low-resolution hyperspectral images remains a critical challenge for models. To address this issue, we propose a multistage dual-attention guided fusion network (MDA-Net), which employs three streams to explore the intra-image characteristics and inter-image correlation simultaneously and merges them in multiple stages, where dual-attention is used to guide the fusion by identifying the useful spatial and spectral contents. Experimental results on both real and simulated datasets demonstrate the superiority of our network. Second, we consider the problem of lacking labels in HSI classification. Deep networks require a large amount of labeled data for optimization. However, collecting sufficient labeled HSI samples is very time-consuming and expensive. To overcome this, we propose a self-supervised learning method, termed cross-domain contrastive learning (XDCL), which performs contrast between the spectral and spatial domains to learn semantic representations. We construct effective signals that represent the two domains, respectively, and demonstrate that semantics are shared across them while other contents tend not to be. Experimental results show that our XDCL learns powerful representations that boost the downstream classification task. Third, we further enhance self-supervised learning in HSI classification. Despite great success in feature extraction without labels, self-supervised pre-training needs to be performed from scratch for each HSI, which is laborious and time-consuming. Moreover, its performance is sensitive to HSI quality. We propose to address these issues through transfer learning. We build a progressive self-supervised pre-training framework that performs sequential training on datasets that are increasingly similar to the target HSI. Specifically, it starts with a large general vision dataset and then moves to a related HSI dataset, acquiring a strong initialization for the final pre-training. Self-supervised elastic weight consolidation is also developed to mitigate catastrophic forgetting in continual pre-training. Apart from improving the convergence speed and representation quality significantly, our framework is applicable to various self-supervised learning algorithms.-
dc.languageeng-
dc.publisherThe University of Hong Kong (Pokfulam, Hong Kong)-
dc.relation.ispartofHKU Theses Online (HKUTO)-
dc.rightsThe author retains all proprietary rights, (such as patent rights) and the right to use in future works.-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subject.lcshHyperspectral imaging-
dc.subject.lcshDeep learning (Machine learning)-
dc.titleDeep learning for pansharpening and classification of hyperspectral images-
dc.typePG_Thesis-
dc.description.thesisnameDoctor of Philosophy-
dc.description.thesislevelDoctoral-
dc.description.thesisdisciplineElectrical and Electronic Engineering-
dc.description.naturepublished_or_final_version-
dc.date.hkucongregation2024-
dc.identifier.mmsid991044809210003414-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats