File Download
Supplementary
-
Citations:
- Appears in Collections:
postgraduate thesis: Deep learning for pansharpening and classification of hyperspectral images
Title | Deep learning for pansharpening and classification of hyperspectral images |
---|---|
Authors | |
Advisors | |
Issue Date | 2024 |
Publisher | The University of Hong Kong (Pokfulam, Hong Kong) |
Citation | Guan, P. [管培岩]. (2024). Deep learning for pansharpening and classification of hyperspectral images. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. |
Abstract | Hyperspectral images (HSIs) are well known for containing abundant spectral information and have found numerous applications in various areas, such as environmental monitoring and remote sensing. Therefore, developing effective HSI processing and analysis techniques is crucial for these applications. Recently, deep learning has achieved great success in image processing and computer vision areas. Compared to traditional models based on hand-crafted features, deep networks rely on more relaxed priors and have much stronger learning capability for different levels of vision tasks, ranging from regression to segmentation, and to classification. The tremendous development of deep learning offers a promising direction to advance the field of hyperspectral vision. This thesis is in pursuit of exploring and proposing deep learning methods for HSI processing and analysis. The research areas range from low-level hyperspectral pansharpening task to high-level HSI classification task. In general, hyperspectral pansharpening involves the enhancement of spatial resolution of HSIs where local patterns matter, while HSI classification aims to predict the category of each pixel where high-level semantics are important.
In the first part of this thesis, we focus on the problem of hyperspectral pansharpening. How to achieve complete and accurate fusion of high-resolution panchromatic images and low-resolution hyperspectral images remains a critical challenge for models. To address this issue, we propose a multistage dual-attention guided fusion network (MDA-Net), which employs three streams to explore the intra-image characteristics and inter-image correlation simultaneously and merges them in multiple stages, where dual-attention is used to guide the fusion by identifying the useful spatial and spectral contents. Experimental results on both real and simulated datasets demonstrate the superiority of our network.
Second, we consider the problem of lacking labels in HSI classification. Deep networks require a large amount of labeled data for optimization. However, collecting sufficient labeled HSI samples is very time-consuming and expensive. To overcome this, we propose a self-supervised learning method, termed cross-domain contrastive learning (XDCL), which performs contrast between the spectral and spatial domains to learn semantic representations. We construct effective signals that represent the two domains, respectively, and demonstrate that semantics are shared across them while other contents tend not to be. Experimental results show that our XDCL learns powerful representations that boost the downstream classification task.
Third, we further enhance self-supervised learning in HSI classification. Despite great success in feature extraction without labels, self-supervised pre-training needs to be performed from scratch for each HSI, which is laborious and time-consuming. Moreover, its performance is sensitive to HSI quality. We propose to address these issues through transfer learning. We build a progressive self-supervised pre-training framework that performs sequential training on datasets that are increasingly similar to the target HSI. Specifically, it starts with a large general vision dataset and then moves to a related HSI dataset, acquiring a strong initialization for the final pre-training. Self-supervised elastic weight consolidation is also developed to mitigate catastrophic forgetting in continual pre-training. Apart from improving the convergence speed and representation quality significantly, our framework is applicable to various self-supervised learning algorithms. |
Degree | Doctor of Philosophy |
Subject | Hyperspectral imaging Deep learning (Machine learning) |
Dept/Program | Electrical and Electronic Engineering |
Persistent Identifier | http://hdl.handle.net/10722/343777 |
DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | Lam, EYM | - |
dc.contributor.advisor | Lee, W | - |
dc.contributor.author | Guan, Peiyan | - |
dc.contributor.author | 管培岩 | - |
dc.date.accessioned | 2024-06-06T01:04:55Z | - |
dc.date.available | 2024-06-06T01:04:55Z | - |
dc.date.issued | 2024 | - |
dc.identifier.citation | Guan, P. [管培岩]. (2024). Deep learning for pansharpening and classification of hyperspectral images. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. | - |
dc.identifier.uri | http://hdl.handle.net/10722/343777 | - |
dc.description.abstract | Hyperspectral images (HSIs) are well known for containing abundant spectral information and have found numerous applications in various areas, such as environmental monitoring and remote sensing. Therefore, developing effective HSI processing and analysis techniques is crucial for these applications. Recently, deep learning has achieved great success in image processing and computer vision areas. Compared to traditional models based on hand-crafted features, deep networks rely on more relaxed priors and have much stronger learning capability for different levels of vision tasks, ranging from regression to segmentation, and to classification. The tremendous development of deep learning offers a promising direction to advance the field of hyperspectral vision. This thesis is in pursuit of exploring and proposing deep learning methods for HSI processing and analysis. The research areas range from low-level hyperspectral pansharpening task to high-level HSI classification task. In general, hyperspectral pansharpening involves the enhancement of spatial resolution of HSIs where local patterns matter, while HSI classification aims to predict the category of each pixel where high-level semantics are important. In the first part of this thesis, we focus on the problem of hyperspectral pansharpening. How to achieve complete and accurate fusion of high-resolution panchromatic images and low-resolution hyperspectral images remains a critical challenge for models. To address this issue, we propose a multistage dual-attention guided fusion network (MDA-Net), which employs three streams to explore the intra-image characteristics and inter-image correlation simultaneously and merges them in multiple stages, where dual-attention is used to guide the fusion by identifying the useful spatial and spectral contents. Experimental results on both real and simulated datasets demonstrate the superiority of our network. Second, we consider the problem of lacking labels in HSI classification. Deep networks require a large amount of labeled data for optimization. However, collecting sufficient labeled HSI samples is very time-consuming and expensive. To overcome this, we propose a self-supervised learning method, termed cross-domain contrastive learning (XDCL), which performs contrast between the spectral and spatial domains to learn semantic representations. We construct effective signals that represent the two domains, respectively, and demonstrate that semantics are shared across them while other contents tend not to be. Experimental results show that our XDCL learns powerful representations that boost the downstream classification task. Third, we further enhance self-supervised learning in HSI classification. Despite great success in feature extraction without labels, self-supervised pre-training needs to be performed from scratch for each HSI, which is laborious and time-consuming. Moreover, its performance is sensitive to HSI quality. We propose to address these issues through transfer learning. We build a progressive self-supervised pre-training framework that performs sequential training on datasets that are increasingly similar to the target HSI. Specifically, it starts with a large general vision dataset and then moves to a related HSI dataset, acquiring a strong initialization for the final pre-training. Self-supervised elastic weight consolidation is also developed to mitigate catastrophic forgetting in continual pre-training. Apart from improving the convergence speed and representation quality significantly, our framework is applicable to various self-supervised learning algorithms. | - |
dc.language | eng | - |
dc.publisher | The University of Hong Kong (Pokfulam, Hong Kong) | - |
dc.relation.ispartof | HKU Theses Online (HKUTO) | - |
dc.rights | The author retains all proprietary rights, (such as patent rights) and the right to use in future works. | - |
dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
dc.subject.lcsh | Hyperspectral imaging | - |
dc.subject.lcsh | Deep learning (Machine learning) | - |
dc.title | Deep learning for pansharpening and classification of hyperspectral images | - |
dc.type | PG_Thesis | - |
dc.description.thesisname | Doctor of Philosophy | - |
dc.description.thesislevel | Doctoral | - |
dc.description.thesisdiscipline | Electrical and Electronic Engineering | - |
dc.description.nature | published_or_final_version | - |
dc.date.hkucongregation | 2024 | - |
dc.identifier.mmsid | 991044809210003414 | - |