File Download
Supplementary
-
Citations:
- Appears in Collections:
postgraduate thesis: Context and adversarial learning in low-level vision
Title | Context and adversarial learning in low-level vision |
---|---|
Authors | |
Advisors | Advisor(s):Yu, Y |
Issue Date | 2020 |
Publisher | The University of Hong Kong (Pokfulam, Hong Kong) |
Citation | Li, H. [李灝峰]. (2020). Context and adversarial learning in low-level vision. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. |
Abstract | In recent years, with increasing network bandwidth and the popularity of portable cameras, massive video blogs and photographs have been produced and uploaded to webcam sites and social media. It is important to develop automated, intelligent and robust tools for digital image processing. How to harvest and exploit spatio-temporal contexts in visual data to achieve high-quality image processing is also an essential problem. On one hand, adversarial learning based attacks could threaten deep learning based models. On the other hand, generative adversarial networks can synthesize realistic image patches via adversarial learning. It is necessary to improve existing image processing tools with generative adversarial networks and against adversarial attacks. Thus, we propose three solutions based on context and adversarial learning for three low-level vision tasks: video salient object detection, robust salient object detection against adversarial attacks and image inpainting.
For detecting salient objects in a video, we introduce a motion guided attention neural network that is integrated with a family of motion guided attention modules. The proposed motion guided attention modules model how motions computed from temporal contexts determine object saliency. The proposed neural network contains a motion branch and an appearance branch, which work by computing intermediate representations and a final saliency map for a pair of optical flow and color images. Our proposed algorithm significantly surpasses the state-of-the-art video salient object detection methods.
For robust salient object detection resistant to adversarial samples, we develop a novel generic framework that boosts the robustness of arbitrary salient object detection models based on fully convolutional networks. The proposed salient object detection framework adopts a segment-wise shielding component to destroy adversarial perturbations in an input image. The framework includes a context-ware restoration component that uses spatial contextual correlations to refine saliency maps. Experimental results demonstrate that our proposed salient object detection framework performs better than other defenses against adversarial attacks.
For semantic image completion, we propose a context-aware semantic inpainting algorithm and two novel evaluation metrics. The proposed algorithm consists of a fully convolutional generative network and a context-aware joint loss function. The generative network resorts to fully convolutional architecture without fully connected bottlenecks to maintain structural features. The joint loss function forces the proposed network to synthesize contents with the same semantic as the surrounding context. Both quantitative and qualitative comparisons suggest that our proposed algorithm obtains state-of-the-art performance. The proposed metrics can better rate an over-smoothed result and the overall semantic of an image. |
Degree | Doctor of Philosophy |
Subject | Image processing - Digital techniques |
Dept/Program | Computer Science |
Persistent Identifier | http://hdl.handle.net/10722/283128 |
DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | Yu, Y | - |
dc.contributor.author | Li, Haofeng | - |
dc.contributor.author | 李灝峰 | - |
dc.date.accessioned | 2020-06-10T01:02:15Z | - |
dc.date.available | 2020-06-10T01:02:15Z | - |
dc.date.issued | 2020 | - |
dc.identifier.citation | Li, H. [李灝峰]. (2020). Context and adversarial learning in low-level vision. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. | - |
dc.identifier.uri | http://hdl.handle.net/10722/283128 | - |
dc.description.abstract | In recent years, with increasing network bandwidth and the popularity of portable cameras, massive video blogs and photographs have been produced and uploaded to webcam sites and social media. It is important to develop automated, intelligent and robust tools for digital image processing. How to harvest and exploit spatio-temporal contexts in visual data to achieve high-quality image processing is also an essential problem. On one hand, adversarial learning based attacks could threaten deep learning based models. On the other hand, generative adversarial networks can synthesize realistic image patches via adversarial learning. It is necessary to improve existing image processing tools with generative adversarial networks and against adversarial attacks. Thus, we propose three solutions based on context and adversarial learning for three low-level vision tasks: video salient object detection, robust salient object detection against adversarial attacks and image inpainting. For detecting salient objects in a video, we introduce a motion guided attention neural network that is integrated with a family of motion guided attention modules. The proposed motion guided attention modules model how motions computed from temporal contexts determine object saliency. The proposed neural network contains a motion branch and an appearance branch, which work by computing intermediate representations and a final saliency map for a pair of optical flow and color images. Our proposed algorithm significantly surpasses the state-of-the-art video salient object detection methods. For robust salient object detection resistant to adversarial samples, we develop a novel generic framework that boosts the robustness of arbitrary salient object detection models based on fully convolutional networks. The proposed salient object detection framework adopts a segment-wise shielding component to destroy adversarial perturbations in an input image. The framework includes a context-ware restoration component that uses spatial contextual correlations to refine saliency maps. Experimental results demonstrate that our proposed salient object detection framework performs better than other defenses against adversarial attacks. For semantic image completion, we propose a context-aware semantic inpainting algorithm and two novel evaluation metrics. The proposed algorithm consists of a fully convolutional generative network and a context-aware joint loss function. The generative network resorts to fully convolutional architecture without fully connected bottlenecks to maintain structural features. The joint loss function forces the proposed network to synthesize contents with the same semantic as the surrounding context. Both quantitative and qualitative comparisons suggest that our proposed algorithm obtains state-of-the-art performance. The proposed metrics can better rate an over-smoothed result and the overall semantic of an image. | - |
dc.language | eng | - |
dc.publisher | The University of Hong Kong (Pokfulam, Hong Kong) | - |
dc.relation.ispartof | HKU Theses Online (HKUTO) | - |
dc.rights | The author retains all proprietary rights, (such as patent rights) and the right to use in future works. | - |
dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
dc.subject.lcsh | Image processing - Digital techniques | - |
dc.title | Context and adversarial learning in low-level vision | - |
dc.type | PG_Thesis | - |
dc.description.thesisname | Doctor of Philosophy | - |
dc.description.thesislevel | Doctoral | - |
dc.description.thesisdiscipline | Computer Science | - |
dc.description.nature | published_or_final_version | - |
dc.date.hkucongregation | 2020 | - |
dc.identifier.mmsid | 991044242097003414 | - |