File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Cyclemlp: A MLP-like architecture for dense prediction
Title | Cyclemlp: A MLP-like architecture for dense prediction |
---|---|
Authors | |
Issue Date | 2022 |
Publisher | IEEE. |
Citation | 10th International Conference on Learning Representations (ICLR) (Virtual), 25-29 April, 2022 How to Cite? |
Abstract | This paper presents a simple MLP-like architecture, CycleMLP, which is a versatile backbone for visual recognition and dense predictions. As compared to modern MLP architectures, e.g. , MLP-Mixer (Tolstikhin et al., 2021), ResMLP (Touvron et al., 2021a), and gMLP (Liu et al., 2021a), whose architectures are correlated to image size and thus are infeasible in object detection and segmentation, CycleMLP has two advantages compared to modern approaches. (1) It can cope with various image sizes. (2) It achieves linear computational complexity to image size by using local windows. In contrast, previous MLPs have O(N2) computations due to fully spatial connections. We build a family of models which surpass existing MLPs and even state-of-the-art Transformer-based models, e.g. Swin Transformer (Liu et al., 2021b), while using fewer parameters and FLOPs. We expand the MLP-like models’ applicability, making them a versatile backbone for dense prediction tasks. CycleMLP achieves competitive results on object detection, instance segmentation, and semantic segmentation. In particular, CycleMLP-Tiny outperforms Swin-Tiny by 1.3% mIoU on ADE20K dataset with fewer FLOPs. Moreover, CycleMLP also shows excellent zero-shot robustness on ImageNet-C dataset. |
Description | Oral presentation |
Persistent Identifier | http://hdl.handle.net/10722/315792 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Chen, S | - |
dc.contributor.author | Xie, E | - |
dc.contributor.author | Ge, C | - |
dc.contributor.author | Liang, DY | - |
dc.contributor.author | Luo, P | - |
dc.date.accessioned | 2022-08-19T09:04:31Z | - |
dc.date.available | 2022-08-19T09:04:31Z | - |
dc.date.issued | 2022 | - |
dc.identifier.citation | 10th International Conference on Learning Representations (ICLR) (Virtual), 25-29 April, 2022 | - |
dc.identifier.uri | http://hdl.handle.net/10722/315792 | - |
dc.description | Oral presentation | - |
dc.description.abstract | This paper presents a simple MLP-like architecture, CycleMLP, which is a versatile backbone for visual recognition and dense predictions. As compared to modern MLP architectures, e.g. , MLP-Mixer (Tolstikhin et al., 2021), ResMLP (Touvron et al., 2021a), and gMLP (Liu et al., 2021a), whose architectures are correlated to image size and thus are infeasible in object detection and segmentation, CycleMLP has two advantages compared to modern approaches. (1) It can cope with various image sizes. (2) It achieves linear computational complexity to image size by using local windows. In contrast, previous MLPs have O(N2) computations due to fully spatial connections. We build a family of models which surpass existing MLPs and even state-of-the-art Transformer-based models, e.g. Swin Transformer (Liu et al., 2021b), while using fewer parameters and FLOPs. We expand the MLP-like models’ applicability, making them a versatile backbone for dense prediction tasks. CycleMLP achieves competitive results on object detection, instance segmentation, and semantic segmentation. In particular, CycleMLP-Tiny outperforms Swin-Tiny by 1.3% mIoU on ADE20K dataset with fewer FLOPs. Moreover, CycleMLP also shows excellent zero-shot robustness on ImageNet-C dataset. | - |
dc.language | eng | - |
dc.publisher | IEEE. | - |
dc.relation.ispartof | International Conference on Learning Representation (ICLR), Oral | - |
dc.rights | . Copyright © IEEE. | - |
dc.rights | ©20xx IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. | - |
dc.title | Cyclemlp: A MLP-like architecture for dense prediction | - |
dc.type | Conference_Paper | - |
dc.identifier.email | Luo, P: pluo@hku.hk | - |
dc.identifier.authority | Luo, P=rp02575 | - |
dc.identifier.hkuros | 335562 | - |
dc.publisher.place | United States | - |