File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1109/TIP.2015.2475625
- Scopus: eid_2-s2.0-84959533227
- WOS: WOS:000362008200015
- Find via
Supplementary
- Citations:
- Appears in Collections:
Article: PCANet: A Simple Deep Learning Baseline for Image Classification?
Title | PCANet: A Simple Deep Learning Baseline for Image Classification? |
---|---|
Authors | |
Keywords | Convolution Neural Network Deep Learning Face Recognition Handwritten Digit Recognition LDA Network Object Classification PCA Network Random Network |
Issue Date | 2015 |
Citation | IEEE Transactions on Image Processing, 2015, v. 24, n. 12, p. 5017-5032 How to Cite? |
Abstract | In this paper, we propose a very simple deep learning network for image classification that is based on very basic data processing components: 1) cascaded principal component analysis (PCA); 2) binary hashing; and 3) blockwise histograms. In the proposed architecture, the PCA is employed to learn multistage filter banks. This is followed by simple binary hashing and block histograms for indexing and pooling. This architecture is thus called the PCA network (PCANet) and can be extremely easily and efficiently designed and learned. For comparison and to provide a better understanding, we also introduce and study two simple variations of PCANet: 1) RandNet and 2) LDANet. They share the same topology as PCANet, but their cascaded filters are either randomly selected or learned from linear discriminant analysis. We have extensively tested these basic networks on many benchmark visual data sets for different tasks, including Labeled Faces in the Wild (LFW) for face verification; the MultiPIE, Extended Yale B, AR, Facial Recognition Technology (FERET) data sets for face recognition; and MNIST for hand-written digit recognition. Surprisingly, for all tasks, such a seemingly naive PCANet model is on par with the state-of-the-art features either prefixed, highly hand-crafted, or carefully learned [by deep neural networks (DNNs)]. Even more surprisingly, the model sets new records for many classification tasks on the Extended Yale B, AR, and FERET data sets and on MNIST variations. Additional experiments on other public data sets also demonstrate the potential of PCANet to serve as a simple but highly competitive baseline for texture classification and object recognition. |
Persistent Identifier | http://hdl.handle.net/10722/327087 |
ISSN | 2023 Impact Factor: 10.8 2023 SCImago Journal Rankings: 3.556 |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Chan, Tsung Han | - |
dc.contributor.author | Jia, Kui | - |
dc.contributor.author | Gao, Shenghua | - |
dc.contributor.author | Lu, Jiwen | - |
dc.contributor.author | Zeng, Zinan | - |
dc.contributor.author | Ma, Yi | - |
dc.date.accessioned | 2023-03-31T05:28:43Z | - |
dc.date.available | 2023-03-31T05:28:43Z | - |
dc.date.issued | 2015 | - |
dc.identifier.citation | IEEE Transactions on Image Processing, 2015, v. 24, n. 12, p. 5017-5032 | - |
dc.identifier.issn | 1057-7149 | - |
dc.identifier.uri | http://hdl.handle.net/10722/327087 | - |
dc.description.abstract | In this paper, we propose a very simple deep learning network for image classification that is based on very basic data processing components: 1) cascaded principal component analysis (PCA); 2) binary hashing; and 3) blockwise histograms. In the proposed architecture, the PCA is employed to learn multistage filter banks. This is followed by simple binary hashing and block histograms for indexing and pooling. This architecture is thus called the PCA network (PCANet) and can be extremely easily and efficiently designed and learned. For comparison and to provide a better understanding, we also introduce and study two simple variations of PCANet: 1) RandNet and 2) LDANet. They share the same topology as PCANet, but their cascaded filters are either randomly selected or learned from linear discriminant analysis. We have extensively tested these basic networks on many benchmark visual data sets for different tasks, including Labeled Faces in the Wild (LFW) for face verification; the MultiPIE, Extended Yale B, AR, Facial Recognition Technology (FERET) data sets for face recognition; and MNIST for hand-written digit recognition. Surprisingly, for all tasks, such a seemingly naive PCANet model is on par with the state-of-the-art features either prefixed, highly hand-crafted, or carefully learned [by deep neural networks (DNNs)]. Even more surprisingly, the model sets new records for many classification tasks on the Extended Yale B, AR, and FERET data sets and on MNIST variations. Additional experiments on other public data sets also demonstrate the potential of PCANet to serve as a simple but highly competitive baseline for texture classification and object recognition. | - |
dc.language | eng | - |
dc.relation.ispartof | IEEE Transactions on Image Processing | - |
dc.subject | Convolution Neural Network | - |
dc.subject | Deep Learning | - |
dc.subject | Face Recognition | - |
dc.subject | Handwritten Digit Recognition | - |
dc.subject | LDA Network | - |
dc.subject | Object Classification | - |
dc.subject | PCA Network | - |
dc.subject | Random Network | - |
dc.title | PCANet: A Simple Deep Learning Baseline for Image Classification? | - |
dc.type | Article | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1109/TIP.2015.2475625 | - |
dc.identifier.scopus | eid_2-s2.0-84959533227 | - |
dc.identifier.volume | 24 | - |
dc.identifier.issue | 12 | - |
dc.identifier.spage | 5017 | - |
dc.identifier.epage | 5032 | - |
dc.identifier.isi | WOS:000362008200015 | - |