International Journal of Fuzzy Logic and Intelligent Systems 2020; 20(1): 8-16
Published online March 25, 2020
https://doi.org/10.5391/IJFIS.2020.20.1.8
© The Korean Institute of Intelligent Systems
Muhammad Aamir1, Nazri Mohd Nawi2, Hairulnizam Bin Mahdin1, Rashid Naseem3, and Muhammad Zulqarnain1
1Faculty of Computer Science and Information Technology, Universiti Tun Hussein Onn Malaysia, Batu Pahat, Malaysia
2Soft Computing and Data Mining Center, Universiti Tun Hussein Onn Malaysia, Batu Pahat, Malaysia
3Department of IT and Computer Science, Pak-Austria Fachhochschule Institute of Applied Sciences and Technology, Haripur, Pakistan
Correspondence to :
Muhammad Aamir (amirr.khan1@gmail.com)
This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
Auto-encoders (AEs) have been proposed for solving many problems in the domain of machine learning and deep learning since the last few decades. Due to their satisfactory performance, their multiple variations have also recently appeared. First, we introduce the conventional AE model and its different variant for learning abstract features from data by using a contrastive divergence algorithm. Second, we present the major differences among the following three popular AE variants: sparse AE (SAE), denoising AE (DAE), and contractive AE (CAE). Third, the main contribution of this study is performing the comparative study of the aforementioned three AE variants on the basis of their mathematical modeling and experiments. All the variants of the standard AE are evaluated on the basis of the MNIST benchmark handwritten digit dataset for classification problem. The observed output reveals the benefit of using the AE model and its variants. From the experiments, it is concluded that CAE achieved better classification accuracy than those of SAE and DAE.
Keywords: Sparse auto-encoder (SAE), Denoising auto-encoder (DAE), Contractive auto-encoder (CAE), MNIST, Classification
International Journal of Fuzzy Logic and Intelligent Systems 2020; 20(1): 8-16
Published online March 25, 2020 https://doi.org/10.5391/IJFIS.2020.20.1.8
Copyright © The Korean Institute of Intelligent Systems.
Muhammad Aamir1, Nazri Mohd Nawi2, Hairulnizam Bin Mahdin1, Rashid Naseem3, and Muhammad Zulqarnain1
1Faculty of Computer Science and Information Technology, Universiti Tun Hussein Onn Malaysia, Batu Pahat, Malaysia
2Soft Computing and Data Mining Center, Universiti Tun Hussein Onn Malaysia, Batu Pahat, Malaysia
3Department of IT and Computer Science, Pak-Austria Fachhochschule Institute of Applied Sciences and Technology, Haripur, Pakistan
Correspondence to:Muhammad Aamir (amirr.khan1@gmail.com)
This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
Auto-encoders (AEs) have been proposed for solving many problems in the domain of machine learning and deep learning since the last few decades. Due to their satisfactory performance, their multiple variations have also recently appeared. First, we introduce the conventional AE model and its different variant for learning abstract features from data by using a contrastive divergence algorithm. Second, we present the major differences among the following three popular AE variants: sparse AE (SAE), denoising AE (DAE), and contractive AE (CAE). Third, the main contribution of this study is performing the comparative study of the aforementioned three AE variants on the basis of their mathematical modeling and experiments. All the variants of the standard AE are evaluated on the basis of the MNIST benchmark handwritten digit dataset for classification problem. The observed output reveals the benefit of using the AE model and its variants. From the experiments, it is concluded that CAE achieved better classification accuracy than those of SAE and DAE.
Keywords: Sparse auto-encoder (SAE), Denoising auto-encoder (DAE), Contractive auto-encoder (CAE), MNIST, Classification
Architecture of the standard AE.
MNIST small subset (basic).
MNIST random-noise background digits (bg-rand).
MNIST rotation and image background digits (bg-img-rot).
ROC for the SAE based on a small subset (basic).
ROC for the SAE based on random noise background digits (bg-rand).
ROC for the SAE based on rotation and image background digits (bg-img-rot).
ROC for the DAE based on MNIST small subset (basic).
ROC for the DAE based on random noise background digits (bg-rand).
ROC for the DAE based on rotation and image background digits (bg-img-rot).
ROC for the CAE based on a small subset (basic).
ROC for the CAE based on random noise background digits (bg-rand).
ROC for the CAE based on rotation and image background digits (bg-img-rot).
Table 1 . Performance evaluation of SAE, DAE, and CAE based on MNIST benchmark datasets.
Datasets | basic | bg-rand | bg-img-rot | ||||||
---|---|---|---|---|---|---|---|---|---|
Approach | Execution time | Training error | Test error | Execution time | Training error | Test error | Execution time | Training error | Test error |
SAE + Softmax | 24m 25s | 5.59 | 13.54 | 31m 25s | 8.92 | 21.54 | 40m 25s | 11.87 | 26.58 |
DAE + Softmax | 31m 20s | 7.73 | 15.78 | 37m 20s | 10.38 | 21.78 | 45m20s | 17.73 | 28.72 |
CAE + Softmax | 28m 55s | 3.35 | 11.65 | 32m 55s | 7.58 | 20.65 | 37m 55s | 11.05 | 24.78 |
Algorithm 1. Auto-encoder.
- No. of hidden layers: - Input feature set: [ - Encoding-activation function: - Decoding-activation function: - Inputs weights: - Biasness values: |
- Compute encoded inputs - Compute biased inputs - Compute |
- Compute decoded outputs - Compute biased outputs - Compute |
- Optimize the value of Eq. (5). |
Nishant Chauhan and Byung-Jae Choi
International Journal of Fuzzy Logic and Intelligent Systems 2021; 21(4): 349-357 https://doi.org/10.5391/IJFIS.2021.21.4.349Jihad Anwar Qadir, Abdulbasit K. Al-Talabani, and Hiwa A. Aziz
International Journal of Fuzzy Logic and Intelligent Systems 2020; 20(4): 272-277 https://doi.org/10.5391/IJFIS.2020.20.4.272Hansoo Lee, Jonggeun Kim, and Sungshin Kim
Int. J. Fuzzy Log. Intell. Syst. 2017; 17(4): 229-234 https://doi.org/10.5391/IJFIS.2017.17.4.229Architecture of the standard AE.
|@|~(^,^)~|@|MNIST small subset (basic).
|@|~(^,^)~|@|MNIST random-noise background digits (bg-rand).
|@|~(^,^)~|@|MNIST rotation and image background digits (bg-img-rot).
|@|~(^,^)~|@|ROC for the SAE based on a small subset (basic).
|@|~(^,^)~|@|ROC for the SAE based on random noise background digits (bg-rand).
|@|~(^,^)~|@|ROC for the SAE based on rotation and image background digits (bg-img-rot).
|@|~(^,^)~|@|ROC for the DAE based on MNIST small subset (basic).
|@|~(^,^)~|@|ROC for the DAE based on random noise background digits (bg-rand).
|@|~(^,^)~|@|ROC for the DAE based on rotation and image background digits (bg-img-rot).
|@|~(^,^)~|@|ROC for the CAE based on a small subset (basic).
|@|~(^,^)~|@|ROC for the CAE based on random noise background digits (bg-rand).
|@|~(^,^)~|@|ROC for the CAE based on rotation and image background digits (bg-img-rot).