International Journal of Fuzzy Logic and Intelligent Systems 2023; 23(4): 465-481
Published online December 25, 2023
https://doi.org/10.5391/IJFIS.2023.23.4.465
© The Korean Institute of Intelligent Systems
Archana P. Kale
Modern Education Society’s College of Engineering, Savitribai Phule Pune University, Maharashtra, India
Correspondence to :
Archana P. Kale (archana.mahantakale@gmail.com)
This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
Fuzzy extreme learning machine (F-ELM) one of the learning algorithm which is specifically used for uncertainty classification. Uncertainty classification is critical problem in the area of machine learning. In various real-time applications, the ambiguity is present in the input dataset itself which affects the systems generalization performance. Data (feature) analysis plays a major role in such types of problems. To solve the said problem in this paper, distance-based Relief algorithm and fuzzy extreme learning algorithm are used for data analysis and classification, respectively which contributes to Relief-based data analysis for F-ELM (RFELM++) and Relief-based data analysis for online sequential ELM (RFOSELM++) for batch mode and sequential input, respectively. Experimental results are calculated by using clinical dataset. Through the results, it is observed that RFELM++ produces increased accuracy in comparison with RELM++ for clinical dataset. The RFOSELM++ maintain accuracy by using 41.5% features as differentiate to OS-ELM for the UCI Repository dataset. The proposed RFOSELM++ algorithm is compared with similar already available sequential based algorithms. As a case study, a novel application of classification of nutrient deficiency and plant disease in which ambiguity presents is considered. Both proposed algorithms are exploited in this expert system which helps the remote farmer with expert advice. The proposed RFELM++ algorithm is tested and validated by using statistical methods.
Keywords: Uncertainty classification problem, Sequential problem, Internet of Things, Feature subset selection problem, Pattern classification
Classification, feature subset selection (FSS), and sequential problems are some of the critical problems in machine learning which has drawn substantial aspect from researchers in recent years. Due to the technological advances, data gets generated at an ever increasing space and the dimensionality or size of the datasets are increased by the day. Therefore, it is mandatory to design and develop the potent and adequate machine learning methods, that can be used to select, analyse and extract only the beneficial features. The existence of the dispensable and immaterial features in input may decreases the performance of systems. FSS is critical important pre-processing technique which selects only prominent (non-redundant and relevant) features. Pattern classification is a critical process which is utilize to classify the input dataset into one of the output classes.
As database and computer technologies advance speedily, various advancing challenges appear with respect to FSS and classification problems as follows:
• Improvement in accuracy by using the same number of attributes;
• Maintain accuracy by downsizing the attributes;
• Improvement in accuracy by reducing the number of attributes.
FSS problem has drawn considerable attention from researchers in current years due to the existence of a huge attributes. Many real-time applications contain high-dimensional dataset. These thousands of features are usually provided to the learning algorithms for the classification task.
In many real-time applications, the attributes present in the dataset are ambiguous. Therefore, it becomes very challenging to classify the instance of input. Such kind of the problems in which the ambiguity presents in the attributes of the input dataset [1] are come under the uncertainty classification problem (UCP). Uncertainty classification is used for one class classification [2], binary classification problems [3,4] and imbalance classification problem via support vector machine (SVM) [5]. A gradient approach is developed in Naive Bayes (NB) for value uncertainty classification learning by Lee [6]. Zhang and Ji [1] developed fuzzy extreme learning machine (F-ELM) for uncertainty classification problem.
F-ELM is the latest classifier which associates the working of fuzzy inference system (FIS) and ELM. For uncertainty classification problem F-ELM algorithm is mostly used [1, 7]. F-ELM uses a membership function for required input which is represented by linguistic variables.
Many research papers show the different FSS techniques for various classification problems. However, FSS with incremental learning for uncertainty classification problem is absent in the vast literature survey.
The original F-ELM and ELM are the batch learning mode which considers that the data is available initially. So, F-ELM and ELM classifiers are unable to handle sequent or serial input. So that, online sequential extreme learning machine (OS-ELM) algorithm is presented by Nan-Ying et. al. for sequent input. Three layers are available for ELM and OS-ELM: input, hidden, and output. Randomly initialization of weights is carried out from the first (input) to second (hidden) layer and logically modulated the weights from the second (hidden) to third (output) layer. On account of random initialization, generalized performance of system may degrade. So that it is necessary to select the prominent features.
To distinguish said problem, in this paper incremental F-ELM (RFELM++) and incremental fuzzy online sequential ELM (RFOSELM++) for the UCP are proposed for batch mode and sequential input respectively. Thus, the key intent of this paper is to design RFELM++ and RFOSELM++ algorithm by using the activation function like radial basis function (RBF) and additive function which is missing in the vast literature survey. Another aspect of the paper is to design an IoT based expert system - Precision Agriculture by exploiting the proposed algorithms.
The overall structure of the paper is the following: in Section 2, the literature survey is detailed in the related work. The proposed incremental F-ELM and RFOSELM++ algorithms are depicted in Section 3. The experimental observations and results of clinical datasets are discussed and compared in Section 4. The validation and testing part of the proposed algorithm are discussed in detail in Section 5 as the results and discussions. Section 6 outlined the conclusion with further research direction.
In various growing areas FSS plays a very critical task. The key role of FSS is to choose prominent/optimal (relevant and non-redundant) feature subset which produces similar or improved (in some of the cases) results as compared to the prime or base data set in which all features are present. It is very hard to decide the significance of attributes without preparatory information [8]. Therefore, FSS is an prime step to select prominent features for efficient machine learning applications [9, 10]. The primary intention of using any FSS techniques is for [11]:
• Improvement in classification accuracy,
• Improvement in prediction performance,
• Improvement in model interpretability,
• Dimensionality reduction,
• Storage management.
The few of the FSS techniques which are mostly used are: filter, wrapper, and hybrid.
Filter techniques rank the features or select feature subset independently of the learning algorithm. Filter techniques are easy to interpret and fast. The characteristics of filter techniques are considered independently, totally independent of which learning algorithm is used etc. For class separability, the evaluation criteria is required which is a classical criterion of the filtering present in the literature. The criteria like distance, information and reliant are mostly utilize to classify the features.
Wrapper techniques use a learning algorithm for FSS, training is required for every feature subset. So, wrapper techniques are computationally heavy and dependent on the type of classifier used. The characteristics of wrapper techniques are computationally expensive, use only heuristic search and dependent on the learning algorithm.
Hybrid techniques use both filter and wrapper techniques. Ranking-wrapper techniques [12], Relief-F with sequential backward search [13] and boosting-based hybrid for feature selection (BBHFS) [14] are few of hybrid techniques.
Tahir and Loo [15] used Relief-F method for feature selection in continual learning framework. Various optimization algorithms like particle swarm intelligence [16] is used to solve feature selection problem and multi-objective particle swarm optimization algorithm [17] is used to solve bioprocess application problems and tumor treatment problems.
Pattern classification is specially used for decision making task. It classifies input entity to one of the predetermined target outputs. A supervised classification is specially used to categorised the given input data into one of the present target classes [18]. Having the data
The original ELM classifier is basically designed for a batch mode. ELM is acknowledged that the complete input data is available previously. However, many real applications require to handle the data sequentially. Therefore, Nan-Ying Liang et. al. have designed OS-ELM algorithm for handling the sequent or serial input. OS-ELM randomly assigns all the weight values in-between first (input) and second (hidden) layer and the weight values in between second (hidden) to third (output layer) are analytically tuned [23–25] as demonstrated in Figure 1 [26]. The global generalization performance may degrade due to the existence of non-prominent attributes and random initialization. Therefore, it is necessary to choose a prominent feature subset for sequential input. However, ELM is insufficient to provide a solution for the sequential and UCP.
OS-ELM is exploited in various applications like estimation of hematocrit [27], DWT domain watermarking [28], traffic profiling [29], etc. However, the literature survey lacks the exploitation of incremental learning ability by using prominent or optimal feature subset selection for sequent or serial input.
Traditional ELM is unable to solve some real-time time problems like the UCP and imbalance problem [1]. So, F-ELM is designed by Zhang and Ji [1] to solve specially UCP. However, in order to solve the UCP for sequential input OS-ELM is designed [30].
Ontiveros-Robles et al. [31] used type II fuzzy classification for non-singleton and also designed hybrid classification by using SVM [32]. Rubio et al. [33] used fuzzy clustering for unsuprvised method.
Incremental learning is used in various ways like incrementally add hidden nodes [34, 35], incrementally add streaming data [36], incremental FSS [37–40], etc. The exploitation of incremental learning in terms of incremental FSS for sequential input is also missing in the literature survey.
Incremental learning comes in various forms and definitions as
• Incrementally add streaming data: Incremental learning is a learning for the data which appear sequentially or over time [36, 41]. Alexander Gepperth describes online learning, incremental learning and concept drift for the supervised learning paradigm in detail.
• Incrementally add hidden nodes: Incremental learning defines as an incrementally insertion of the new neurons to the hidden layer. Error-minimized ELM (EM-ELM) [42], Convex incremental ELM (CI-ELM) [35] and uncremental extreme learning machine (RELM++) [34] are some of the examples of incremental learning. Yang et al. [43] designed a bidirectional extreme learning machines (B-ELM) algorithm for regression.
• Incremental feature subset selection: Incremental learning in the form of incremental feature subset selection is used in many papers like [37–40]. Baluja [44] have developed a population-based incremental learning method which integrated learning and the random search based function optimization [44]. Tahir and Loo [15] used an incremental ELM is used for food recognition. Wang et al. [45] have designed incremental wrapper based gene selection with Markov blanket. Incremental wrapper subset selection (IWSS) uses a filter measure for feature ranking by using a sequential search [46]. Then, a SFS is used for incrementally insertion of the features.
Wang et al. [47] designed a framework for smart contracts based on a novel six-layer architecture.
This section is categorized into two subsections, RFELM++ and RFOSELM++ algorithms.
In this section, incremental learning is used to resolve the uncertainty classification problem. As already discussed in Section 2, F-ELM specifically uses for the uncertainty classification problem. Hence, RFELM++ algorithm is initiated. The paradigm of RFELM++ is broadly categorised into four subsystems: input, incremental learning for FSS, fuzzification, and classification.
In this section, the detailed description of the input is described in detail. In order to resolve the UCP, it is necessary to use the datasets in which uncertainty is present. Therefore, Statlog Heart Disease (SHD) and Pima Indian Diabetes (PID) datasets are used [48]. The PID data set contains total 768 instances and 8 attributes. The datasets consists of several medical predictor variables and one target variable, outcome. Predictor variables includes the number of pregnancies the patient has had, their body mass index (BMI), insulin level, age, and so on. The SHD data set contains 270 instances and 13 attributes. Additionaly, the Cleveland Heart Disease (CHD) dataset is used for multiclass classification.
The generalization performance of any system is depend on the input. If the input to the system is right the performance of system gets increased. Therefore, it is very important task to provide a right or correct input features to the system.
In data preprocessing FSS algorithm is used. To select only prominent features relief algorithm is used. Relief algorithm is one of the FSS algorithms based on distance measure. The measure of distance (D) is basically used to search the features which can differentiated between two target classes. If the D is equal to 0, then the two attributes are identical. The Relief algorithm is used the joint relationship to compute weights with the output [49, 50] which is famous due to it’s simplicity and effectiveness. In Relief (feature weight-based algorithm), a statistical method is used to choose the relevant attributes [51].
After applying relief algorithm on PID and SHD datasets, we get feature order. For example, PID dataset has 8 attributes like age, DPF, BMI, insulin, skin, BP, Glu, and pregnancy. After Relief algorithm we get 〈2, 6, 4, 8, 1, 7, 5, 3〉 feature order. With this order it is understood that 2 is the most required or more relevant feature whereas 3 is the least required feature.
As per already described in the previous section, prominent feature order is calculated by considering distance measures i.e., Relief algorithm. Suppose the attribute order for SHD and PID data sets is 〈13, 12, 3, 9, 10, 11, 2, 1, 7, 4, 5, 6, 8〉 and 〈2, 6, 8, 1, 7, 5, 4, 3〉 respectively. The organised attributes are incrementally inserted one by one by using SFS for sequential space. Experimental observation shows that from all the subsets {1, 2, 3, 4, 5, 7, 9, 10, 11, 12, 13} and {1, 2, 4, 5, 6, 7, 8} are the prominent feature subset for SHD and PID datasets, respectively, as it provides improved classification accuracy.
The attributes of PID and SHD datasets are translated into linguistic attributes (fuzzified features) with the help of trapezoidal membership function [52]. The PID (8 attributes) and SHD (13 attributes) datasets are translated into 25 and 39 linguistic attributes, respectively. Input dataset is distributed into two datasets like training with 70% and testing with 30 %. Hence for PID ans SHD datasets 577 and 177 instances are used for traing and 577 and 177 instances are used for testing, respectively. By using trapazoidal membership function the features are converted into fuzzified features. For classification ELM is used.
ELM is a batch learning mode classifier which is not suitable for Sequential input. In this section, incremental learning for FSS is used to solve the UCP as well as sequential problem. Hence, incremental FSS techniques for the UCP is proposed for sequential input which contributes to RFOSELM++ algorithm.
In RFOSELM++, the ranked features are inserted sequentially i.e., one by one. The accuracy is calculated for all feature subsets which are equal to the number of attributes. The proposed RFOSELM++ learning algorithm is applied for every subset and predicted accuracy is calculated. The paradigm of RFOSELM++ is mostly same as RFELM++. Only the classification subsystem is different one. Through observations, the prominent feature subset is one which provides maximum accuracy.
For experimental observation, Waikato Environment for Knowledge Analysis (WEKA) and MATLAB R2014a are used. Precision, Recall and classification accuracy are used as an evaluation measure. The calculations of evaluation measures are depend upon the contingency table or confusion matrix of classifier. Confusion matrix contains both true and false positives (TP/FP) also true and false negatives (TN/FN) [52]. FP correspond to negative samples falsely identified as positive. FN correspond to positive samples falsely resembles as negative. TP are the samples perfectly identified as positive. TN refer to negative samples perfectly identified as negative. The perfectness of identified examples are measured by using precision. The measure of the all positive identified examples is called recall [53]. Accuracy, precision and recall (sensitivity) are measured by evaluating the respective values in
In this section, the incremental learning capability for batch and sequential mode is exploited i.e., incremental learning for F-ELM (RFELM++) algorithm and incremental learning for fuzzy online sequential extreme learning machine, respectively.
The experimental results of the proposed RFELM++ algorithm are enumerated. In order to prove the effectiveness of an initiated algorithm the same steps are implemented for ELM classifier (RELM++). The results of both RELM++ and RFELM++ are compared with sigmoidal activity function by considering PID and SHD datasets. The overall comparison of RELM++ and RFELM++ for PID dataset is as shown in Figure 2. The same results are evaluated for SHD dataset also. Through the results, it is observed that RFELM++ produces 6.266% and 0.855% increased accuracy in comparison of RELM++ for PID and SHD datasets consequently.
The overall performance of the FELM classifier is analysed by considering the precision, accuracy and recall. Figure 3 shows an overall comparison of performance for ELM and F-ELM with all features. The results are calculated by using all activity functions like sigmoid function (sig), radial basis function (rbf), sine function (sin), hardlim function (hardlim), triangular basis function (tribas) for SHD and PID datasets. By observations it is apprise that, F-ELM issues an improved performance as compared to ELM classifier.
Various FSS methods like half selection (HS), mean selection (MS), particle swarm optimization (PSO) based FS, neural network for threshold, self-regulated learning (SR-PSO) [54] are used to select prominent features. Table 1 indicates the output of all given FSS methods with ELM and FELM classifiers. For evaluation, the sigmoidal activation function is used for PID dataset. From experimented results, it is noticed that the F-ELM achieves 4.65% improved classification accuracy in comparison with ELM for the same number of feature. The result for RFELM++ algorithm for multiclass classification problem for CHD dataset is as shown in Figure 5.
In order to prove the effectiveness of the proposed RFELM++ algorithm, experiment results are derived and analysed by using the traditional classifiers–MLP, random forest (RF), Bayes net (BN), SVM, NB, RBF, and J4.8–by using prominent features. The prominent features are selected by using Relief algorithm. Five-fold cross-validation method is used for evaluation. Figure 4 shows the variation of the proposed RFELM++ algorithm with existing classifiers by using relief FSS algorithm for PID and SHD data set. It is evidently prove that the proposed RFELM++ algorithm enables maximum predictive accuracy with minimum number of features.
PID dataset is used as an input dataset. Normalization is used in pre-process. The algorithm uses relief algorithm (distance measure) for ranking the features. IWSS [39, 40] is used which acquires a ranking of the features over to the target class. IWSS is used SFS search over the ranked attributes by incrementally inserting those features one by one. The wrapper way is used to measure the relevance of the new attribute. The classification accuracy is calculated by using OS-ELM classifier.
Experimental results of incremental attribute learning for sequential input is calculated for both ELM and F-ELM classifier. Table 2 shows the parameters required for the evaluation. Figures 6 and 7 show the comparison of RELM++ and RFELM++ for training and testing accuracy, respectively. Through observation, it is noticed that the proposed RFELM++ issues an improved as well as stable (constant) results as compared to RELM++.
In order to prove the validation of the proposed RFELM++ algorithm, experimental observations need to be validated and hypothetically tested [55]. F-ELM classifier is used after the feature selection by using statistical methods. The PID and SHD datasets are categorized into four samples like S4 (SHD ELM), S3 (PID ELM), S2 (SHD F-ELM) and S1 (PID F-ELM). The comparative analysis of before and after relief feature selection algorithm is as shown in Figure 8. Through the results it is noticed that, RFELM++ algorithm with feature subset produces improved performance as compared to with all features.
The correlation analysis technique is used to compute the robustness of the relationship between 2 or more variables. All the time the values of correlation are in between −1 to +1. The abbreviation of +1 is that the two variables are strongly related with positive linear, −1 is that two variables are resolutely related with negative linear and 0 is that no relationship in between the two variables. According to Evans classification, the correlation coefficient are broadly categorized into very strong, strong, moderate, weak, very weak [56].
The
where ‘
The coefficient (
When the relationship (
where ‘
Hypothesis of testing [57] RFELM++:
The
For performance comparison, accuracy of RFELM++ by using prominent feature subset is compared with F-ELM [59], modified fuzzy min-max neural network-FIS (MFMM-FIS) [60] and enhanced generalized adaptive resonance theory FIS (EGART-FIS) [61] for the PID dataset. Table 3 describes the comparative classification accuracy for the same. With the comparative analysis, it is observed that the testing and training accuracy is 83.47% to 86.80%, respectively by using only 62.5% features.
With the intention to evince the effectiveness of the proposed RFELM++ algorithm, it is mandatory to have comparison the result of RFELM++ with the similar approach like NN for threshold selection (NN) [62,63], multilayer NN (MLNN) by using Levenberg–Marquardt (LM) [64], generalized discriminant analysis least square (GDA-LS) [65], HS method, least square-ELM (LS-ELM) [65], linear discriminant analysis and Morlet wavelet (LDA-MW), probabilistic neural network (PNN) [64], MS methods. Table 4 shows RFELM++ and other classifiers comparison for PID input dataset. Table 5 illustrates the same comparison for SHD dataset with evolutionary product-unit NN (EPUNN), MS, evolutionary sigmoidal unit NN (ESUNN), HS, multilogistic regression by means of evolutionary product-unit NN (MR+EPUNN) for SHD dataset. Through comparative average analysis, it is noticed that the introduced RFELM++ issues 8.82% and 6.63% increased accuracy by reducing 6.41% and 10.71% for SHD adn PID datasets, respectively.
The comparative analysis in between F-ELM and ELM by using different FSS algorithms is as shown in Figure 9 for PID dataset. HS, PSO, MS, NN, T-test and self-regulated learning (SR-PSO) [54] are used as FSS algorithms. For PID dataset, the comparison of ELM and F-ELM classifier in terms of accuracy measure by using the feature subset as shown in Figure 9 in which F-ELM produces increased performance for the same number of features.
The introduced RFELM++ algorithm is exploited in the real time application which is represented as a case study in the following section.
Agriculture is the backbone of Indian economy where approximate 60% of people are dependent directly or indirectly on agriculture. The expert advice is required for distinguishing the plant disease damage and nutrient imbalance. It is observed that, the conventional judgmental analysis is not enough while deciding the quantity of chemical or fertilizer to be used. The mis-proportional dose harms directly the health of the crop and hence the living beings. To overcome the said problem, this section is developed an IoT-based precision agriculture expert system by inferring proposed algorithms with improving the accuracy by reducing the features or by selecting the prominent features. Grape leaf diseases like downy mildew and nutrient deficiency which contains ambiguity is considered as a case study. Both are spread with yellow spot on leaf. The conceptual framework for the proposed IoT-based precision agriculture expert system is as shown in Figure 10.
In IoT-based precision agriculture expert system, the system is partitioned into three rough phases namely image processing, FSS by using incremental learning and classification. In image processing, the acquisition step is used to input the image. The image is acquired by using cell phone or any digital camera. The resolution of the image is 96 dpi (dots per inch). For interfacing, the ARDUINO hardware is used with various sensors like temperature, humidity etc. The leaf image is used to extract the feature points and are stored in the database.
Data normalization is used as the preprocessing method. In image analysis, image segmentation and feature extraction methods are considered. For image segmentation, region-based segmentation is used to separate the healthy and diseased region by using leaf color. The features extracted from an images are color, shape and texture [71]. Once features are extracted, the database is generated with integrated information from temperature and humidity sensors. On that created dataset the proposed algorithms are applied for feature selection and classification.
In order to increase the generalization performance of the system, the proposed RELM++ and RFOSELM++ algorithms are exploited. The performance of precision agriculture is evaluated and tested by using RELM++ and RFOSELM++ algorithms. The comparison in-between RELM++ and ELM for proposed system (plant disease and nutrient deficiency) is as shown in Figure 11 in terms of testing accuracy. Through this experimentation, it is intended that the RELM++ produces 5.71% increased accuracy and RFOSELM++ maintains (same accuracy) the accuracy with only 27.27% features. However, for Tribas activation function the classical method provides good results as compared to the proposed one. Tribas provided an optimal result only for certain number of hidden neurons for ELM [72].
The prime intention of this paper is to exploit the two proposed algorithms like RFELM++ and RFOSELM++ into the new developed IoT-based precision agriculture Expert system. The expert system helps farmers in decision making. The experimental results for ELM, OS-ELM, RELM++, RFOSELM++, RFELM++ and RFOSELM++ are calculated. RFELM++ produces 6.266% and 0.855% increased accuracy in comparison of RELM++ for PID and SHD datasets consequently. Additionally, various FSS methods like like HS, MS, PSO based FS, neural network for threshold, self-regulated learning (SR-PSO) [54] are used to select prominent features. Through the experimented results, it is noticed that the F-ELM achieves 4.65% improved classification accuracy in comparison with ELM for the same number of feature. The proposed RFOSELM++ maintain the accuracy by using 41.5% features as differentiate to OS-ELM for UCI Repository dataset. For benchmark problem, the results of initiated RFELM++ algorithm is compared with the similar existing approaches. From the comparative average analysis, it is noticed that the introduced RFELM++ issues 8.82% and 6.63% increased accuracy by reducing 6.41% and 10.71% for SHD and PID datasets, respectively. Table 3 describes the comparative classification accuracy for the same. With the comparative analysis, it is observed that the testing and training accuracy is 83.47% to 86.80%, respectively by using only 62.5% features. Hypothetically it is proved that there is no variance between the generalization performance of F-ELM (with all features or before feature selection) and RFELM++ (after feature selection) for four samples. A novel application, IoT-based precision agriculture expert system, is developed by exploiting the proposed algorithms RFELM++ and RFOSELM++. Currently, the algorithms are used for binary classification in future it can be used for multi-class classification. The precision system can be improved by considering the rain removal [73] and by using CNN [74].
No potential conflict of interest relevant to this article was reported.
Comparison of ELM and F-ELM: (a) accuracy of PID dataset, (b) accuracy of SHD dataset, (c) precision of PID dataset, (d) precision of SHD dataset, (e) recall of PID dataset, and (f) recall of SHD dataset.
Accuracy measure comparison of F-ELM with various classifiers by using Relief-FSS algorithm: (a) input PID dataset and (b) input SHD dataset.
Table 1. Comparative results of ELM and F-ELM with different FS methods for input PID dataset.
FSS Methods | ELM | F-ELM | Feature subset |
---|---|---|---|
MS | 78.26 | 79.13 | 2, 6, 8 |
HS; l | 76.08 | 82.6 | 1, 2, 6, 8 |
NN for threshold selection | 78.26 | 82.6 | 1, 2, 6, 8 |
PSO | 78.26 | 82.6 | 1, 2, 6, 8 |
SRLPSO | 78.26 | 79.13 | 2, 6, 8 |
76.52 | 83.47 | 1, 2, 6, 7, 8 | |
Without FSS Method | 71.73 | 80.43 | All |
Avg | 76.76 | 81.42 | - |
Table 2. Parameters for RFELM++.
Parameter | Features |
---|---|
Input neuron | Same as input features |
Output neuron | 2 |
Hidden neuron | 10:5:200 |
Activity function | Sigmoidal |
Sequential mode | 1 by 1, 20 by 20, 10 by 30 |
Dataset division | Training (70%)-Testing (30%) |
Table 3. Comparison of RFELM++ with existing learning algorithms for input PID dataset.
Methods/algorithm | Accuracy of training | Accuracy of testing | Total number of attributes used |
---|---|---|---|
RFELM++ | 86.80 | 83.47 | 05 (62.5%) |
F-ELM [59] | 75.35 | 74.09 | 08 (100%) |
MFMM-FIS [60] | n/a | 72.92 | 08 (100%) |
EGART-FIS [61] | n/a | 73.05 | 08 (100%) |
MFMM-FIS [62] | n/a | 72.92 | 08 (100%) |
Table 4. RFELM++ and existing learning algorithms comparison for input PID dataset.
Method/algorithm | Accuracy (testing) | Total number of attributes used | Feature subset |
---|---|---|---|
RFELM++ | 83.47 | 05 | {1,2,6,7,8} |
RELM++ | 76.52 | 05 | {1,2,6,7,8} |
EGART-FIS [61] | 73.05 | 08 | {1,2,3,4,5,6,7,8} |
F-ELM [59] | 74.09 | 08 | {1,2,3,4,5,6,7,8} |
MFMM-FIS [60] | 72.92 | 08 | {1,2,3,4,5,6,7,8} |
LS-ELM [65] | 78.21 | 04 | {1,2,6,8} |
GDA-LS-ELM [65] | 79.16 | 05 | {1,2,6,7,8} |
MLNN with LM [64] | 79.62 | 04 | {1,2,6,8} |
PNN [64] | 78.05 | 03 | {2,6,8} |
MS [62] | 76.04 | 03 | {2,6,8} |
HS [62] | 75.91 | 04 | {1,2,6,8} |
NN [62] | 76.04 | 03 | {2,6,8} |
GP-KNN [66] | 80.50 | 08 | {1,2,3,4,5,6,7,8} |
Gini-Fuzzy [67] | 75.80 | 08 | {1,2,3,4,5,6,7,8} |
FMNN-CART-RF [25] | 78.39 | 08 | {1,2,3,4,5,6,7,8} |
Generic statistical approach [68] | 78.04 | 08 | {1,2,3,4,5,6,7,8} |
76.84 | 26.79 | - |
Table 5. RFELM++ and existing learning algorithms comparison for Input SHD dataset.
Method/algorithm | Accuracy (testing) | Number of features | Feature subset |
---|---|---|---|
RFELM++ | 92.59 | 06 | {3,9,10,11,12,13} |
RELM++ | 90.74 | 05 | {3,9,10,12,13} |
ESUNN [69] | 83.22 | 05 | {3,8,9,11,12} |
EPUNN [69] | 81.89 | 04 | {8,9,11,12} |
MR+EPUNN [70] | 83.12 | 05 | {8,9,11,12,13} |
MS [62] | 84.44 | 06 | {3,8,9,11,12,13} |
HS [62] | 84.81 | 07 | {3,8,9,10,11,12,13} |
NN [62] | 85.19 | 04 | {3,11,12,13} |
83.77 | 60.26 | - |
Table 6. Evaluation comparative analysis for ELM and F-ELM with different FSS for input PID dataset.
FSS Method | ELM | F-ELM | Feature subset |
---|---|---|---|
LS [65] | 78.21 | 82.6 | {1,2,6,8} |
GDA-LS [65] | 79.16 | 82.17 | {1,2,6,7,8} |
MLNN with LM [64] | 79.62 | 82.6 | {1,2,6,8} |
PNN [64] | 78.05 | 79.13 | {2,6,8} |
LDA-MW [62] | 89.74 | 82.17 | {1,2,6,7,8} |
MS [62] | 76.04 | 79.13 | {2,6,8} |
HS [62] | 75.91 | 82.6 | {1,2,6,8} |
NN [62] | 76.04 | 79.13 | {2,6,8} |
International Journal of Fuzzy Logic and Intelligent Systems 2023; 23(4): 465-481
Published online December 25, 2023 https://doi.org/10.5391/IJFIS.2023.23.4.465
Copyright © The Korean Institute of Intelligent Systems.
Archana P. Kale
Modern Education Society’s College of Engineering, Savitribai Phule Pune University, Maharashtra, India
Correspondence to:Archana P. Kale (archana.mahantakale@gmail.com)
This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted noncommercial use, distribution, and reproduction in any medium, provided the original work is properly cited.
Fuzzy extreme learning machine (F-ELM) one of the learning algorithm which is specifically used for uncertainty classification. Uncertainty classification is critical problem in the area of machine learning. In various real-time applications, the ambiguity is present in the input dataset itself which affects the systems generalization performance. Data (feature) analysis plays a major role in such types of problems. To solve the said problem in this paper, distance-based Relief algorithm and fuzzy extreme learning algorithm are used for data analysis and classification, respectively which contributes to Relief-based data analysis for F-ELM (RFELM++) and Relief-based data analysis for online sequential ELM (RFOSELM++) for batch mode and sequential input, respectively. Experimental results are calculated by using clinical dataset. Through the results, it is observed that RFELM++ produces increased accuracy in comparison with RELM++ for clinical dataset. The RFOSELM++ maintain accuracy by using 41.5% features as differentiate to OS-ELM for the UCI Repository dataset. The proposed RFOSELM++ algorithm is compared with similar already available sequential based algorithms. As a case study, a novel application of classification of nutrient deficiency and plant disease in which ambiguity presents is considered. Both proposed algorithms are exploited in this expert system which helps the remote farmer with expert advice. The proposed RFELM++ algorithm is tested and validated by using statistical methods.
Keywords: Uncertainty classification problem, Sequential problem, Internet of Things, Feature subset selection problem, Pattern classification
Classification, feature subset selection (FSS), and sequential problems are some of the critical problems in machine learning which has drawn substantial aspect from researchers in recent years. Due to the technological advances, data gets generated at an ever increasing space and the dimensionality or size of the datasets are increased by the day. Therefore, it is mandatory to design and develop the potent and adequate machine learning methods, that can be used to select, analyse and extract only the beneficial features. The existence of the dispensable and immaterial features in input may decreases the performance of systems. FSS is critical important pre-processing technique which selects only prominent (non-redundant and relevant) features. Pattern classification is a critical process which is utilize to classify the input dataset into one of the output classes.
As database and computer technologies advance speedily, various advancing challenges appear with respect to FSS and classification problems as follows:
• Improvement in accuracy by using the same number of attributes;
• Maintain accuracy by downsizing the attributes;
• Improvement in accuracy by reducing the number of attributes.
FSS problem has drawn considerable attention from researchers in current years due to the existence of a huge attributes. Many real-time applications contain high-dimensional dataset. These thousands of features are usually provided to the learning algorithms for the classification task.
In many real-time applications, the attributes present in the dataset are ambiguous. Therefore, it becomes very challenging to classify the instance of input. Such kind of the problems in which the ambiguity presents in the attributes of the input dataset [1] are come under the uncertainty classification problem (UCP). Uncertainty classification is used for one class classification [2], binary classification problems [3,4] and imbalance classification problem via support vector machine (SVM) [5]. A gradient approach is developed in Naive Bayes (NB) for value uncertainty classification learning by Lee [6]. Zhang and Ji [1] developed fuzzy extreme learning machine (F-ELM) for uncertainty classification problem.
F-ELM is the latest classifier which associates the working of fuzzy inference system (FIS) and ELM. For uncertainty classification problem F-ELM algorithm is mostly used [1, 7]. F-ELM uses a membership function for required input which is represented by linguistic variables.
Many research papers show the different FSS techniques for various classification problems. However, FSS with incremental learning for uncertainty classification problem is absent in the vast literature survey.
The original F-ELM and ELM are the batch learning mode which considers that the data is available initially. So, F-ELM and ELM classifiers are unable to handle sequent or serial input. So that, online sequential extreme learning machine (OS-ELM) algorithm is presented by Nan-Ying et. al. for sequent input. Three layers are available for ELM and OS-ELM: input, hidden, and output. Randomly initialization of weights is carried out from the first (input) to second (hidden) layer and logically modulated the weights from the second (hidden) to third (output) layer. On account of random initialization, generalized performance of system may degrade. So that it is necessary to select the prominent features.
To distinguish said problem, in this paper incremental F-ELM (RFELM++) and incremental fuzzy online sequential ELM (RFOSELM++) for the UCP are proposed for batch mode and sequential input respectively. Thus, the key intent of this paper is to design RFELM++ and RFOSELM++ algorithm by using the activation function like radial basis function (RBF) and additive function which is missing in the vast literature survey. Another aspect of the paper is to design an IoT based expert system - Precision Agriculture by exploiting the proposed algorithms.
The overall structure of the paper is the following: in Section 2, the literature survey is detailed in the related work. The proposed incremental F-ELM and RFOSELM++ algorithms are depicted in Section 3. The experimental observations and results of clinical datasets are discussed and compared in Section 4. The validation and testing part of the proposed algorithm are discussed in detail in Section 5 as the results and discussions. Section 6 outlined the conclusion with further research direction.
In various growing areas FSS plays a very critical task. The key role of FSS is to choose prominent/optimal (relevant and non-redundant) feature subset which produces similar or improved (in some of the cases) results as compared to the prime or base data set in which all features are present. It is very hard to decide the significance of attributes without preparatory information [8]. Therefore, FSS is an prime step to select prominent features for efficient machine learning applications [9, 10]. The primary intention of using any FSS techniques is for [11]:
• Improvement in classification accuracy,
• Improvement in prediction performance,
• Improvement in model interpretability,
• Dimensionality reduction,
• Storage management.
The few of the FSS techniques which are mostly used are: filter, wrapper, and hybrid.
Filter techniques rank the features or select feature subset independently of the learning algorithm. Filter techniques are easy to interpret and fast. The characteristics of filter techniques are considered independently, totally independent of which learning algorithm is used etc. For class separability, the evaluation criteria is required which is a classical criterion of the filtering present in the literature. The criteria like distance, information and reliant are mostly utilize to classify the features.
Wrapper techniques use a learning algorithm for FSS, training is required for every feature subset. So, wrapper techniques are computationally heavy and dependent on the type of classifier used. The characteristics of wrapper techniques are computationally expensive, use only heuristic search and dependent on the learning algorithm.
Hybrid techniques use both filter and wrapper techniques. Ranking-wrapper techniques [12], Relief-F with sequential backward search [13] and boosting-based hybrid for feature selection (BBHFS) [14] are few of hybrid techniques.
Tahir and Loo [15] used Relief-F method for feature selection in continual learning framework. Various optimization algorithms like particle swarm intelligence [16] is used to solve feature selection problem and multi-objective particle swarm optimization algorithm [17] is used to solve bioprocess application problems and tumor treatment problems.
Pattern classification is specially used for decision making task. It classifies input entity to one of the predetermined target outputs. A supervised classification is specially used to categorised the given input data into one of the present target classes [18]. Having the data
The original ELM classifier is basically designed for a batch mode. ELM is acknowledged that the complete input data is available previously. However, many real applications require to handle the data sequentially. Therefore, Nan-Ying Liang et. al. have designed OS-ELM algorithm for handling the sequent or serial input. OS-ELM randomly assigns all the weight values in-between first (input) and second (hidden) layer and the weight values in between second (hidden) to third (output layer) are analytically tuned [23–25] as demonstrated in Figure 1 [26]. The global generalization performance may degrade due to the existence of non-prominent attributes and random initialization. Therefore, it is necessary to choose a prominent feature subset for sequential input. However, ELM is insufficient to provide a solution for the sequential and UCP.
OS-ELM is exploited in various applications like estimation of hematocrit [27], DWT domain watermarking [28], traffic profiling [29], etc. However, the literature survey lacks the exploitation of incremental learning ability by using prominent or optimal feature subset selection for sequent or serial input.
Traditional ELM is unable to solve some real-time time problems like the UCP and imbalance problem [1]. So, F-ELM is designed by Zhang and Ji [1] to solve specially UCP. However, in order to solve the UCP for sequential input OS-ELM is designed [30].
Ontiveros-Robles et al. [31] used type II fuzzy classification for non-singleton and also designed hybrid classification by using SVM [32]. Rubio et al. [33] used fuzzy clustering for unsuprvised method.
Incremental learning is used in various ways like incrementally add hidden nodes [34, 35], incrementally add streaming data [36], incremental FSS [37–40], etc. The exploitation of incremental learning in terms of incremental FSS for sequential input is also missing in the literature survey.
Incremental learning comes in various forms and definitions as
• Incrementally add streaming data: Incremental learning is a learning for the data which appear sequentially or over time [36, 41]. Alexander Gepperth describes online learning, incremental learning and concept drift for the supervised learning paradigm in detail.
• Incrementally add hidden nodes: Incremental learning defines as an incrementally insertion of the new neurons to the hidden layer. Error-minimized ELM (EM-ELM) [42], Convex incremental ELM (CI-ELM) [35] and uncremental extreme learning machine (RELM++) [34] are some of the examples of incremental learning. Yang et al. [43] designed a bidirectional extreme learning machines (B-ELM) algorithm for regression.
• Incremental feature subset selection: Incremental learning in the form of incremental feature subset selection is used in many papers like [37–40]. Baluja [44] have developed a population-based incremental learning method which integrated learning and the random search based function optimization [44]. Tahir and Loo [15] used an incremental ELM is used for food recognition. Wang et al. [45] have designed incremental wrapper based gene selection with Markov blanket. Incremental wrapper subset selection (IWSS) uses a filter measure for feature ranking by using a sequential search [46]. Then, a SFS is used for incrementally insertion of the features.
Wang et al. [47] designed a framework for smart contracts based on a novel six-layer architecture.
This section is categorized into two subsections, RFELM++ and RFOSELM++ algorithms.
In this section, incremental learning is used to resolve the uncertainty classification problem. As already discussed in Section 2, F-ELM specifically uses for the uncertainty classification problem. Hence, RFELM++ algorithm is initiated. The paradigm of RFELM++ is broadly categorised into four subsystems: input, incremental learning for FSS, fuzzification, and classification.
In this section, the detailed description of the input is described in detail. In order to resolve the UCP, it is necessary to use the datasets in which uncertainty is present. Therefore, Statlog Heart Disease (SHD) and Pima Indian Diabetes (PID) datasets are used [48]. The PID data set contains total 768 instances and 8 attributes. The datasets consists of several medical predictor variables and one target variable, outcome. Predictor variables includes the number of pregnancies the patient has had, their body mass index (BMI), insulin level, age, and so on. The SHD data set contains 270 instances and 13 attributes. Additionaly, the Cleveland Heart Disease (CHD) dataset is used for multiclass classification.
The generalization performance of any system is depend on the input. If the input to the system is right the performance of system gets increased. Therefore, it is very important task to provide a right or correct input features to the system.
In data preprocessing FSS algorithm is used. To select only prominent features relief algorithm is used. Relief algorithm is one of the FSS algorithms based on distance measure. The measure of distance (D) is basically used to search the features which can differentiated between two target classes. If the D is equal to 0, then the two attributes are identical. The Relief algorithm is used the joint relationship to compute weights with the output [49, 50] which is famous due to it’s simplicity and effectiveness. In Relief (feature weight-based algorithm), a statistical method is used to choose the relevant attributes [51].
After applying relief algorithm on PID and SHD datasets, we get feature order. For example, PID dataset has 8 attributes like age, DPF, BMI, insulin, skin, BP, Glu, and pregnancy. After Relief algorithm we get 〈2, 6, 4, 8, 1, 7, 5, 3〉 feature order. With this order it is understood that 2 is the most required or more relevant feature whereas 3 is the least required feature.
As per already described in the previous section, prominent feature order is calculated by considering distance measures i.e., Relief algorithm. Suppose the attribute order for SHD and PID data sets is 〈13, 12, 3, 9, 10, 11, 2, 1, 7, 4, 5, 6, 8〉 and 〈2, 6, 8, 1, 7, 5, 4, 3〉 respectively. The organised attributes are incrementally inserted one by one by using SFS for sequential space. Experimental observation shows that from all the subsets {1, 2, 3, 4, 5, 7, 9, 10, 11, 12, 13} and {1, 2, 4, 5, 6, 7, 8} are the prominent feature subset for SHD and PID datasets, respectively, as it provides improved classification accuracy.
The attributes of PID and SHD datasets are translated into linguistic attributes (fuzzified features) with the help of trapezoidal membership function [52]. The PID (8 attributes) and SHD (13 attributes) datasets are translated into 25 and 39 linguistic attributes, respectively. Input dataset is distributed into two datasets like training with 70% and testing with 30 %. Hence for PID ans SHD datasets 577 and 177 instances are used for traing and 577 and 177 instances are used for testing, respectively. By using trapazoidal membership function the features are converted into fuzzified features. For classification ELM is used.
ELM is a batch learning mode classifier which is not suitable for Sequential input. In this section, incremental learning for FSS is used to solve the UCP as well as sequential problem. Hence, incremental FSS techniques for the UCP is proposed for sequential input which contributes to RFOSELM++ algorithm.
In RFOSELM++, the ranked features are inserted sequentially i.e., one by one. The accuracy is calculated for all feature subsets which are equal to the number of attributes. The proposed RFOSELM++ learning algorithm is applied for every subset and predicted accuracy is calculated. The paradigm of RFOSELM++ is mostly same as RFELM++. Only the classification subsystem is different one. Through observations, the prominent feature subset is one which provides maximum accuracy.
For experimental observation, Waikato Environment for Knowledge Analysis (WEKA) and MATLAB R2014a are used. Precision, Recall and classification accuracy are used as an evaluation measure. The calculations of evaluation measures are depend upon the contingency table or confusion matrix of classifier. Confusion matrix contains both true and false positives (TP/FP) also true and false negatives (TN/FN) [52]. FP correspond to negative samples falsely identified as positive. FN correspond to positive samples falsely resembles as negative. TP are the samples perfectly identified as positive. TN refer to negative samples perfectly identified as negative. The perfectness of identified examples are measured by using precision. The measure of the all positive identified examples is called recall [53]. Accuracy, precision and recall (sensitivity) are measured by evaluating the respective values in
In this section, the incremental learning capability for batch and sequential mode is exploited i.e., incremental learning for F-ELM (RFELM++) algorithm and incremental learning for fuzzy online sequential extreme learning machine, respectively.
The experimental results of the proposed RFELM++ algorithm are enumerated. In order to prove the effectiveness of an initiated algorithm the same steps are implemented for ELM classifier (RELM++). The results of both RELM++ and RFELM++ are compared with sigmoidal activity function by considering PID and SHD datasets. The overall comparison of RELM++ and RFELM++ for PID dataset is as shown in Figure 2. The same results are evaluated for SHD dataset also. Through the results, it is observed that RFELM++ produces 6.266% and 0.855% increased accuracy in comparison of RELM++ for PID and SHD datasets consequently.
The overall performance of the FELM classifier is analysed by considering the precision, accuracy and recall. Figure 3 shows an overall comparison of performance for ELM and F-ELM with all features. The results are calculated by using all activity functions like sigmoid function (sig), radial basis function (rbf), sine function (sin), hardlim function (hardlim), triangular basis function (tribas) for SHD and PID datasets. By observations it is apprise that, F-ELM issues an improved performance as compared to ELM classifier.
Various FSS methods like half selection (HS), mean selection (MS), particle swarm optimization (PSO) based FS, neural network for threshold, self-regulated learning (SR-PSO) [54] are used to select prominent features. Table 1 indicates the output of all given FSS methods with ELM and FELM classifiers. For evaluation, the sigmoidal activation function is used for PID dataset. From experimented results, it is noticed that the F-ELM achieves 4.65% improved classification accuracy in comparison with ELM for the same number of feature. The result for RFELM++ algorithm for multiclass classification problem for CHD dataset is as shown in Figure 5.
In order to prove the effectiveness of the proposed RFELM++ algorithm, experiment results are derived and analysed by using the traditional classifiers–MLP, random forest (RF), Bayes net (BN), SVM, NB, RBF, and J4.8–by using prominent features. The prominent features are selected by using Relief algorithm. Five-fold cross-validation method is used for evaluation. Figure 4 shows the variation of the proposed RFELM++ algorithm with existing classifiers by using relief FSS algorithm for PID and SHD data set. It is evidently prove that the proposed RFELM++ algorithm enables maximum predictive accuracy with minimum number of features.
PID dataset is used as an input dataset. Normalization is used in pre-process. The algorithm uses relief algorithm (distance measure) for ranking the features. IWSS [39, 40] is used which acquires a ranking of the features over to the target class. IWSS is used SFS search over the ranked attributes by incrementally inserting those features one by one. The wrapper way is used to measure the relevance of the new attribute. The classification accuracy is calculated by using OS-ELM classifier.
Experimental results of incremental attribute learning for sequential input is calculated for both ELM and F-ELM classifier. Table 2 shows the parameters required for the evaluation. Figures 6 and 7 show the comparison of RELM++ and RFELM++ for training and testing accuracy, respectively. Through observation, it is noticed that the proposed RFELM++ issues an improved as well as stable (constant) results as compared to RELM++.
In order to prove the validation of the proposed RFELM++ algorithm, experimental observations need to be validated and hypothetically tested [55]. F-ELM classifier is used after the feature selection by using statistical methods. The PID and SHD datasets are categorized into four samples like S4 (SHD ELM), S3 (PID ELM), S2 (SHD F-ELM) and S1 (PID F-ELM). The comparative analysis of before and after relief feature selection algorithm is as shown in Figure 8. Through the results it is noticed that, RFELM++ algorithm with feature subset produces improved performance as compared to with all features.
The correlation analysis technique is used to compute the robustness of the relationship between 2 or more variables. All the time the values of correlation are in between −1 to +1. The abbreviation of +1 is that the two variables are strongly related with positive linear, −1 is that two variables are resolutely related with negative linear and 0 is that no relationship in between the two variables. According to Evans classification, the correlation coefficient are broadly categorized into very strong, strong, moderate, weak, very weak [56].
The
where ‘
The coefficient (
When the relationship (
where ‘
Hypothesis of testing [57] RFELM++:
The
For performance comparison, accuracy of RFELM++ by using prominent feature subset is compared with F-ELM [59], modified fuzzy min-max neural network-FIS (MFMM-FIS) [60] and enhanced generalized adaptive resonance theory FIS (EGART-FIS) [61] for the PID dataset. Table 3 describes the comparative classification accuracy for the same. With the comparative analysis, it is observed that the testing and training accuracy is 83.47% to 86.80%, respectively by using only 62.5% features.
With the intention to evince the effectiveness of the proposed RFELM++ algorithm, it is mandatory to have comparison the result of RFELM++ with the similar approach like NN for threshold selection (NN) [62,63], multilayer NN (MLNN) by using Levenberg–Marquardt (LM) [64], generalized discriminant analysis least square (GDA-LS) [65], HS method, least square-ELM (LS-ELM) [65], linear discriminant analysis and Morlet wavelet (LDA-MW), probabilistic neural network (PNN) [64], MS methods. Table 4 shows RFELM++ and other classifiers comparison for PID input dataset. Table 5 illustrates the same comparison for SHD dataset with evolutionary product-unit NN (EPUNN), MS, evolutionary sigmoidal unit NN (ESUNN), HS, multilogistic regression by means of evolutionary product-unit NN (MR+EPUNN) for SHD dataset. Through comparative average analysis, it is noticed that the introduced RFELM++ issues 8.82% and 6.63% increased accuracy by reducing 6.41% and 10.71% for SHD adn PID datasets, respectively.
The comparative analysis in between F-ELM and ELM by using different FSS algorithms is as shown in Figure 9 for PID dataset. HS, PSO, MS, NN, T-test and self-regulated learning (SR-PSO) [54] are used as FSS algorithms. For PID dataset, the comparison of ELM and F-ELM classifier in terms of accuracy measure by using the feature subset as shown in Figure 9 in which F-ELM produces increased performance for the same number of features.
The introduced RFELM++ algorithm is exploited in the real time application which is represented as a case study in the following section.
Agriculture is the backbone of Indian economy where approximate 60% of people are dependent directly or indirectly on agriculture. The expert advice is required for distinguishing the plant disease damage and nutrient imbalance. It is observed that, the conventional judgmental analysis is not enough while deciding the quantity of chemical or fertilizer to be used. The mis-proportional dose harms directly the health of the crop and hence the living beings. To overcome the said problem, this section is developed an IoT-based precision agriculture expert system by inferring proposed algorithms with improving the accuracy by reducing the features or by selecting the prominent features. Grape leaf diseases like downy mildew and nutrient deficiency which contains ambiguity is considered as a case study. Both are spread with yellow spot on leaf. The conceptual framework for the proposed IoT-based precision agriculture expert system is as shown in Figure 10.
In IoT-based precision agriculture expert system, the system is partitioned into three rough phases namely image processing, FSS by using incremental learning and classification. In image processing, the acquisition step is used to input the image. The image is acquired by using cell phone or any digital camera. The resolution of the image is 96 dpi (dots per inch). For interfacing, the ARDUINO hardware is used with various sensors like temperature, humidity etc. The leaf image is used to extract the feature points and are stored in the database.
Data normalization is used as the preprocessing method. In image analysis, image segmentation and feature extraction methods are considered. For image segmentation, region-based segmentation is used to separate the healthy and diseased region by using leaf color. The features extracted from an images are color, shape and texture [71]. Once features are extracted, the database is generated with integrated information from temperature and humidity sensors. On that created dataset the proposed algorithms are applied for feature selection and classification.
In order to increase the generalization performance of the system, the proposed RELM++ and RFOSELM++ algorithms are exploited. The performance of precision agriculture is evaluated and tested by using RELM++ and RFOSELM++ algorithms. The comparison in-between RELM++ and ELM for proposed system (plant disease and nutrient deficiency) is as shown in Figure 11 in terms of testing accuracy. Through this experimentation, it is intended that the RELM++ produces 5.71% increased accuracy and RFOSELM++ maintains (same accuracy) the accuracy with only 27.27% features. However, for Tribas activation function the classical method provides good results as compared to the proposed one. Tribas provided an optimal result only for certain number of hidden neurons for ELM [72].
The prime intention of this paper is to exploit the two proposed algorithms like RFELM++ and RFOSELM++ into the new developed IoT-based precision agriculture Expert system. The expert system helps farmers in decision making. The experimental results for ELM, OS-ELM, RELM++, RFOSELM++, RFELM++ and RFOSELM++ are calculated. RFELM++ produces 6.266% and 0.855% increased accuracy in comparison of RELM++ for PID and SHD datasets consequently. Additionally, various FSS methods like like HS, MS, PSO based FS, neural network for threshold, self-regulated learning (SR-PSO) [54] are used to select prominent features. Through the experimented results, it is noticed that the F-ELM achieves 4.65% improved classification accuracy in comparison with ELM for the same number of feature. The proposed RFOSELM++ maintain the accuracy by using 41.5% features as differentiate to OS-ELM for UCI Repository dataset. For benchmark problem, the results of initiated RFELM++ algorithm is compared with the similar existing approaches. From the comparative average analysis, it is noticed that the introduced RFELM++ issues 8.82% and 6.63% increased accuracy by reducing 6.41% and 10.71% for SHD and PID datasets, respectively. Table 3 describes the comparative classification accuracy for the same. With the comparative analysis, it is observed that the testing and training accuracy is 83.47% to 86.80%, respectively by using only 62.5% features. Hypothetically it is proved that there is no variance between the generalization performance of F-ELM (with all features or before feature selection) and RFELM++ (after feature selection) for four samples. A novel application, IoT-based precision agriculture expert system, is developed by exploiting the proposed algorithms RFELM++ and RFOSELM++. Currently, the algorithms are used for binary classification in future it can be used for multi-class classification. The precision system can be improved by considering the rain removal [73] and by using CNN [74].
Extreme learning machine.
RELM++ and RFELM++ by using relief algorithm.
Comparison of ELM and F-ELM: (a) accuracy of PID dataset, (b) accuracy of SHD dataset, (c) precision of PID dataset, (d) precision of SHD dataset, (e) recall of PID dataset, and (f) recall of SHD dataset.
Accuracy measure comparison of F-ELM with various classifiers by using Relief-FSS algorithm: (a) input PID dataset and (b) input SHD dataset.
Evaluation performance: (a) CHD accuracy, (b) CHD precision, and (c) CHD recall.
ELM and F-ELM performance comparison with incremental learning for training.
ELM and F-ELM performance comparison with incremental learning for testing.
Performance metrics of RFELM++ before and after using Relief-FSS algorithm.
F-ELM and ELM in combination with different FSS methods for PID dataset.
Conceptual framework for the IoT-based precision agriculture expert system.
ELM and RELM++ (testing accuracy) for plant disease and nutrient deficiency.
Table 1 . Comparative results of ELM and F-ELM with different FS methods for input PID dataset.
FSS Methods | ELM | F-ELM | Feature subset |
---|---|---|---|
MS | 78.26 | 79.13 | 2, 6, 8 |
HS; l | 76.08 | 82.6 | 1, 2, 6, 8 |
NN for threshold selection | 78.26 | 82.6 | 1, 2, 6, 8 |
PSO | 78.26 | 82.6 | 1, 2, 6, 8 |
SRLPSO | 78.26 | 79.13 | 2, 6, 8 |
76.52 | 83.47 | 1, 2, 6, 7, 8 | |
Without FSS Method | 71.73 | 80.43 | All |
Avg | 76.76 | 81.42 | - |
Table 2 . Parameters for RFELM++.
Parameter | Features |
---|---|
Input neuron | Same as input features |
Output neuron | 2 |
Hidden neuron | 10:5:200 |
Activity function | Sigmoidal |
Sequential mode | 1 by 1, 20 by 20, 10 by 30 |
Dataset division | Training (70%)-Testing (30%) |
Table 3 . Comparison of RFELM++ with existing learning algorithms for input PID dataset.
Methods/algorithm | Accuracy of training | Accuracy of testing | Total number of attributes used |
---|---|---|---|
RFELM++ | 86.80 | 83.47 | 05 (62.5%) |
F-ELM [59] | 75.35 | 74.09 | 08 (100%) |
MFMM-FIS [60] | n/a | 72.92 | 08 (100%) |
EGART-FIS [61] | n/a | 73.05 | 08 (100%) |
MFMM-FIS [62] | n/a | 72.92 | 08 (100%) |
Table 4 . RFELM++ and existing learning algorithms comparison for input PID dataset.
Method/algorithm | Accuracy (testing) | Total number of attributes used | Feature subset |
---|---|---|---|
RFELM++ | 83.47 | 05 | {1,2,6,7,8} |
RELM++ | 76.52 | 05 | {1,2,6,7,8} |
EGART-FIS [61] | 73.05 | 08 | {1,2,3,4,5,6,7,8} |
F-ELM [59] | 74.09 | 08 | {1,2,3,4,5,6,7,8} |
MFMM-FIS [60] | 72.92 | 08 | {1,2,3,4,5,6,7,8} |
LS-ELM [65] | 78.21 | 04 | {1,2,6,8} |
GDA-LS-ELM [65] | 79.16 | 05 | {1,2,6,7,8} |
MLNN with LM [64] | 79.62 | 04 | {1,2,6,8} |
PNN [64] | 78.05 | 03 | {2,6,8} |
MS [62] | 76.04 | 03 | {2,6,8} |
HS [62] | 75.91 | 04 | {1,2,6,8} |
NN [62] | 76.04 | 03 | {2,6,8} |
GP-KNN [66] | 80.50 | 08 | {1,2,3,4,5,6,7,8} |
Gini-Fuzzy [67] | 75.80 | 08 | {1,2,3,4,5,6,7,8} |
FMNN-CART-RF [25] | 78.39 | 08 | {1,2,3,4,5,6,7,8} |
Generic statistical approach [68] | 78.04 | 08 | {1,2,3,4,5,6,7,8} |
76.84 | 26.79 | - |
Table 5 . RFELM++ and existing learning algorithms comparison for Input SHD dataset.
Method/algorithm | Accuracy (testing) | Number of features | Feature subset |
---|---|---|---|
RFELM++ | 92.59 | 06 | {3,9,10,11,12,13} |
RELM++ | 90.74 | 05 | {3,9,10,12,13} |
ESUNN [69] | 83.22 | 05 | {3,8,9,11,12} |
EPUNN [69] | 81.89 | 04 | {8,9,11,12} |
MR+EPUNN [70] | 83.12 | 05 | {8,9,11,12,13} |
MS [62] | 84.44 | 06 | {3,8,9,11,12,13} |
HS [62] | 84.81 | 07 | {3,8,9,10,11,12,13} |
NN [62] | 85.19 | 04 | {3,11,12,13} |
83.77 | 60.26 | - |
Table 6 . Evaluation comparative analysis for ELM and F-ELM with different FSS for input PID dataset.
FSS Method | ELM | F-ELM | Feature subset |
---|---|---|---|
LS [65] | 78.21 | 82.6 | {1,2,6,8} |
GDA-LS [65] | 79.16 | 82.17 | {1,2,6,7,8} |
MLNN with LM [64] | 79.62 | 82.6 | {1,2,6,8} |
PNN [64] | 78.05 | 79.13 | {2,6,8} |
LDA-MW [62] | 89.74 | 82.17 | {1,2,6,7,8} |
MS [62] | 76.04 | 79.13 | {2,6,8} |
HS [62] | 75.91 | 82.6 | {1,2,6,8} |
NN [62] | 76.04 | 79.13 | {2,6,8} |
Jung-Sook Kim and Junho Jeong
International Journal of Fuzzy Logic and Intelligent Systems 2019; 19(4): 342-348 https://doi.org/10.5391/IJFIS.2019.19.4.342Minyoung Kim
Int. J. Fuzzy Log. Intell. Syst. 2017; 17(1): 10-16 https://doi.org/10.5391/IJFIS.2017.17.1.10Jae Hoon Cho, Dae-Jong Lee, Jin-Il Park, and Myung-Geun Chun
Int. J. Fuzzy Log. Intell. Syst. 2013; 13(1): 69-77 https://doi.org/10.5391/IJFIS.2013.13.1.69Extreme learning machine.
|@|~(^,^)~|@|RELM++ and RFELM++ by using relief algorithm.
|@|~(^,^)~|@|Comparison of ELM and F-ELM: (a) accuracy of PID dataset, (b) accuracy of SHD dataset, (c) precision of PID dataset, (d) precision of SHD dataset, (e) recall of PID dataset, and (f) recall of SHD dataset.
|@|~(^,^)~|@|Accuracy measure comparison of F-ELM with various classifiers by using Relief-FSS algorithm: (a) input PID dataset and (b) input SHD dataset.
|@|~(^,^)~|@|Evaluation performance: (a) CHD accuracy, (b) CHD precision, and (c) CHD recall.
|@|~(^,^)~|@|ELM and F-ELM performance comparison with incremental learning for training.
|@|~(^,^)~|@|ELM and F-ELM performance comparison with incremental learning for testing.
|@|~(^,^)~|@|Performance metrics of RFELM++ before and after using Relief-FSS algorithm.
|@|~(^,^)~|@|F-ELM and ELM in combination with different FSS methods for PID dataset.
|@|~(^,^)~|@|Conceptual framework for the IoT-based precision agriculture expert system.
|@|~(^,^)~|@|ELM and RELM++ (testing accuracy) for plant disease and nutrient deficiency.