Extreme Learning Machine: A Review
Extreme Learning Machine: A Review
4610-4623
© Research India Publications. http://www.ripublication.com
a
CAIT, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi, Selangor, Malaysia.
4610
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
4611
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
Feature learning
β1 βi βL
Clustering
Problem based Regression
optimization constraints Classification
1 .. d d Input nodes
Xj
Figure 2: ELM architecture; the hidden nodes in ELM can be combinatorial nodes that are made up of different types of
computational nodes (Huang & Chen 2007)
Table 1: Commonly used mapping functions in ELM utilizes kernel functions to map features, or deep neural
networks (Bengio 2009), which utilise Auto-Encoders/Auto-
Decoders or Restricted Boltzmann Machines (RBM) for
𝟏 feature learning. In ELM, the nonlinear mapping functions
𝐆(𝐚, 𝐛, 𝐱) = can be any of the nonlinear piecewise continuous functions.
Sigmoid function 𝟏 + 𝐞𝐱𝐩(−(𝐚・𝐱 + 𝐛))
Table 1 shows some of the most commonly used ones.
1 − exp(−(a・x + b))
Hyperbolic tangent G(a, b, x) =
1 + exp(−(a・x + b))
function
Multiquadric
G(a, b, x) = (∥ x − a ∥ + b2)1/2
function
4612
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
Basic Extreme Learning Machine is done using the least-squares solution. Furthermore, they
have been defined by the outputs of the targets and the hidden
(Huang et al. 2006) proposed the original ELM algorithm that
layer. Figure 4 shows an overview of the training algorithm
can be used to train SLFN. In ELM, the main idea involves
and the ELM structure. A brief description of ELM will be
the hidden layer weights. Furthermore, the biases are
given in the next section.
randomly generated and the calculation of the output weights
Input xj1
2
Input xj2
3 Output tj
Input xj3
...
4
Input xjn
...
𝐿 𝐿
∑ 𝛽𝑖 𝑔𝑖 (𝑋𝑗 ) = ∑ 𝛽𝑖 𝑔𝑖 (𝑊𝑖 . 𝑋𝑗 + 𝑏𝑖 ) (3) From the equations given above for N, it can then be
𝑖=1 𝑖=1 presented as follows:
J = 1… N. H𝛽 =T (5)
Where: Where:
Wi = [Wi1, Wi2… Win] T represents the weight vector that H (W1… WL, b1…bL, X1… XN)
connects the hidden node and the ith input nodes.
𝛽 𝑖 = [𝛽 𝑖1, 𝛽 𝑖2,……, 𝛽 𝑖𝑚] Trepresents the weight vector that g(W1 . X1 + b1 ) … g(WL . X1 + bL )
connects the hidden node and the ith output nodes.
= …
…
4613
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
Equation (5) then turns into a linear system. Furthermore, the Pruning ELM
output weights β can be determined analytically by
The use of too few/many hidden nodes may result into issues
discovering a least square solution in the following way:
of underfitting/overfitting in pattern classification.(Rong et al.
𝛽 = 𝐻†T 2008) developed a pruned-ELM (P-ELM) algorithm that
provides an automated and systematic way to design an ELM
Where 𝐻 † is represents the Moore–Penrose generalised
network. P-ELM starts with a large amount of hidden nodes
inverse for H. Thus, the output weights are calculated via a
before it gets rid of the lowly relevant or irrelevant hidden
mathematical transformation. This makes sure that the lengthy
nodes by taking into account their relevance to the class labels
training phrase when network parameters are iteratively
in the learning process. As a result, one can automate the
adjusted with some suitable learning parameters (like
architectural design of ELM. Simulation results revealed that
iterations and learning rate) is done away with.
the P-ELM resulted in compact network classifiers that are
Huang et al. (2006) enumerated the variables, where H capable of generating robust prediction accuracy and fast
represents the output matrix of the neural network’s hidden response on unseen data compared to the standard BP, ELM,
layer; in H, the ith column is used to describe the ith hidden and MRAN. P-ELM is mostly suitable for pattern
layer nodes in terms of the input nodes. If L ≤ N represents classification problems. Given the fact that too few/many
the desired number of hidden nodes, the activation function g hidden nodes can result into issues of underfitting/overfitting
becomes infinitely differentiable. in pattern classification, Rong et al. (2008) proposed a pruned-
ELM (P-ELM) algorithm. This algorithm is a systematic and
VARIANTS OF ELM
automated way to design an ELM network. P-ELM starts with
This section summarises and briefly introduces several typical a large amount of hidden nodes before it gets rid of the lowly
variants of ELM. relevant or irrelevant hidden nodes by taking into account
their relevance to the class labels in the learning process. As a
result, one can automate the architectural design of ELM.
Incremental ELM Simulation results revealed that the P-ELM resulted in
(Huang et al. 2006) developed an Incremental Extreme compact network classifiers that are capable of generating
Learning Machine (I-ELM) to create a feedforward network robust prediction accuracy and fast response on unseen data
that is incremental. I-ELM added nodes randomly to the compared to the standard BP, ELM, and MRAN. P-ELM is
hidden layer. This addition was done one by one. It then froze mostly suitable for pattern classification problems.
the existing hidden nodes’ output weights during the addition
of a new hidden node. I-ELM is efficient not only for SLFN
Error-minimised ELM
having continuous activation functions (as well as
differentiable), but they are also efficient for SLFNs that have (Feng et al. 2009) developed an error-minimisation-based
piecewise continuous activation functions (like threshold). method that can be used for ELM (EM-ELM). This method is
Given this context of I-ELM, Huang et al. presented the able to grow hidden nodes group by group or one by one and
convex I-ELM (CI-ELM) and enhance I-ELM (EI-ELM). automatically know the amount of hidden nodes that can be
Unlike I-ELM, CI-ELM (Huang & Chen 2007) recalculates found in generalised SLFNs. During network growth,
the existing hidden nodes’ output weights when a new hidden updating of the output weights is done incrementally, which
node is added. Compared to I-ELM, CI-ELM is able to pointedly lowers the computational complexity. For sigmoid
achieve more compact network architectures and faster type hidden nodes, the simulation results revealed that this
convergence rates while retaining the efficiency and technique could significantly lower ELM’s computational
simplicity of I-ELM. EI-ELM (Huang & Chen 2008) allows complexity and help formulate an efficient ELM
for the maximum amount of hidden nodes. Furthermore, users implementation.
need not set control parameters manually. Unlike the original
I-ELM, EI-ELM chooses the optimal hidden node. This
means that the smallest residual error is obtained at every Two-stage ELM
learning step among several of the hidden nodes that were
To achieve a parsimonious solution for the preliminary
randomly generated. EI-ELM is able to achieve a much more
ELM’s network structure, (Lan et al. 2010) proposed a
compact network architecture and faster convergence rate.
4614
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
systematic two-stage algorithm (named TS-ELM). During the called the evolutionary extreme learning machine (E-ELM) to
first stage, they applied a forward recursive algorithm to optimize the hidden biases and input weights and determine
choose the hidden nodes from the randomly generated the weights of the output. In E-ELM, the input weights and
candidates in each step. These hidden nodes were then added hidden biases were optimised using the modified differential
to the network until they met the stopping criterion. evolutionary (DE) algorithm. The output weights were
Consequently, each hidden node’s significance was analytically determined using the Moore– Penrose (MP)
determined by their net contribution after being added to the generalised inverse. Experimental results revealed that E-
network. During the second stage, a review of the selected ELM was capable of achieving good generalisation
hidden nodes was done in order to get rid of the unimportant performance that has more compact networks and which is
nodes in the network. This step significantly reduced the superior to other algorithms like GALS, BP, and the original
complexity of the network. The empirical studies conducted ELM.
on the six cases revealed that TS-ELM having a significantly
smaller network structure may be able to achieve similar or
better performance than EM-ELM. Voting-based ELM
Since the hidden nodes in ELM’s learning parameters are
randomly assigned and stay the same during the training
Online sequential ELM
process, ELM may not achieve the optimal classification
When using the conventional ELM, all of the training data boundary. As such, the samples nearest to the classification
must be available for training purposes. However, the training boundary run the risk of being misclassified. Therefore, Cao
data in real applications may be obtained one by one or chunk et al. (2012) developed an improved algorithm referred to as
by chunk. (Liang et al. 2006) proposed a sequential learning the voting-based extreme learning machine (V-ELM). The
algorithm called the online sequential extreme learning aim of this algorithm is to lessen the amount of misclassified
machine (OS-ELM). This algorithm is able to work with both samples that are found near the classification boundary. In V-
RBF and additive nodes in a unified framework. OSELM that ELM, the main idea is to conduct multiple independent ELM
has additive nodes randomly generates the input weights that trainings rather than just performing a single ELM training
connect the hidden nodes and biases to the input nodes. It then and taking a final decision that is based on the results of the
analytically determines the output weights based on the majority voting method (Cao et al. 2012). V-ELM was able to
hidden nodes’ output. Unlike the other kinds of sequential improve the classification performance, lessen the amount of
learning algorithms, the OS-ELM only needs the specification misclassified samples, and reduce variance among the
for the number of hidden nodes. This is also similar to the different realisations. Based on the simulations conducted on
conventional ELM. To enhance the OSELM’s performance numerous real-world classification datasets, it was observed
and acquaint the ensemble networks with the sequential that V-ELM generally performed better than the original ELM
learning mode, (Lan et al. 2009) developed an integrated algorithm and even other recent classification algorithms.
network structure that is referred to as the ensemble of online
sequential extreme learning machine (EOS-ELM). EOS-ELM
is made up of several OS-ELM networks. For these networks, Ordinal ELM
the network performance’s final measurement is computed
To study the ELM algorithm further for ordinal regression
based on the average value of the outputs for every OS-ELM
problems, (Deng et al. 2010) introduced three ELM-based
in the ensemble. Furthermore, to show the training data’s
ordinal regression algorithms and an encoding-based ordinal
timeliness in the learning process,(Zhao et al. 2012)
regression framework. The paper developed an encoding-
developed an improved EOS-ELM called the online
based framework that can be used for ordinal regression and
sequential extreme learning machine with forgetting
that contained three encoding schemes: multiple binary
mechanism (FOS-ELM). This algorithm can retain EOS-
classifications having a one-against-all decomposition
ELM’s advantages and enhance its learning effects by quickly
method, single multi-output classifier, and one-against-one
getting rid of the out-dated data during the learning process in
method. The framework was used as the basis for the
order to lessen their bad affection to the next learning process.
redesigning of the SLFN for ordinal regression problems.
Extreme learning machine was then used to train the
algorithms. Experiments conducted on the three types of
Evolutionary ELM
datasets revealed that ordinal ELM is capable of achieving
Typically, the amount of hidden neurons is randomly good generalisation ability and extremely rapid training speed.
determined during the application of ELM. However, ELM
may require higher amounts of hidden neurons as a result of
the random determination of the hidden biases and input
weights. (Zhu et al. 2005) proposed a novel learning algorithm
4615
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
4616
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
with that of SVM. They found out that ELM has higher ELM for image processing
accuracy and faster execution. (Han et al. 2014) also
ELM is also considered an attractive technique for image
attempted to identify the speaker’s emotion using extreme
processing. For instance, (An & Bhanu 2012) introduced an
learning machine as classifier and DNN as feature extractor.
efficient image super resolution method that has its basis in
They found that ELM and Kernel ELM (KELM), when
ELM. The aim of their approach is to generate high resolution
combined with DNN, have the highest accuracies compared to
images from inputs with low-resolution. In the training
all the other baseline approaches. (Muthusamy et al. 2015)
process, the input was extracted from the image features.
utilised ELM with another classifier on various types of
Furthermore, the high frequency components that were taken
audio-related classification problems. They also addressed
from the original images with high-resolution were utilized as
emotion recognition based on the speaker’s audio. They used
the target values. Then, ELM learns a model that is capable of
GMM model features as inputs for the classifier. The authors
mapping the interpolated image and imposing it on the high-
stress the power of GMM-based features in offering
frequency components. Once training is done, the learned
discriminative factors that can be used to classify emotions.
model can predict the high-frequency components using low-
resolution images. (Li et al. 2013) used ELM to burn state
recognition of rotary kiln. (Chang et al. 2010) used ELM for
ELM for medical/biomedical applications
change detection of land cover and land use. Moreover, ELM
Medical or biomedical data typically have high dimensional was utilised for image classification by (Cao et al. 2013; Bazi
features or a large amount of samples. Thus, medical or et al. 2014). ELM was used to assess the perceived image
biomedical data analysis often utilise advanced machine quality (Decherchi et al. 2013; Suresh et al. 2009). ELM was
learning techniques like SVM, Since ELM offers many also utilised in the detection of semantic concept for videos
advantages compared to other learning algorithms, its (Lu et al. 2013). Image deblurring can also be done using
application in this area could be an interesting thing to see. filters that are learned by ELM (Wang et al. 2011). SAE-ELM
Indeed, many encouraging results on the application of ELM was utilised in the coal mine water inrush’s multi-level
to predict protein–protein interactions (You et al. 2013), EEG- forecasting model (Zhao & Hu 2014).
based vigilance estimation(Shi & Lu 2013),epileptic EEG
patterns recognition (Yuan et al. 2011; Song et al. 2012; Song
& Zhang 2013), transmembrane beta-barrel chains detection ELM for system modelling and prediction
(Savojardo et al. 2011), an eye-control method for eye-based
Because traditional neural networks have had wide uses in
computer interaction (Barea et al. 2012), spike sorting with
system prediction and modelling, ELM also has great
overlap resolution that is based on a hybrid noise-assisted
potential in the development of accurate and efficient models
methodology (Adamos et al. 2010), lie detection (Gao et al.
for these applications. (Xu et al. 2013) proposed an ELM-
2013), an electrocardiogram ECG (Karpagachelvi et al. 2012),
based predictor that can be used in the actual frequency
liver parenchyma segmentation (Huang et al. 2012), diagnosis
stability assessment (FSA) of power systems. The predictor’s
of hepatitis (Kaya & Uyar 2013), thyroid (Li et al. 2012),
inputs are the power system operational parameters, while the
arrhythmia classification in ECG (Kim et al. 2009), detection
output is set as the frequency stability margin. This margin
of mycobacterium tuberculosis in tissue sections (Osman et al.
measures the power system’s stability degree, subject to a
2012), protein secondary structure prediction (Saraswathi et
contingency. Using off-line training and a frequency stability
al. 2012), and metagenomics taxonomic classification
database, one can apply the predictor online for real-time
(Rasheed & Rangwala 2012) have been observed in recent
FSA. They tested the predictor on New England’s 10-
years.
generator 39-bus test system. The results of this simulation
revealed that it is capable of accurately and efficiently
predicting the frequency stability. ELM was also utilised for
ELM for computer vision
electricity price forecasting (Chen et al. 2012), sales
ELM has had successful applications in various computer forecasting (Wong & Guo 2010; Chen & Ou 2011),
vision tasks, such as human action recognition (Minhas et al. temperature prediction of molten steel (Tian & Mao 2010),
2010; Minhas et al. 2012), face recognition (Mohammed et al. security assessment of wind power system (Xu et al. 2012; Xu
2011; Zong & Huang 2011; Choi et al. 2012; Baradarani et al. et al. 2012), drying system modelling (Balbay et al. 2012), etc.
2013; Marqués & Graña 2013; He et al. 2014), terrain-based Because of its notable advantages, many other applications
navigation (Kan et al. 2013), and matching of fingerprints have adopted ELM. Based on past literature, we can witness
(Yang et al. 2013). its successful applications in control system design, text
analysis, chemical process monitor, feature selection,
clustering, ranking, and representational learning.
Furthermore, ELM can be applied to more potential fields as
well.
4617
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
4618
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
Rodríguez-Ascariz 2012. EOG-based eye [21] Cortes, C. & V. Vapnik 1995. Support vector
movements codification for human computer machine. Machine learning20(3): 273-297.
interaction. Expert Systems with Applications39(3): [22] Daliri, M. R. 2012. A hybrid automatic system for
2677-2683.
the diagnosis of lung cancer based on genetic
[8] Bazi, Y., N. Alajlan, F. Melgani, H. AlHichri, S. algorithm and fuzzy extreme learning machines.
Malek & R. R. Yager 2014. Differential evolution Journal of medical systems36(2): 1001-1005.
extreme learning machine for the classification of [23] Decherchi, S., P. Gastaldo, R. Zunino, E. Cambria &
hyperspectral images. IEEE Geoscience and Remote
J. Redi 2013. Circular-ELM for the reduced-
Sensing Letters 11(6): 1066-1070.
reference assessment of perceived image quality.
[9] Bengio, Y. 2009. Learning deep architectures for AI. Neurocomputing102: 78-89.
Foundations and trends® in Machine Learning2(1): [24] Deng, J., K. Li & G. W. Irwin 2011. Fast automatic
1-127.
two-stage nonlinear model identification based on
[10] BenoíT, F., M. Van Heeswijk, Y. Miche, M. the extreme learning machine.
Verleysen & A. Lendasse 2013. Feature selection for Neurocomputing74(16): 2422-2429.
nonlinear models with extreme learning machines. [25] Deng, W.-Y., Z. Bai, G.-B. Huang & Q.-H. Zheng
Neurocomputing102: 111-124.
2016. A fast SVD-hidden-nodes based extreme
[11] Bodyanskiy, Y., O. Vynokurova, I. Pliss, G. Setlak & learning machine for large-scale data analytics.
P. Mulesa 2016. Fast learning algorithm for deep Neural Networks77: 14-28.
evolving GMDH-SVM neural network in data [26] Deng, W.-Y., Q.-H. Zheng, S. Lian, L. Chen & X.
stream mining tasks. Data Stream Mining &
Wang 2010. Ordinal extreme learning machine.
Processing (DSMP), IEEE First International
Neurocomputing74(1): 447-456.
Conference on. pp. 257-262.
[27] Ding, S., X. Xu & R. Nie 2014. Extreme learning
[12] Branke, J. 1995. Evolutionary algorithms for neural
machine and its applications. Neural computing &
network design and training. In Proceedings of the
applications25.
First Nordic Workshop on Genetic Algorithms and
its Applications. [28] Du, D., K. Li, G. W. Irwin & J. Deng 2013. A novel
automatic two-stage locally regularized classifier
[13] Cao, F., B. Liu & D. S. Park 2013. Image
construction method using the extreme learning
classification based on effective extreme learning
machine. Neurocomputing102: 10-22.
machine. Neurocomputing 102: 90-97.
[29] Ebtehaj, I., H. Bonakdari & S. Shamshirband 2016.
[14] Cao, J., Z. Lin & G.-b. Huang 2010. Composite Extreme learning machine assessment for estimating
function wavelet neural networks with extreme
sediment transport in open channels. Engineering
learning machine. Neurocomputing73(7): 1405-1416.
with Computers32(4): 691-704.
[15] Cao, J., Z. Lin, G.-B. Huang & N. Liu 2012. Voting [30] Feng, G., G.-B. Huang, Q. Lin & R. Gay 2009. Error
based extreme learning machine. Information minimized extreme learning machine with growth of
Sciences185(1): 66-77.
hidden nodes and incremental learning. IEEE
[16] Chang, N.-B., M. Han, W. Yao, L.-C. Chen & S. Xu Transactions on Neural Networks20(8): 1352-1357.
2010. Change detection of land use and land cover in [31] Fernández-Delgado, M., E. Cernadas, S. Barro, J.
an urban region with SPOT-5 images and partial Ribeiro & J. Neves 2014. Direct Kernel Perceptron
Lanczos extreme learning machine. Journal of
(DKP): Ultra-fast kernel ELM-based classification
Applied Remote Sensing4(1): 043551.
with non-iterative closed-form weight calculation.
[17] Chen, F. & T. Ou 2011. Sales forecasting system Neural Networks50: 60-71.
based on Gray extreme learning machine with
[32] Gao, J., Z. Wang, Y. Yang, W. Zhang, C. Tao, J.
Taguchi method in retail industry. Expert Systems
Guan & N. Rao 2013. A novel approach for lie
with Applications38(3): 1336-1345.
detection based on F-score and extreme learning
[18] Chen, S., C. F. Cowan & P. M. Grant 1991. machine. PloS one8(6): e64704.
Orthogonal least squares learning algorithm for
[33] Hagan, M. T. & M. B. Menhaj 1994. Training
radial basis function networks. IEEE Transactions on
feedforward networks with the Marquardt algorithm.
neural networks2(2): 302-309.
IEEE transactions on Neural Networks5(6): 989-
[19] Chen, X., Z. Y. Dong, K. Meng, Y. Xu, K. P. Wong 993.
& H. Ngan 2012. Electricity price forecasting with
[34] Han, K., D. Yu & I. Tashev 2014. Speech emotion
extreme learning machine and bootstrapping. IEEE
recognition using deep neural network and extreme
Transactions on Power Systems27(4): 2055-2062.
learning machine. Interspeech. pp. 223-227.
[20] Choi, K., K.-A. Toh & H. Byun 2012. Incremental [35] He, B., D. Xu, R. Nian, M. van Heeswijk, Q. Yu, Y.
face recognition for large-scale social network
Miche & A. Lendasse 2014. Fast Face Recognition
services. Pattern Recognition45(8): 2868-2883.
4619
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
Via Sparse Coding and Extreme Learning Machine. approach to the segmentation of liver parenchyma
Cognitive Computation6(2): 264-277. from 3D CT images with Extreme Learning
[36] He, Q., C. Du, Q. Wang, F. Zhuang & Z. Shi 2011. A Machine. Engineering in Medicine and Biology
Society (EMBC), 2012 Annual International
parallel incremental extreme SVM classifier.
Conference of the IEEE. pp. 3752-3755.
Neurocomputing74(16): 2532-2540.
[37] Horata, P., S. Chiewchanwattana & K. Sunat 2013. [51] Kan, E. M., M. H. Lim, Y. S. Ong, A. H. Tan & S. P.
Robust extreme learning machine. Yeo 2013. Extreme learning machine terrain-based
navigation for unmanned aerial vehicles. Neural
Neurocomputing102: 31-44.
computing & applications: 1-9.
[38] Huang, G.-B. 2014. An insight into extreme learning
machines: random neurons, random features and [52] Karpagachelvi, S., M. Arthanari & M. Sivakumar
kernels. Cognitive Computation6(3): 376-390. 2012. Classification of electrocardiogram signals
with support vector machines and extreme learning
[39] Huang, G.-B. 2015. What are extreme learning machine. Neural Computing & Applications21(6):
machines? Filling the gap between Frank 1331-1339.
Rosenblatt’s dream and John von Neumann’s puzzle.
Cognitive Computation7(3): 263-278. [53] Kasun, L. L. C., H. Zhou, G.-B. Huang & C. M.
Vong 2013. Representational learning with ELMs for
[40] Huang, G.-B. & L. Chen 2007. Convex incremental big data.
extreme learning machine. Neurocomputing70(16):
3056-3062. [54] Kaya, Y. & M. Uyar 2013. A hybrid decision support
system based on rough set and extreme learning
[41] Huang, G.-B. & L. Chen 2008. Enhanced random machine for diagnosis of hepatitis disease. Applied
search based incremental extreme learning machine. Soft Computing13(8): 3429-3438.
Neurocomputing71(16): 3460-3468.
[55] Kim, J., H. S. Shin, K. Shin & M. Lee 2009. Robust
[42] Huang, G.-B., L. Chen & C. K. Siew 2006. Universal algorithm for arrhythmia classification in ECG using
approximation using incremental constructive extreme learning machine. Biomedical engineering
feedforward networks with random hidden nodes. online8(1): 31.
IEEE Trans. Neural Networks17(4): 879-892.
[56] Lahoz, D., B. Lacruz & P. M. Mateo 2013. A multi-
[43] Huang, G.-B., M.-B. Li, L. Chen & C.-K. Siew 2008. objective micro genetic ELM algorithm.
Incremental extreme learning machine with fully Neurocomputing111: 90-103.
complex hidden nodes. Neurocomputing71(4): 576-
583. [57] Lan, Y., Z. Hu, Y. C. Soh & G.-B. Huang 2013. An
extreme learning machine approach for speaker
[44] Huang, G.-B., D. H. Wang & Y. Lan 2011. Extreme recognition. Neural Computing and
learning machines: a survey. International journal of Applications22(3-4): 417-425.
machine learning and cybernetics2(2): 107-122.
[58] Lan, Y., Y. C. Soh & G.-B. Huang 2009. Ensemble
[45] Huang, G.-B., H. Zhou, X. Ding & R. Zhang 2012. of online sequential extreme learning machine.
Extreme learning machine for regression and Neurocomputing72(13): 3391-3395.
multiclass classification. IEEE Transactions on
[59] Lan, Y., Y. C. Soh & G.-B. Huang 2010. Two-stage
Systems, Man, and Cybernetics, Part B
(Cybernetics)42(2): 513-529. extreme learning machine for regression.
Neurocomputing73(16): 3028-3038.
[46] Huang, G.-B., Q.-Y. Zhu & C.-K. Siew 2004.
[60] Le, Q., T. Sarlós & A. Smola 2013. Fastfood-
Extreme learning machine: a new learning scheme of
approximating kernel expansions in loglinear time.
feedforward neural networks. Neural Networks,
2004. Proceedings. 2004 IEEE International Joint Proceedings of the international conference on
machine learning. 85.
Conference on. 2 pp. 985-990.
[61] Li, B., Y. Li & X. Rong 2013. The extreme learning
[47] Huang, G.-B., Q.-Y. Zhu & C.-K. Siew 2006.
Extreme learning machine: theory and applications. machine learning algorithm with tunable activation
Neurocomputing70(1): 489-501. function. Neural computing & applications: 1-9.
[62] Li, K., J.-X. Peng & G. W. Irwin 2005. A fast
[48] Huang, G., G.-B. Huang, S. Song & K. You 2015.
nonlinear model identification method. IEEE
Trends in extreme learning machines: A review.
Neural Networks61: 32-48. Transactions on Automatic Control50(8): 1211-
1216.
[49] Huang, G., S. Song, J. N. Gupta & C. Wu 2014.
[63] Li, L.-N., J.-H. Ouyang, H.-L. Chen & D.-Y. Liu
Semi-supervised and unsupervised extreme learning
2012. A computer aided diagnosis system for thyroid
machines. IEEE transactions on cybernetics44(12):
2405-2417. disease using extreme learning machine. Journal of
medical systems36(5): 3327-3337.
[50] Huang, W., Z. Tan, Z. Lin, G.-B. Huang, J. Zhou, C.
[64] Li, M.-B., G.-B. Huang, P. Saratchandran & N.
Chui, Y. Su & S. Chang 2012. A semi-automatic
4620
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
Sundararajan 2005. Fully complex extreme learning extreme learning machine. IEEE transactions on
machine. Neurocomputing68: 306-314. neural networks21(1): 158-162.
[65] Li, W., D. Wang & T. Chai 2013. Burning state [79] Minhas, R., A. Baradarani, S. Seifzadeh & Q. J. Wu
recognition of rotary kiln using ELMs with 2010. Human action recognition using extreme
heterogeneous features. Neurocomputing102: 144- learning machine based on visual vocabularies.
153. Neurocomputing73(10): 1906-1917.
[66] Liang, N.-Y., G.-B. Huang, P. Saratchandran & N. [80] Minhas, R., A. A. Mohammed & Q. J. Wu 2012.
Sundararajan 2006. A fast and accurate online Incremental learning in human action recognition
sequential learning algorithm for feedforward based on snippets. IEEE Transactions on Circuits
networks. IEEE Transactions on neural and Systems for Video Technology22(11): 1529-
networks17(6): 1411-1423. 1541.
[67] Lin, S., X. Liu, J. Fang & Z. Xu 2015. Is extreme [81] Mohammed, A. A., R. Minhas, Q. J. Wu & M. A.
learning machine feasible? A theoretical assessment Sid-Ahmed 2011. Human face recognition based on
(Part II). IEEE Transactions on Neural Networks and multidimensional PCA and extreme learning
Learning Systems26(1): 21-34. machine. Pattern Recognition44(10): 2588-2597.
[68] Liu, D., Y. Wu & H. Jiang 2016. FP-ELM: An online [82] Muthusamy, H., K. Polat & S. Yaacob 2015.
sequential learning algorithm for dealing with Improved emotion recognition using Gaussian
concept drift. Neurocomputing207: 322-334. Mixture Model and extreme learning machine in
[69] Liu, X., C. Gao & P. Li 2012. A comparative speech and glottal signals. Mathematical Problems in
Engineering2015.
analysis of support vector machines and extreme
learning machines. Neural Networks33: 58-66. [83] Oneto, L., F. Bisio, E. Cambria & D. Anguita 2016.
[70] Liu, X., P. Li & C. Gao 2013. Symmetric extreme Statistical learning theory and ELM for big social
data analysis. ieee CompUTATionAl inTelliGenCe
learning machine. Neural computing & applications:
mAGAzine11(3): 45-55.
1-8.
[71] Liu, X., S. Lin, J. Fang & Z. Xu 2015. Is extreme [84] Osman, M. K., M. Y. Mashor & H. Jaafar 2012.
learning machine feasible? A theoretical assessment Performance comparison of extreme learning
machine algorithms for mycobacterium tuberculosis
(Part I). IEEE Transactions on Neural Networks and
detection in tissue sections. Journal of Medical
Learning Systems26(1): 7-20.
Imaging and Health Informatics2(3): 307-312.
[72] Lu, B., G. Wang, Y. Yuan & D. Han 2013. Semantic
concept detection for video based on extreme [85] Qu, Y., C. Shang, W. Wu & Q. Shen 2011.
Evolutionary Fuzzy Extreme Learning Machine for
learning machine. Neurocomputing102: 176-183.
Mammographic Risk Analysis. International Journal
[73] Malathi, V., N. Marimuthu & S. Baskar 2010. of Fuzzy Systems13(4).
Intelligent approaches using support vector machine
and extreme learning machine for transmission line [86] Rahimi, A. & B. Recht 2008. Random features for
large-scale kernel machines. Advances in neural
protection. Neurocomputing73(10): 2160-2167.
information processing systems. pp. 1177-1184.
[74] Malathi, V., N. Marimuthu, S. Baskar & K. Ramar
2011. Application of extreme learning machine for [87] Rahimi, A. & B. Recht 2008. Uniform approximation
series compensated transmission line protection. of functions with random bases. Communication,
Control, and Computing, 2008 46th Annual Allerton
Engineering Applications of Artificial
Conference on. pp. 555-561.
Intelligence24(5): 880-887.
[75] Man, Z., K. Lee, D. Wang, Z. Cao & C. Miao 2011. [88] Rahimi, A. & B. Recht 2009. Weighted sums of
random kitchen sinks: Replacing minimization with
A new robust training algorithm for a class of single-
randomization in learning. Advances in neural
hidden layer feedforward neural networks.
information processing systems. pp. 1313-1320.
Neurocomputing74(16): 2491-2501.
[76] Marqués, I. & M. Graña 2013. Fusion of lattice [89] Rasheed, Z. & H. Rangwala 2012. Metagenomic
taxonomic classification using extreme learning
independent and linear features improving face
machines. Journal of bioinformatics and
identification. Neurocomputing114: 80-85.
computational biology10(05): 1250015.
[77] MartíNez-MartíNez, J. M., P. Escandell-Montero, E.
Soria-Olivas, J. D. MartíN-Guerrero, R. Magdalena- [90] Rong, H.-J., G.-B. Huang, N. Sundararajan & P.
Saratchandran 2009. Online sequential fuzzy
Benedito & J. GóMez-Sanchis 2011. Regularized
extreme learning machine for function
extreme learning machine for regression problems.
approximation and classification problems. IEEE
Neurocomputing74(17): 3716-3721.
Transactions on Systems, Man, and Cybernetics,
[78] Miche, Y., A. Sorjamaa, P. Bas, O. Simula, C. Jutten Part B (Cybernetics)39(4): 1067-1072.
& A. Lendasse 2010. OP-ELM: optimally pruned
[91] Rong, H.-J., Y.-S. Ong, A.-H. Tan & Z. Zhu 2008. A
4621
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
4622
International Journal of Applied Engineering Research ISSN 0973-4562 Volume 12, Number 14 (2017) pp. 4610-4623
© Research India Publications. http://www.ripublication.com
4623