American Journal of Intelligent Systems
p-ISSN: 2165-8978 e-ISSN: 2165-8994
2014; 4(4): 154-158
doi:10.5923/j.ajis.20140404.05
1Department of Business, Faculty of Economic and Administrative Sciences, Ondokuz Mayis University, Samsun, Turkey
2Department of Statistics, Faculty of Arts and Science, Marmara University, Istanbul, Turkey
Correspondence to: Erol Egrioglu, Department of Statistics, Faculty of Arts and Science, Marmara University, Istanbul, Turkey.
| Email: | ![]() |
Copyright © 2014 Scientific & Academic Publishing. All Rights Reserved.
Partial least square regression (PLSR) is a latent variable based multivariate statistical method that is a combination of partial least square (PLS) and multiple linear regressions. It accounts for small sample size, large number of predictor variables, correlated variables and several response variables. It is almost used commonly in all area. But in some complicated data sets linear PLS methods do not give satisfactory results so nonlinear PLS approaches were examined in literature. Feed forward artificial neural networks based nonlinear PLS method was proposed in the literature. In this study, the method of nonlinear PLS is improved to make a suggestion of a new nonlinear PLS method which is based on Elman feedback artificial neural networks. The proposed method is applied to data set of “30 young football players enrolled in the league of Football Players who are Candidates of Professional Leagues” and compare with some PLS methods.
Keywords: Partial least squares regression, Elman neural network, Prediction, Feed forward neural network
Cite this paper: Elif Bulut, Erol Egrioglu, A New Partial Least Square Method Based on Elman Neural Network, American Journal of Intelligent Systems, Vol. 4 No. 4, 2014, pp. 154-158. doi: 10.5923/j.ajis.20140404.05.
and
without any assumptions about their dimensions (N, M or K).
is
,
is
where N also represents the number of rows (observations), M also represents the number of columns (predictors), and K is the number of response variables. The standard procedure in PLS method is to centered and scaled matrices by subtracting their averages and dividing their standard deviations. The PLS regression method is an iterative method except on single y variable. In this study, NIPALS algorithm was performed in analysis. This algorithm can be explained as follows: It starts with the centered and scaled X and Y matrices as
and
. NIPALS algorithm compose of two loops. The inner loop is used to attain latent variables. The corresponding weight vectors w and c for latent variables are obtained by multiplying the latent variables through the specific matrix as
and
. u is taken as the first column or column with the biggest variance of Y matrix and w and c are scaled to length 1. t latent variable is obtained as
. New u latent variable is defined as
. Then a convergence is tested on the change in u. If convergence has been reached the outer loop is used sequentially to extract p loading vectors from X matrices with the new pairs of latent variables. Otherwise, this loop is repeated until a convergence is reached. Loadings are obtained as
. In this loop, it is possible to calculate the new t as
. In this algorithm a regression model between latent variables is written as
and named as inner model. Here b is the regression coefficient of inner relation and computed by
. Loadings are calculated to obtain the residual matrices that will be used in the next iteration as
and
. In these equations the subtracted parts represent decomposition of matrices X and Y into biliner products and named as outer model. These residual matrices are used to obtain new t and u latent variables. The whole set of latent variables has been found when E residual matrix became a null matrix. For more information look (Geladi et al., [6]).An neural network partial least square (NNPLS) modeling approach which was proposed by Qin et al. [12] was considered in this study. They proposed an NNPLS modeling approach by keeping the outer relation in linear PLS while using neural networks as the inner regressors:
. Here,
stands for the nonlinear relation represented by a neural network. h and rh represents iteration number and residual.![]() | Figure 1. Multilayer feed forward artificial neural network with one output neuron |
![]() | Figure 2. Elman recurrent artificial neural network |
and h=1.Step 2. For each factor h, take
.Step 3. PLS outer transform:in matrix X:
, normalize
to norm 1.
• in matrix Y:
, normalize
to norm 1.
, Iterate this step until it converges. Step 4. Calculate the X loadings and rescale the variables:
, normalize
,
. Step 5. Find inner network model: train inner network such that the following error function is minimized by Levenberg-Marquardt method.
Here,
is the output of the Elman artificial neural network. For example, if the number of hidden layer was 1, Elman Neural Network that was used in obtaining
was given in Figure 3. The architecture of Elman type artificial neutral network was given in Figure 3 and logistic activation function was used in all neurons of Elman type artificial neutral network. Also, the number of hidden layer unit has been determined obtained by trial-error method. ![]() | Figure 3. Inner Model |
for matrix Y,
where
.Step 7. Let h=h+1, return to Step 2 until all principal factors are calculated. In the proposed method, FFANNs are used in every iteration of algorithm. Because of this, a lot of FFANN were employed in the proposed method. Convergence of FFANN is provided by using 100 iteration at least for each neural network.
, Y:
. Vertical jumping was measured in centimeter, broad jumping was measured in meter. Length and circumference measurements were measured in centimeters and skinfold was measured in millimeter. Randomly selected 27 observations were used to obtain the models. 3 observations were used as test set (ntest=3). That is, 27 observations were used in modeling, 3 were used in prediction. As a comparison criterion RMSE (root mean square error) was used. RMSE values for test set by predicting with PLSR method appear in Table 4.
Firstly, prediction was made with FFANN method in MATLAB R2011b. Input number of FFANN is the number of explanatory variables. On the other hand, the numbers of hidden layer neurons vary between 1 and 73, the 73 different FFANN architectures are used for obtaining predictions. The FFANN was trained by using Levenberg-Marquardt algorithm with 500 maximum number of iterations. The best result of FFANN is the architecture (73-8-2) which has 73 inputs, 8 hidden layer neurons and two outputs.
|