American Journal of Intelligent Systems
p-ISSN: 2165-8978 e-ISSN: 2165-8994
2012; 2(4): 40-44
doi: 10.5923/j.ajis.20120204.02
E. Shafiei , H. Jazayeri-Rad
Department of Automation and Instrumentation, Petroleum University of Technology, Ahwaz, Iran
Correspondence to: E. Shafiei , Department of Automation and Instrumentation, Petroleum University of Technology, Ahwaz, Iran.
| Email: | ![]() |
Copyright © 2012 Scientific & Academic Publishing. All Rights Reserved.
Modelling or identification of industrial plants is the first and most crucial step in their implementation process. Artificial neural networks (ANNs) as a powerful tool for modelling have been offered in recent years. Industrial processes are often so complicated that using a single neural network (SNN) is not optimal. SNNs in dealing with complex processes do not perform as required. For example the process models with this method are not accurate enough or the dynamic characteristics of the system are not adequately represented. SNNs are generally non-robust and they are sometimes over fitted. So in this paper, we use multiple neural networks (MNNs) for modelling. Bagging and boosting are two methods employed to construct MNNs. Here, we concentrate on the use of these two methods in modelling a continuous stirred tank reactor (CSTR) and compare the results against the SNN model. Simulation results show that the use of MNNs improves the model performance.most popular ones include some elaboration of bagging[6-9], and boosting[10-18]. In this work, we parallel the use of bagging and boosting methods in modelling a chemical plant (CSTR) and compare the results against the corresponding SNN model.
Keywords: System identification, Industrial processes,Neural networks, Bagging, Boosting, Bootstrapping
, where x and y are input and output variables, respectively. It is required to acquire B bootstrap datasets. As a first step, each instance in T is assigned a probability of 1/N, and the training set for each of the bootstrap member TB is created by sampling with replacement N times from the original dataset T using the above probabilities. Hence, each bootstrap dataset TB may have many instances in T repeated a number of times, while other instances may be omitted. Individual neural network models are then trained on each of TB. Therefore, for any given input vector, the bootstrap algorithm offers B different outputs. The bagging estimate is then computed by determining the mean of B model predictions (see Figure 1).![]() | Figure 1. Bagging neural network |
3) For i = 1 to B do:3.1) Make a new training set TB by sampling N items evenly at random with replacement from the data set T.3.2) Train an estimator
with the set TB and add it to the ensemble.4) For any new testing pattern, the bagged ensemble output is given by the equation:![]() | (1) |
, where x and y are input and output variables, respectively. Initially each value in the dataset is allocated the same probability value so that each instance in the initial dataset has an equal likelihood of being sampled in the first training set; that is, sampling distribution,
at step
is equal to
over all i, where
to N.![]() | Figure 2. Boosting neural network |
such that
for
.2) For
: (B can be determined by a trial and error method).2.1) Fill the new training set with the distribution wtand obtain a hypothesis
2.2) Compute the adjusted error
for each instance:![]() | (2) |
![]() | (3) |
if
, stop and set
2.4) Let
2.5) Bring up to date the weight vector according to the equation:![]() | (4) |
= the weighted median of
for
, using
as the weight for the hypothesisht.![]() | Figure 3. Input variable (coolant flow rate) |
![]() | Figure 4. Output variable (concentration) |
in the present time. In this case, the inputs to the network are the inputs in the present time
the previous input
and the previous output
. So, the neural network consists of three inputs
and one output
.To determine the final predicted output of the trained ensemble, an average is taken over the predictions from individual networks.The results and conclusions are given in the following sections.For system identification using the Adaboost.R2 algorithm, the number of iterations B should be determined. We select B to be equal to 20. All sequential networks are trained using the Levenberg-Marquardtback-propagationalgorithm. Each network consists of two layers. Activation functions in the output layer are “purelin” and in the hidden layer are “tansig”or “logsig”. The stopping goal for the single and every individual network is the mean squared error (mse = 0.000001). As described in the previous section, the lag space is equal to one; therefore, the number of sampled data points employed for modelling is 7499. So, the input-output data points are in the form of
. At the first iteration, all of the data points have equal chance to be selected. So, the probability of each data point to be elected is equal to
. Using the Adaboost.R2 algorithm described in the previous section, we will see that the value of
prior to the final iteration is less than the threshold value of 0.5, however at the final iteration
. Hence, at this point the algorithm stops.
|
|
prior to the eleventh iteration is less than the threshold value of 0.5, however at the eleventh iteration 
Hence, at this point the algorithm stops. Details of all the steps and the final results are shown in this table. The second column of the table shows that using the boosting algorithm leads to reductions of both of the error variance and modeling error in the MNN when compared against the SNN. It is clear from this table that regression of the MNN is better than the SNN (R-Squared closer to 1).| [1] | Kadkhodaie A., Rezaee M., Rahimpour H., “A committee neural network for prediction of normalized oil content from well log data: An example from South Pars Gas Field, Persian Gulf”, Journal of Petroleum Science and Engineering, vol. 65, no. 1-2, pp. 23-32, 2009. |
| [2] | Siwek K., Osowski S., Szupiluk R., “Ensemble Neural Network for Accurate Load Forecasting in a Power System”,International Journal of Applied Mathematics and Computer Science, vol. 19, no. 2, pp. 303-315, 2009. |
| [3] | Ahmad Z., Roslin F., “Modeling of Real pH Neutralization Process Using Multiple Neural Networks (MNN) Combination”, International Conference on Control, Instrumentation and Mechatronics Engineering, Malaysia, 2007. |
| [4] | Cunningham P., Carney J., Jacob S., “Stability problems with artificial neural networks and the ensemble solution”, Journal of Artificial Intelligence in Medicine, vol. 20, no. 3 , pp. 217-225, 2000. |
| [5] | Breiman L., “Bagging predictors”, Journal of Machine Learning, vol. 24, no. 2, pp. 123-140, 1996. |
| [6] | Amanda J., Sharkey A., “Combining Artificial Neural Nets: Ensemble and Modular Multi-net Systems”, Springer, New York, USA, 1999. |
| [7] | Coelho A, Nascimento D., “On the evolutionary design of heterogeneous Bagging models”, International journal of Neurocomputing, vol. 73, no. 16-18, pp. 3319-3322, 2010. |
| [8] | Chen H., Yao X., “Regularized Negative Correlation Learning for Neural Network Ensembles”, IEEE Transactions On Neural Networks, vol. 20, no. 12, pp. 1962-1979, 2009. |
| [9] | Chandrahasan R.K., Christobel Y.A., Sridhar U.R, Arockiam L., An Empirical Comparison of Boosting and BaggingAlgorithms, (IJCSIS) International Journal of Computer Science and Information Security, vol. 9, no. 11, pp. 147-152, 2011. |
| [10] | Freund Y., Schapire R., “Experiments with a new boosting algorithm”, Proceedings of the Thirteenth International Conference on Machine Learning , San Francisco, USA, 1996. |
| [11] | Bühlmann, P., Hothorn, T., “Twin Boosting: Improved Feature Selection and Prediction”, Statistics and Computing, vol. 20, no. 2, pp. 119–138, 2010. |
| [12] | Chang W.C., Cho C.W, “Online Boosting for Vehicle Detection”, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 40, no. 3, pp. 892-902, 2010. |
| [13] | Dimitrakakis C., Bengio S., “Phoneme and Sentence-Level Ensembles for Speech Recognition”, EURASIP Journal on Audio, Speech, and Music Processing,vol. 2011, no.1, 2011. |
| [14] | Gao C., Sang N., Tang Q., “On Selection and Combination of Weak Learners in AdaBoost”, Pattern Recognition Letters, vol. 31, no. 9, pp. 991-1001, 2010. |
| [15] | Landesa-Vázquez I., Alba-Castro J.L., “Shedding Light on the Asymmetric Learning Capability of AdaBoost”, Pattern Recognition Letters, vol. 33, no. 3, pp. 247-255, 2012. |
| [16] | Padmaja P., Jyothi P.N., Shilpa D., “Comparative Study of Network Boosting with AdaBoost and Bagging”, International Journal of Engineering Science and Technology vol. 2, no. 8, pp. 3447-3450, 2010. |
| [17] | Shen C., Li H., “On the Dual Formulation of Boosting Algorithms”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 32, no. 12, pp. 2216-2231, 2010. |
| [18] | Tran T.P, Cao L., Tran D., Nguyen C.D, “Novel Intrusion Detection using Probabilistic Neural Network and Adaptive Boosting”, (IJCSIS) International Journal of Computer Science and Information Security, vol. 6, No. 1, pp. 83-91, 2009. |
| [19] | Barutcoglu Z., Alpaydın E., “A Comparison of Model Aggregation Methods for Regression”, Proceedings of the 13th joint international conference on Artificial neural networks and neural information processing, Istanbul, Turkey, 2003. |
| [20] | Ahmad Z., Zhang J., “A Comparison of Different Methods for Combining Multiple Neural Networks Models”, Proceeding of the International Joint Conference on Neural Networks, Berlin, Germany, 2002. |
| [21] | Rashid T., “A Heterogeneous Ensemble Network Using Machine Learning Techniques”, International Journal of Computer Science and Network Security, vol. 9, no. 8, pp. 335-339, 2009. |
| [22] | Solomatine D. P., Shrestha D. L., “AdaBoost.RT: a Boosting Algorithm for Regression Problems”, Proceedings on International Joint Conference, Budapest Hungary, 25-29 July 2004. |
| [23] | Schapire R., “The Strength of Weak Learn ability”, Journal of Machine Learning, vol. 5, no. 2, pp. 197-227,1990. |
| [24] | Freund Y., Schapire R., “A decision-theoreticgeneralisation of on-line learning and an application of boosting”, Journal of Computer and System Sciences, vol. 55, no, 1, pp. 119-139, 1997. |
| [25] | Drucker H., “Improving regressor using boosting”, Proceedings of the 14th International Conferences on Machine Learning, Nashville , USA ,1997. |
| [26] | Bailly K., Milgram M., “Boosting feature selection for Neural Networks based regression”, International journal of Neural Networks, vol. 22, no. 5-6, pp. 748-756, 2009. |
| [27] | http://homes.esat.kuleuven.be/~smc/daisy/ |