SlideShare a Scribd company logo
Sundarapandian et al. (Eds): CoNeCo,WiMo, NLP, CRYPSIS, ICAIT, ICDIP, ITCSE, CS & IT 07,
pp. 393–398, 2012. © CS & IT-CSCP 2012 DOI : 10.5121/csit.2012.2438
BACKPROPAGATION LEARNING ALGORITHM
BASED ON LEVENBERG MARQUARDT
ALGORITHM
S.Sapna1
, Dr.A.Tamilarasi2
and M.Pravin Kumar3
1
Assistant Professor, Department of Master of Computer Applications, K.S.R
College of Engineering, Tiruchengode-637215, Tamilnadu, India.
sapnaanjusrijumani@rediffmail.com, sapnaanjusri@gmail.com
2
Prof. & Head, Department of Master of Computer Applications, Kongu
Engineering College, Perundurai, Tamilnadu, India.
3
Assistant Professor, Department of Electronics and Communication
Engineering, K.S.R College of Engineering, Tiruchengode-637215,
Tamilnadu, India
erodepravin@gmail.com
ABSTRACT
Data Mining aims at discovering knowledge out of data and presenting it in a form that is easily
compressible to humans. Data Mining represents a process developed to examine large amounts
of data routinely collected. The term also refers to a collection of tools used to perform the
process. One of the useful applications in the field of medicine is the incurable chronic disease
diabetes. Data Mining algorithm is used for testing the accuracy in predicting diabetic status.
Fuzzy Systems are been used for solving a wide range of problems in different application
domain and Genetic Algorithm for designing. Fuzzy systems allows in introducing the learning
and adaptation capabilities. Neural Networks are efficiently used for learning membership
functions. Diabetes occurs throughout the world, but Type 2 is more common in the most
developed countries. The greater increase in prevalence is however expected in Asia and Africa
where most patients will likely be found by 2030. This paper is proposed on the Levenberg –
Marquardt algorithm which is specifically designed to minimize sum-of-square error functions.
Levernberg-Marquardt algorithm gives the best performance in the prediction of diabetes
compared to any other backpropogation algorithm.
KEYWORDS
Data Mining, Diabetes, Fuzzy Systems, Genetic Algorithm (GA), Levenberg – Marquardt,
Neural Networks.
1. INTRODUCTION
In this paper an intellectual and efficient diabetic prediction method with the support of fuzzy
neural network is proposed. The network is trained using multi layer feed forward back
propagation algorithm to test its performance. The proposed method will not only assist medical
practitioners but also supports special educators, occupational therapists and psychologist in
better assessment of diabetes disease. Usually the decisions made by the medical experts are
seldom based on the single symptom because of the complication of the human body, as one
symptom could indicate any number of problems. A skilled medical expert is far more likely to
make a sound assessment than a beginner, because from his past knowledge he knows what to
394 Computer Science & Information Technology ( CS & IT )
look out for and what to inquire, and may have etched on his mind a past inaccuracy, which he
neither will nor repeat. Thus the senior medical expert is in a higher position than the trainee.
Similarly it would be helpful if machines, too, could use past events as part of the factors on
which their decisions are based, and this is the role that neural network seeks to fill.
The back propagation algorithm was developed by Paul Werbos in 1974 and rediscovered
independently by Rumelhart and Parker [4]. Since its rediscovery, the back propagation algorithm
has been widely used as a learning algorithm in feed forward multilayer neural networks. ANN
using the Back Propagation (BP) algorithm performs parallel training for improving the
efficiency of Multilayer Perceptron (MLP) network. It is the most popular, effective, and easy to
learn model for complex, multilayered networks. A Backpropagation is a supervised learning
technique which is based on the Gradient Descent (GD) method that attempts to minimize the
error of the network by moving down the gradient of the error curve as stated [2,6]. The most
popular in the supervised learning architecture because of the weight error correct rules [3,5]. It is
considered a generalization of the delta rule for nonlinear activation functions and multilayer
networks.
The structure of layered feedforward neural networks is considered and each of these networks
consists of a set of inputs and one or more layers of parallel neurons. Inputs are connected only to
neurons in the first layer with the exception of the special input X0, representing the bias of each
neuron, which is connected to all neurons in the network. Neurons in one layer are connected to
all neurons in the next layer. No other connections exist in neural networks of this type. The last
layer, which produces the output of the network, is called an output layer. Any layers that precede
the output layer are called hidden layers. The set of inputs is sometimes referred to as an input
layer. Inputs don’t do any computation; their only role is to feed input patterns into the rest of the
network.
2. BACK PROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG
MARQUARDT ALGORITHM (LM)
Levenberg – Marquardt algorithm is specifically designed to minimize sum-of-square error
functions [1], of the form.
( ) 2
e
2
12
kek
2
1E =∑=
Where ek is the error in the kth exemplar or pattern and e is a vector with element ek. If the
difference between the pervious weight vector and the new weight vector is small, the error
vector can be expanded to first order by means of a Taylor series.
( ) ( ) ( ) ( )( )jw1jw
iw/kejeije −+∂∂+=+
As a consequence, the error function can be expressed as
( ) ( ) ( )( )2
jw1jwiw/keje
2
1E −+∂∂+=
Minimizing the error function with respect to the new weight vector, gives
… (2.1)
… (2.2)
… (2.3)
Computer Science & Information Technology ( CS & IT ) 395
( ) ( ) ( )jeTZ
1
ZTZjw1jw
−




−=+
where ( ) iw/kekiZ ∂∂≡
Since the Hessian for the sum-of-square error function is
( ) ( )( ) }{∑ ∂∂∂+∂∂∂∂=∂∂∂= wjiw/ke
2
keiw/keiw/kewjiw/E2
ijH
Neglecting the second term, the Hessian can be written as ZTZH =
Updating of the weights therefore involves the inverse Hessian or an approximation thereof for
nonlinear networks. The Hessian is relatively easy to compute, since it is based on first order
derivatives with respect to the network weights that are easily accommodated by
backpropagation. Although the updating formula could be applied iteratively to minimize the
error function, this may result in a large step size, which would invalidated the linear
approximation on which the formula is based.
In the Levenberg-Marquardt algorithm, the error function is minimized, while the step size is kept
small in order to ensure the validity of the linear approximation. This is accomplished by use of a
modified error function of the form.
( ) ( ) ( )( ) ( ) ( )
2
jw1jw
2
jw1jw
iw/keje
2
1E −+λ+−+∂∂+=
where λ is a parameter governing the step size. Minimizing the modified error with respect to
( )1jw + gives
( ) ( ) ( )jeTZ
1
IZTZjw1jw
−




 λ+−=+
very large values of λ amount to standard gradient descent, while very small values λ of amount
to the Newton method.
3. PERFORMANCE EVALUATION USING LEVENBERG-MARQUARDT
ALGORITHM
For this implementation, the data set collected from SRC Diabetic Care Centre, Erode District,
Tamilnadu in South India is used. Levenberg-Marquardt back propagation algorithm is used for
training the network [7,8]. Training automatically stops when generalization stops improving, as
indicated by an increase in the Mean Square Error (MSE) of the validation samples. The Mean
Squared Error (MSE) is the average squared difference between outputs and targets. Lower
values are better while zero means no error. Regression R analysis is performed to measure the
correlation between outputs and targets.
… (2.4)
… (2.5)
… (2.6)
… (2.7)
… (2.8)
396 Computer Science & Information Technology ( CS & IT )
Figure 3.1 Performance of Levenberg-Marquardt Backpropagation Algorithm
An R value of 1 means a close relationship, 0 a random relationship. The performance of the
proposed network when trained with Levenberg-Marquardt backpropagation algorithm using
Matlab R2007b, is shown in Figure 3.1.
From Figure 3.1 it is observed that the best validation performance 0.00012359 at epoch 9 is
obtained. The Regression plot shown in Figure 3.2 shows the perfect correlation between the
outputs and the targets.
Figure 3.2 Regression Analysis Plot - Levenberg-Marquardt Backpropagation Algorithm
Computer Science & Information Technology ( CS & IT ) 397
Table 3.1 The Mean Square Error (MSE) and Regression (R) values for the Training, Validation
and Testing
MSE R
Training 2.25880e-4 0.902515
Validation 8.83702e-5 1.000000
Testing 6.15467e-4 1.000000
The result for training, validation and testing samples is illustrated in Table 3.1. It is observed that
the value of R is closest to 1 indicating the accurate prediction. When the data set was trained in
Levenberg – Marquardt algorithms the performance obtained was in 9 epochs. Levenberg –
Marquardt algorithm (LM) is the most widely used optimization algorithm. LM algorithm is an
iterative technique that locates a local minimum of a multivariate function that is expressed as the
sum of squares of several non-linear, real-valued functions. It has become a standard technique
for non linear least-square problems, widely adopted in various disciplines for dealing data-fitting
applications. Levenberg-Marquardt curve-fitting method is actually a combination of two
minimization methods the gradient descent method and the Gauss-Newton method.
This paper presented a standard nonlinear least squares optimization algorithm, and showed how
to include it into the backpropagation algorithm. The Marquardt algorithm was experienced on
several function approximation problems, and it was compared with the conjugate gradient
algorithm and with variable learning rate backpropagation. The results indicated that the
Marquardt algorithm is very efficient when training networks which have up to a few hundred
weights. Although the computational requirements are much higher for each iteration of the
Marquardt algorithm, this is more than made up for by the increased efficiency. This is specially
true when high accuracy is essential. It is also found that in many cases the Marquardt algorithm
converged when the conjugate gradient and variable learning rate algorithms failed to converge.
4. CONCLUSION
This paper aimed to evaluate the artificial neural network in predicting diabetes disease. The feed-
forward backpropagation neural network with supervised learning is proposed to diagnose the
disease. The reliability of the proposed neural network method depends on data collected from
the patients and experts opinion. Backpropagation learning algorithm is used to train the
feedforward neural network to perform a given task based on Levenberg-Marquardt algorithm
and also the performance is analyzed. It is analyzed that Levernberg-Marquardt algorithm gives
the best performance in the prediction of diabetes compared to any other backpropogation
algorithm. The proposed diagnosis based on neural network showed significant results in
identifying the diabetes.
REFERENCE
[1] Aldrich .C (2002), “Exploratory Analysis of Metallurgical Process Data with Neural Networks and
Related Methods”, Elsevier Science, British Library Cataloguing in Publication Data, Netherlands,
Vol.1, pp.56-57.
[2] Alsmadi M. K. S., K. Omar, and S. A. Noah (2009), “Back Propagation Algorithm: The Best
Algorithm among the Multi-layer Perceptron Algorithm”, International Journal of Computer Science
and Network Security, pp. 378-383.
398 Computer Science & Information Technology ( CS & IT )
[3] Insung Jung, Lockjo Koo and Gi-Nam Wang (2007), “Two States Mapping Based Neural Network
Model for Decreasing of Prediction Residual Error”, International Journal of Electrical and Computer
Engineering, Vol.2, No.8, pp.531-537.
[4] Kevin L. Priddy and Paul E.Keller (2005), “Artificial Neural Networks An Introduction”, SPIE-The
International Society of Optical Engineering, Washington.
[5] Mutasem khalil Sari Alsmadi, Khairuddin Bin Omar and Shahrul Azman Noah (2009), “Back
Propagation Algorithm: The Best Algorithm among the Multi-layer Perceptron Algorithm”,
International Journal of Computer Science and Network Security, Vol.9, No.4, pp.378-383.
[6] Norhamreeza Abdul Hamid and Nazir Mohd Nawi (2011), “Accelerating Learning Performance of
Back Propagation Algorithm by Using Adaptive Gain Together with Adaptive Momentum and
Adaptive Learning Rate on Classification Problems”, International Journal of Software Engineering
and its Applications, Vol.5, No.4, pp.31-44.
[7] Martin T. Hagan and Mohammad B. Menhaj (1994), “Training Feedforward Networks with the
Marquardt Algorithm”, IEEE Transactions on Neural Networks, Vol. 5, No. 6, November 1994.
[8] Sivanandam .S.N, Sumathi.S, Deepa .S.N (2008), “Introduction to Neural Network Using MATLAB
6.0”, Tata McGraw-Hill Publishing Company Limited, New Delhi.
BIOGRAPHY
S.Sapna received her B.Sc Degree, M.C.A and M.Phil Degree, from Bharathiar University.
She is currently working as Assistant Professor in the Department of MCA, K.S.R. College
of Engineering. She has presented more than 20 papers on various topics including
national, international conference and journals. She is a research scholar of Mother Teresa
Women’s University, Kodaikanal. Her research interest includes Soft Computing, Data
Mining, Mathematical Computations and Networks. She is a life member of ISTE and CSI.
Dr.A.Tamilarasi, currently serving as Professor and Head, Department of M.C.A, Kongu
Engineering College. She has published various papers in the filed of Fuzzy Logics. She
has published various books in mathematical field. She is guiding several research scholars
in her area of interest like Data Mining, Soft Computing and Networks.
M.Pravin Kumar received his B.E & M.E Degree, from Anna University. He is
currently working as Assistant Professor in the Department of ECE, K.S.R. College of
Engineering. He has presented more than 8 papers on various topics including national,
international conference and journals. He is a research scholar of Anna University,
Coimbatore. His research interest includes Soft Computing and Networks. He is a life
member of ISTE.

More Related Content

PDF
Comparison of Neural Network Training Functions for Hematoma Classification i...
PDF
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
PDF
Black-box modeling of nonlinear system using evolutionary neural NARX model
PDF
Efficiency of Neural Networks Study in the Design of Trusses
PDF
Neural Network Implementation Control Mobile Robot
PDF
Optimal Load Shedding Using an Ensemble of Artifcial Neural Networks
PDF
Jf3515881595
PDF
Simulation of Single and Multilayer of Artificial Neural Network using Verilog
Comparison of Neural Network Training Functions for Hematoma Classification i...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Black-box modeling of nonlinear system using evolutionary neural NARX model
Efficiency of Neural Networks Study in the Design of Trusses
Neural Network Implementation Control Mobile Robot
Optimal Load Shedding Using an Ensemble of Artifcial Neural Networks
Jf3515881595
Simulation of Single and Multilayer of Artificial Neural Network using Verilog

What's hot (18)

PDF
Electricity Demand Forecasting Using ANN
PDF
Electricity Demand Forecasting Using Fuzzy-Neural Network
PDF
DCT AND DFT BASED BIOMETRIC RECOGNITION AND MULTIMODAL BIOMETRIC SECURITY
PDF
IRJET- Image Classification – Cat and Dog Images
PDF
Survey on Artificial Neural Network Learning Technique Algorithms
PPTX
Face Recognition: From Scratch To Hatch / Эдуард Тянтов (Mail.ru Group)
PDF
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
PDF
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...
PDF
PERFORMANCE ASSESSMENT OF ANFIS APPLIED TO FAULT DIAGNOSIS OF POWER TRANSFORMER
PDF
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
PDF
Optimal neural network models for wind speed prediction
PDF
DSR Routing Decisions for Mobile Ad Hoc Networks using Fuzzy Inference System
PDF
Long Zhou - 2017 - Neural System Combination for Machine Transaltion
PDF
Design of c slotted microstrip antenna using
PDF
IRJET - Machine Learning Algorithms for the Detection of Diabetes
PDF
A Learning Linguistic Teaching Control for a Multi-Area Electric Power System
PPTX
Perceptron and Sigmoid Neurons
PPT
Artificial neural networks in hydrology
Electricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using Fuzzy-Neural Network
DCT AND DFT BASED BIOMETRIC RECOGNITION AND MULTIMODAL BIOMETRIC SECURITY
IRJET- Image Classification – Cat and Dog Images
Survey on Artificial Neural Network Learning Technique Algorithms
Face Recognition: From Scratch To Hatch / Эдуард Тянтов (Mail.ru Group)
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...
PERFORMANCE ASSESSMENT OF ANFIS APPLIED TO FAULT DIAGNOSIS OF POWER TRANSFORMER
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Optimal neural network models for wind speed prediction
DSR Routing Decisions for Mobile Ad Hoc Networks using Fuzzy Inference System
Long Zhou - 2017 - Neural System Combination for Machine Transaltion
Design of c slotted microstrip antenna using
IRJET - Machine Learning Algorithms for the Detection of Diabetes
A Learning Linguistic Teaching Control for a Multi-Area Electric Power System
Perceptron and Sigmoid Neurons
Artificial neural networks in hydrology
Ad

Similar to BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM (20)

PDF
24csit38.pdf
PDF
Poster_Reseau_Neurones_Journees_2013
PDF
Hidden Layer Leraning Vector Quantizatio
PDF
Neural network based numerical digits recognization using nnt in matlab
PDF
A Fletcher-Reeves conjugate gradient algorithm-based neuromodel for smart gri...
PDF
Fpga implementation of optimal step size nlms algorithm and its performance a...
PDF
Fpga implementation of optimal step size nlms algorithm and its performance a...
PDF
Power system transient stability margin estimation using artificial neural ne...
PDF
Adaptive modified backpropagation algorithm based on differential errors
PDF
Multimode system condition monitoring using sparsity reconstruction for quali...
PDF
Artificial Neural Networks Deep Learning Report
PDF
Initial Optimal Parameters of Artificial Neural Network and Support Vector Re...
PDF
F017533540
PDF
MARGINAL PERCEPTRON FOR NON-LINEAR AND MULTI CLASS CLASSIFICATION
PPTX
Introduction Of Artificial neural network
PDF
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
PDF
International Refereed Journal of Engineering and Science (IRJES)
PDF
A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...
PPTX
SC_U2_PPT.pptxE13RWT4SDGHYFYJGVKBHJL.J,HGRFEDWSEZRFXDGTCHVYJBKNLM.LK,JKHM
PDF
Predicting rainfall using ensemble of ensembles
24csit38.pdf
Poster_Reseau_Neurones_Journees_2013
Hidden Layer Leraning Vector Quantizatio
Neural network based numerical digits recognization using nnt in matlab
A Fletcher-Reeves conjugate gradient algorithm-based neuromodel for smart gri...
Fpga implementation of optimal step size nlms algorithm and its performance a...
Fpga implementation of optimal step size nlms algorithm and its performance a...
Power system transient stability margin estimation using artificial neural ne...
Adaptive modified backpropagation algorithm based on differential errors
Multimode system condition monitoring using sparsity reconstruction for quali...
Artificial Neural Networks Deep Learning Report
Initial Optimal Parameters of Artificial Neural Network and Support Vector Re...
F017533540
MARGINAL PERCEPTRON FOR NON-LINEAR AND MULTI CLASS CLASSIFICATION
Introduction Of Artificial neural network
An Optimized Parallel Algorithm for Longest Common Subsequence Using Openmp –...
International Refereed Journal of Engineering and Science (IRJES)
A ROBUST MISSING VALUE IMPUTATION METHOD MIFOIMPUTE FOR INCOMPLETE MOLECULAR ...
SC_U2_PPT.pptxE13RWT4SDGHYFYJGVKBHJL.J,HGRFEDWSEZRFXDGTCHVYJBKNLM.LK,JKHM
Predicting rainfall using ensemble of ensembles
Ad

More from cscpconf (20)

PDF
ANALYSIS OF LAND SURFACE DEFORMATION GRADIENT BY DINSAR
PDF
4D AUTOMATIC LIP-READING FOR SPEAKER'S FACE IDENTIFCATION
PDF
MOVING FROM WATERFALL TO AGILE PROCESS IN SOFTWARE ENGINEERING CAPSTONE PROJE...
PDF
PROMOTING STUDENT ENGAGEMENT USING SOCIAL MEDIA TECHNOLOGIES
PDF
A SURVEY ON QUESTION ANSWERING SYSTEMS: THE ADVANCES OF FUZZY LOGIC
PDF
DYNAMIC PHONE WARPING – A METHOD TO MEASURE THE DISTANCE BETWEEN PRONUNCIATIONS
PDF
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
PDF
TWO DISCRETE BINARY VERSIONS OF AFRICAN BUFFALO OPTIMIZATION METAHEURISTIC
PDF
DETECTION OF ALGORITHMICALLY GENERATED MALICIOUS DOMAIN
PDF
GLOBAL MUSIC ASSET ASSURANCE DIGITAL CURRENCY: A DRM SOLUTION FOR STREAMING C...
PDF
IMPORTANCE OF VERB SUFFIX MAPPING IN DISCOURSE TRANSLATION SYSTEM
PDF
EXACT SOLUTIONS OF A FAMILY OF HIGHER-DIMENSIONAL SPACE-TIME FRACTIONAL KDV-T...
PDF
AUTOMATED PENETRATION TESTING: AN OVERVIEW
PDF
CLASSIFICATION OF ALZHEIMER USING fMRI DATA AND BRAIN NETWORK
PDF
VALIDATION METHOD OF FUZZY ASSOCIATION RULES BASED ON FUZZY FORMAL CONCEPT AN...
PDF
PROBABILITY BASED CLUSTER EXPANSION OVERSAMPLING TECHNIQUE FOR IMBALANCED DATA
PDF
CHARACTER AND IMAGE RECOGNITION FOR DATA CATALOGING IN ECOLOGICAL RESEARCH
PDF
SOCIAL MEDIA ANALYTICS FOR SENTIMENT ANALYSIS AND EVENT DETECTION IN SMART CI...
PDF
SOCIAL NETWORK HATE SPEECH DETECTION FOR AMHARIC LANGUAGE
PDF
GENERAL REGRESSION NEURAL NETWORK BASED POS TAGGING FOR NEPALI TEXT
ANALYSIS OF LAND SURFACE DEFORMATION GRADIENT BY DINSAR
4D AUTOMATIC LIP-READING FOR SPEAKER'S FACE IDENTIFCATION
MOVING FROM WATERFALL TO AGILE PROCESS IN SOFTWARE ENGINEERING CAPSTONE PROJE...
PROMOTING STUDENT ENGAGEMENT USING SOCIAL MEDIA TECHNOLOGIES
A SURVEY ON QUESTION ANSWERING SYSTEMS: THE ADVANCES OF FUZZY LOGIC
DYNAMIC PHONE WARPING – A METHOD TO MEASURE THE DISTANCE BETWEEN PRONUNCIATIONS
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
TWO DISCRETE BINARY VERSIONS OF AFRICAN BUFFALO OPTIMIZATION METAHEURISTIC
DETECTION OF ALGORITHMICALLY GENERATED MALICIOUS DOMAIN
GLOBAL MUSIC ASSET ASSURANCE DIGITAL CURRENCY: A DRM SOLUTION FOR STREAMING C...
IMPORTANCE OF VERB SUFFIX MAPPING IN DISCOURSE TRANSLATION SYSTEM
EXACT SOLUTIONS OF A FAMILY OF HIGHER-DIMENSIONAL SPACE-TIME FRACTIONAL KDV-T...
AUTOMATED PENETRATION TESTING: AN OVERVIEW
CLASSIFICATION OF ALZHEIMER USING fMRI DATA AND BRAIN NETWORK
VALIDATION METHOD OF FUZZY ASSOCIATION RULES BASED ON FUZZY FORMAL CONCEPT AN...
PROBABILITY BASED CLUSTER EXPANSION OVERSAMPLING TECHNIQUE FOR IMBALANCED DATA
CHARACTER AND IMAGE RECOGNITION FOR DATA CATALOGING IN ECOLOGICAL RESEARCH
SOCIAL MEDIA ANALYTICS FOR SENTIMENT ANALYSIS AND EVENT DETECTION IN SMART CI...
SOCIAL NETWORK HATE SPEECH DETECTION FOR AMHARIC LANGUAGE
GENERAL REGRESSION NEURAL NETWORK BASED POS TAGGING FOR NEPALI TEXT

Recently uploaded (20)

PDF
Open folder Downloads.pdf yes yes ges yes
PDF
Pre independence Education in Inndia.pdf
PDF
Insiders guide to clinical Medicine.pdf
PDF
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table
PDF
102 student loan defaulters named and shamed – Is someone you know on the list?
PDF
O7-L3 Supply Chain Operations - ICLT Program
PPTX
Open Quiz Monsoon Mind Game Prelims.pptx
PPTX
Introduction to Child Health Nursing – Unit I | Child Health Nursing I | B.Sc...
PDF
Abdominal Access Techniques with Prof. Dr. R K Mishra
PPTX
Revamp in MTO Odoo 18 Inventory - Odoo Slides
PPTX
Open Quiz Monsoon Mind Game Final Set.pptx
PDF
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
PDF
3rd Neelam Sanjeevareddy Memorial Lecture.pdf
PPTX
Pharmacology of Heart Failure /Pharmacotherapy of CHF
PPTX
Renaissance Architecture: A Journey from Faith to Humanism
PDF
Mark Klimek Lecture Notes_240423 revision books _173037.pdf
PPTX
Introduction_to_Human_Anatomy_and_Physiology_for_B.Pharm.pptx
PPTX
IMMUNITY IMMUNITY refers to protection against infection, and the immune syst...
PPTX
Introduction and Scope of Bichemistry.pptx
PDF
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student
Open folder Downloads.pdf yes yes ges yes
Pre independence Education in Inndia.pdf
Insiders guide to clinical Medicine.pdf
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table
102 student loan defaulters named and shamed – Is someone you know on the list?
O7-L3 Supply Chain Operations - ICLT Program
Open Quiz Monsoon Mind Game Prelims.pptx
Introduction to Child Health Nursing – Unit I | Child Health Nursing I | B.Sc...
Abdominal Access Techniques with Prof. Dr. R K Mishra
Revamp in MTO Odoo 18 Inventory - Odoo Slides
Open Quiz Monsoon Mind Game Final Set.pptx
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
3rd Neelam Sanjeevareddy Memorial Lecture.pdf
Pharmacology of Heart Failure /Pharmacotherapy of CHF
Renaissance Architecture: A Journey from Faith to Humanism
Mark Klimek Lecture Notes_240423 revision books _173037.pdf
Introduction_to_Human_Anatomy_and_Physiology_for_B.Pharm.pptx
IMMUNITY IMMUNITY refers to protection against infection, and the immune syst...
Introduction and Scope of Bichemistry.pptx
grade 11-chemistry_fetena_net_5883.pdf teacher guide for all student

BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM

  • 1. Sundarapandian et al. (Eds): CoNeCo,WiMo, NLP, CRYPSIS, ICAIT, ICDIP, ITCSE, CS & IT 07, pp. 393–398, 2012. © CS & IT-CSCP 2012 DOI : 10.5121/csit.2012.2438 BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM S.Sapna1 , Dr.A.Tamilarasi2 and M.Pravin Kumar3 1 Assistant Professor, Department of Master of Computer Applications, K.S.R College of Engineering, Tiruchengode-637215, Tamilnadu, India. [email protected], [email protected] 2 Prof. & Head, Department of Master of Computer Applications, Kongu Engineering College, Perundurai, Tamilnadu, India. 3 Assistant Professor, Department of Electronics and Communication Engineering, K.S.R College of Engineering, Tiruchengode-637215, Tamilnadu, India [email protected] ABSTRACT Data Mining aims at discovering knowledge out of data and presenting it in a form that is easily compressible to humans. Data Mining represents a process developed to examine large amounts of data routinely collected. The term also refers to a collection of tools used to perform the process. One of the useful applications in the field of medicine is the incurable chronic disease diabetes. Data Mining algorithm is used for testing the accuracy in predicting diabetic status. Fuzzy Systems are been used for solving a wide range of problems in different application domain and Genetic Algorithm for designing. Fuzzy systems allows in introducing the learning and adaptation capabilities. Neural Networks are efficiently used for learning membership functions. Diabetes occurs throughout the world, but Type 2 is more common in the most developed countries. The greater increase in prevalence is however expected in Asia and Africa where most patients will likely be found by 2030. This paper is proposed on the Levenberg – Marquardt algorithm which is specifically designed to minimize sum-of-square error functions. Levernberg-Marquardt algorithm gives the best performance in the prediction of diabetes compared to any other backpropogation algorithm. KEYWORDS Data Mining, Diabetes, Fuzzy Systems, Genetic Algorithm (GA), Levenberg – Marquardt, Neural Networks. 1. INTRODUCTION In this paper an intellectual and efficient diabetic prediction method with the support of fuzzy neural network is proposed. The network is trained using multi layer feed forward back propagation algorithm to test its performance. The proposed method will not only assist medical practitioners but also supports special educators, occupational therapists and psychologist in better assessment of diabetes disease. Usually the decisions made by the medical experts are seldom based on the single symptom because of the complication of the human body, as one symptom could indicate any number of problems. A skilled medical expert is far more likely to make a sound assessment than a beginner, because from his past knowledge he knows what to
  • 2. 394 Computer Science & Information Technology ( CS & IT ) look out for and what to inquire, and may have etched on his mind a past inaccuracy, which he neither will nor repeat. Thus the senior medical expert is in a higher position than the trainee. Similarly it would be helpful if machines, too, could use past events as part of the factors on which their decisions are based, and this is the role that neural network seeks to fill. The back propagation algorithm was developed by Paul Werbos in 1974 and rediscovered independently by Rumelhart and Parker [4]. Since its rediscovery, the back propagation algorithm has been widely used as a learning algorithm in feed forward multilayer neural networks. ANN using the Back Propagation (BP) algorithm performs parallel training for improving the efficiency of Multilayer Perceptron (MLP) network. It is the most popular, effective, and easy to learn model for complex, multilayered networks. A Backpropagation is a supervised learning technique which is based on the Gradient Descent (GD) method that attempts to minimize the error of the network by moving down the gradient of the error curve as stated [2,6]. The most popular in the supervised learning architecture because of the weight error correct rules [3,5]. It is considered a generalization of the delta rule for nonlinear activation functions and multilayer networks. The structure of layered feedforward neural networks is considered and each of these networks consists of a set of inputs and one or more layers of parallel neurons. Inputs are connected only to neurons in the first layer with the exception of the special input X0, representing the bias of each neuron, which is connected to all neurons in the network. Neurons in one layer are connected to all neurons in the next layer. No other connections exist in neural networks of this type. The last layer, which produces the output of the network, is called an output layer. Any layers that precede the output layer are called hidden layers. The set of inputs is sometimes referred to as an input layer. Inputs don’t do any computation; their only role is to feed input patterns into the rest of the network. 2. BACK PROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM (LM) Levenberg – Marquardt algorithm is specifically designed to minimize sum-of-square error functions [1], of the form. ( ) 2 e 2 12 kek 2 1E =∑= Where ek is the error in the kth exemplar or pattern and e is a vector with element ek. If the difference between the pervious weight vector and the new weight vector is small, the error vector can be expanded to first order by means of a Taylor series. ( ) ( ) ( ) ( )( )jw1jw iw/kejeije −+∂∂+=+ As a consequence, the error function can be expressed as ( ) ( ) ( )( )2 jw1jwiw/keje 2 1E −+∂∂+= Minimizing the error function with respect to the new weight vector, gives … (2.1) … (2.2) … (2.3)
  • 3. Computer Science & Information Technology ( CS & IT ) 395 ( ) ( ) ( )jeTZ 1 ZTZjw1jw −     −=+ where ( ) iw/kekiZ ∂∂≡ Since the Hessian for the sum-of-square error function is ( ) ( )( ) }{∑ ∂∂∂+∂∂∂∂=∂∂∂= wjiw/ke 2 keiw/keiw/kewjiw/E2 ijH Neglecting the second term, the Hessian can be written as ZTZH = Updating of the weights therefore involves the inverse Hessian or an approximation thereof for nonlinear networks. The Hessian is relatively easy to compute, since it is based on first order derivatives with respect to the network weights that are easily accommodated by backpropagation. Although the updating formula could be applied iteratively to minimize the error function, this may result in a large step size, which would invalidated the linear approximation on which the formula is based. In the Levenberg-Marquardt algorithm, the error function is minimized, while the step size is kept small in order to ensure the validity of the linear approximation. This is accomplished by use of a modified error function of the form. ( ) ( ) ( )( ) ( ) ( ) 2 jw1jw 2 jw1jw iw/keje 2 1E −+λ+−+∂∂+= where λ is a parameter governing the step size. Minimizing the modified error with respect to ( )1jw + gives ( ) ( ) ( )jeTZ 1 IZTZjw1jw −      λ+−=+ very large values of λ amount to standard gradient descent, while very small values λ of amount to the Newton method. 3. PERFORMANCE EVALUATION USING LEVENBERG-MARQUARDT ALGORITHM For this implementation, the data set collected from SRC Diabetic Care Centre, Erode District, Tamilnadu in South India is used. Levenberg-Marquardt back propagation algorithm is used for training the network [7,8]. Training automatically stops when generalization stops improving, as indicated by an increase in the Mean Square Error (MSE) of the validation samples. The Mean Squared Error (MSE) is the average squared difference between outputs and targets. Lower values are better while zero means no error. Regression R analysis is performed to measure the correlation between outputs and targets. … (2.4) … (2.5) … (2.6) … (2.7) … (2.8)
  • 4. 396 Computer Science & Information Technology ( CS & IT ) Figure 3.1 Performance of Levenberg-Marquardt Backpropagation Algorithm An R value of 1 means a close relationship, 0 a random relationship. The performance of the proposed network when trained with Levenberg-Marquardt backpropagation algorithm using Matlab R2007b, is shown in Figure 3.1. From Figure 3.1 it is observed that the best validation performance 0.00012359 at epoch 9 is obtained. The Regression plot shown in Figure 3.2 shows the perfect correlation between the outputs and the targets. Figure 3.2 Regression Analysis Plot - Levenberg-Marquardt Backpropagation Algorithm
  • 5. Computer Science & Information Technology ( CS & IT ) 397 Table 3.1 The Mean Square Error (MSE) and Regression (R) values for the Training, Validation and Testing MSE R Training 2.25880e-4 0.902515 Validation 8.83702e-5 1.000000 Testing 6.15467e-4 1.000000 The result for training, validation and testing samples is illustrated in Table 3.1. It is observed that the value of R is closest to 1 indicating the accurate prediction. When the data set was trained in Levenberg – Marquardt algorithms the performance obtained was in 9 epochs. Levenberg – Marquardt algorithm (LM) is the most widely used optimization algorithm. LM algorithm is an iterative technique that locates a local minimum of a multivariate function that is expressed as the sum of squares of several non-linear, real-valued functions. It has become a standard technique for non linear least-square problems, widely adopted in various disciplines for dealing data-fitting applications. Levenberg-Marquardt curve-fitting method is actually a combination of two minimization methods the gradient descent method and the Gauss-Newton method. This paper presented a standard nonlinear least squares optimization algorithm, and showed how to include it into the backpropagation algorithm. The Marquardt algorithm was experienced on several function approximation problems, and it was compared with the conjugate gradient algorithm and with variable learning rate backpropagation. The results indicated that the Marquardt algorithm is very efficient when training networks which have up to a few hundred weights. Although the computational requirements are much higher for each iteration of the Marquardt algorithm, this is more than made up for by the increased efficiency. This is specially true when high accuracy is essential. It is also found that in many cases the Marquardt algorithm converged when the conjugate gradient and variable learning rate algorithms failed to converge. 4. CONCLUSION This paper aimed to evaluate the artificial neural network in predicting diabetes disease. The feed- forward backpropagation neural network with supervised learning is proposed to diagnose the disease. The reliability of the proposed neural network method depends on data collected from the patients and experts opinion. Backpropagation learning algorithm is used to train the feedforward neural network to perform a given task based on Levenberg-Marquardt algorithm and also the performance is analyzed. It is analyzed that Levernberg-Marquardt algorithm gives the best performance in the prediction of diabetes compared to any other backpropogation algorithm. The proposed diagnosis based on neural network showed significant results in identifying the diabetes. REFERENCE [1] Aldrich .C (2002), “Exploratory Analysis of Metallurgical Process Data with Neural Networks and Related Methods”, Elsevier Science, British Library Cataloguing in Publication Data, Netherlands, Vol.1, pp.56-57. [2] Alsmadi M. K. S., K. Omar, and S. A. Noah (2009), “Back Propagation Algorithm: The Best Algorithm among the Multi-layer Perceptron Algorithm”, International Journal of Computer Science and Network Security, pp. 378-383.
  • 6. 398 Computer Science & Information Technology ( CS & IT ) [3] Insung Jung, Lockjo Koo and Gi-Nam Wang (2007), “Two States Mapping Based Neural Network Model for Decreasing of Prediction Residual Error”, International Journal of Electrical and Computer Engineering, Vol.2, No.8, pp.531-537. [4] Kevin L. Priddy and Paul E.Keller (2005), “Artificial Neural Networks An Introduction”, SPIE-The International Society of Optical Engineering, Washington. [5] Mutasem khalil Sari Alsmadi, Khairuddin Bin Omar and Shahrul Azman Noah (2009), “Back Propagation Algorithm: The Best Algorithm among the Multi-layer Perceptron Algorithm”, International Journal of Computer Science and Network Security, Vol.9, No.4, pp.378-383. [6] Norhamreeza Abdul Hamid and Nazir Mohd Nawi (2011), “Accelerating Learning Performance of Back Propagation Algorithm by Using Adaptive Gain Together with Adaptive Momentum and Adaptive Learning Rate on Classification Problems”, International Journal of Software Engineering and its Applications, Vol.5, No.4, pp.31-44. [7] Martin T. Hagan and Mohammad B. Menhaj (1994), “Training Feedforward Networks with the Marquardt Algorithm”, IEEE Transactions on Neural Networks, Vol. 5, No. 6, November 1994. [8] Sivanandam .S.N, Sumathi.S, Deepa .S.N (2008), “Introduction to Neural Network Using MATLAB 6.0”, Tata McGraw-Hill Publishing Company Limited, New Delhi. BIOGRAPHY S.Sapna received her B.Sc Degree, M.C.A and M.Phil Degree, from Bharathiar University. She is currently working as Assistant Professor in the Department of MCA, K.S.R. College of Engineering. She has presented more than 20 papers on various topics including national, international conference and journals. She is a research scholar of Mother Teresa Women’s University, Kodaikanal. Her research interest includes Soft Computing, Data Mining, Mathematical Computations and Networks. She is a life member of ISTE and CSI. Dr.A.Tamilarasi, currently serving as Professor and Head, Department of M.C.A, Kongu Engineering College. She has published various papers in the filed of Fuzzy Logics. She has published various books in mathematical field. She is guiding several research scholars in her area of interest like Data Mining, Soft Computing and Networks. M.Pravin Kumar received his B.E & M.E Degree, from Anna University. He is currently working as Assistant Professor in the Department of ECE, K.S.R. College of Engineering. He has presented more than 8 papers on various topics including national, international conference and journals. He is a research scholar of Anna University, Coimbatore. His research interest includes Soft Computing and Networks. He is a life member of ISTE.