SlideShare a Scribd company logo
2
Most read
3
Most read
20
Most read
CLASSIFICATION USING
BACK-PROPAGATION
Backpropagation Algorithms
 The back-propagation learning algorithm is one of the most important
developments in neural networks.
 Backpropagation is the generalization of the Widrow-Hoff learning
rule to multiple-layer networks and nonlinear differentiable transfer
functions.
 This learning algorithm is applied to multilayer feed-forward networks
consisting of processing elements with continuous differentiable
activation functions.
 The networks associated with back-propagation algorithm are also
called back-propagation networks(BPNs).
Backpropagation Algorithms
 The Aim Of The Neural Network Is To Train The Net Ot
Achieve A Balance Between The Net’s Ability To
Respond(memorization) And Its Ability To Give
Resasonable Responses To The Input That Is Similar But
Not Identical To The One That Is Use In Trianing
(Generalization).
Architecture
This section presents the architecture of the network that is most
commonly used with the backpropagation algorithm –
the multilayer feedforward network
Architecture
Feedforward Network
Feedforward networks often have one or more hidden layers of sigmoid neurons followed
by an output layer of linear neurons.
Multiple layers of neurons with nonlinear transfer functions allow the network to learn
nonlinear and linear relationships between input and output vectors.
The linear output layer lets the network produce values outside the range -1 to +1. On the
other hand, if you want to constrain the outputs of a network (such as between 0 and 1),
then the output layer should use a sigmoid transfer function (such as logsig).
Learning Algorithm:
Backpropagation
The following slides describes teaching process of multi-layer neural network
employing backpropagation algorithm. To illustrate this process the three layer neural
network with two inputs and one output,which is shown in the picture below, is used:
Learning Algorithm:
Backpropagation
Each neuron is composed of two units. First unit adds products of weights coefficients and
input signals. The second unit realise nonlinear function, called neuron transfer (activation)
function. Signal e is adder output signal, and y = f(e) is output signal of nonlinear element.
Signal y is also output signal of neuron.
Learning Algorithm:
Backpropagation
To teach the neural network we need training data set. The training data set consists of input
signals (x1 and x2 ) assigned with corresponding target (desired output) z.
The network training is an iterative process. In each iteration weights coefficients of nodes
are modified using new data from training data set. Modification is calculated using
algorithm described below:
Each teaching step starts with forcing both input signals from training set. After this stage we
can determine output signals values for each neuron in each network layer.
Learning Algorithm:
Backpropagation
Pictures below illustrate how signal is propagating through the network,
Symbols w(xm)n represent weights of connections between network input xm and
neuron n in input layer. Symbols yn represents output signal of neuron n.
Learning Algorithm:
Backpropagation
Learning Algorithm:
Backpropagation
Learning Algorithm:
Backpropagation
Propagation of signals through the hidden layer. Symbols wmn represent weights
of connections between output of neuron m and input of neuron n in the next
layer.
Learning Algorithm:
Backpropagation
Learning Algorithm:
Backpropagation
Learning Algorithm:
Backpropagation
Propagation of signals through the output layer.
Learning Algorithm:
Backpropagation
In the next algorithm step the output signal of the network y is compared
with the desired output value (the target), which is found in training data
set. The difference is called error signal d of output layer neuron
Learning Algorithm:
Backpropagation
The idea is to propagate error signal d (computed in single teaching step)
back to all neurons, which output signals were input for discussed
neuron.
Learning Algorithm:
Backpropagation
The idea is to propagate error signal d (computed in single teaching step)
back to all neurons, which output signals were input for discussed
neuron.
Learning Algorithm:
Backpropagation
The weights' coefficients wmn used to propagate errors back are equal to
this used during computing output value. Only the direction of data flow
is changed (signals are propagated from output to inputs one after the
other). This technique is used for all network layers. If propagated errors
came from few neurons they are added. The illustration is below:
Learning Algorithm:
Backpropagation
When the error signal for each neuron is computed, the weights
coefficients of each neuron input node may be modified. In formulas
below df(e)/de represents derivative of neuron activation function (which
weights are modified).
Learning Algorithm:
Backpropagation
When the error signal for each neuron is computed, the weights
coefficients of each neuron input node may be modified. In formulas
below df(e)/de represents derivative of neuron activation function (which
weights are modified).
Learning Algorithm:
Backpropagation
When the error signal for each neuron is computed, the weights
coefficients of each neuron input node may be modified. In formulas
below df(e)/de represents derivative of neuron activation function (which
weights are modified).
Backpropagation
applications
 They have been successful on a wide array of real-world data,
including
 handwritten character recognition,
 pathology and laboratory medicine,
 and training a computer to pronounce English text.
Backpropagation
terminologies
 EPOCH
 EPOCHUPDATION
 TERMINATING CONDITION
Ad

Recommended

MD5 ALGORITHM.pptx
MD5 ALGORITHM.pptx
Rajapriya82
 
Algorithms and Flowcharts
Algorithms and Flowcharts
Deva Singh
 
Open Source Software
Open Source Software
Ali Yavari
 
Measurement in research
Measurement in research
Bikram Pradhan
 
Perceptron & Neural Networks
Perceptron & Neural Networks
NAGUR SHAREEF SHAIK
 
Mobile tower radiation and its impact on environment
Mobile tower radiation and its impact on environment
Nishat Fatima
 
Concurrency control
Concurrency control
Soumyajit Dutta
 
Role of artificial intelligence in health care
Role of artificial intelligence in health care
Prachi Gupta
 
Support Vector Machines ( SVM )
Support Vector Machines ( SVM )
Mohammad Junaid Khan
 
Mc culloch pitts neuron
Mc culloch pitts neuron
Siksha 'O' Anusandhan (Deemed to be University )
 
backpropagation in neural networks
backpropagation in neural networks
Akash Goel
 
Recognition-of-tokens
Recognition-of-tokens
Dattatray Gandhmal
 
Feedforward neural network
Feedforward neural network
Sopheaktra YONG
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
Francesco Collova'
 
Back propagation
Back propagation
Nagarajan
 
Ensemble methods in machine learning
Ensemble methods in machine learning
SANTHOSH RAJA M G
 
Adaline madaline
Adaline madaline
Nagarajan
 
Artificial Neural Network ANN
Artificial Neural Network ANN
Abdullah al Mamun
 
Machine Learning with Decision trees
Machine Learning with Decision trees
Knoldus Inc.
 
Hetro associative memory
Hetro associative memory
DEEPENDRA KORI
 
Adaptive Resonance Theory
Adaptive Resonance Theory
Naveen Kumar
 
Artifical Neural Network and its applications
Artifical Neural Network and its applications
Sangeeta Tiwari
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
Mohammed Bennamoun
 
Decision tree induction \ Decision Tree Algorithm with Example| Data science
Decision tree induction \ Decision Tree Algorithm with Example| Data science
MaryamRehman6
 
Unsupervised learning clustering
Unsupervised learning clustering
Arshad Farhad
 
Huffman Coding
Huffman Coding
anithabalaprabhu
 
Neural Networks
Neural Networks
NikitaRuhela
 
Frames
Frames
amitp26
 
Artificial neural network
Artificial neural network
mustafa aadel
 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...
bihira aggrey
 

More Related Content

What's hot (20)

Support Vector Machines ( SVM )
Support Vector Machines ( SVM )
Mohammad Junaid Khan
 
Mc culloch pitts neuron
Mc culloch pitts neuron
Siksha 'O' Anusandhan (Deemed to be University )
 
backpropagation in neural networks
backpropagation in neural networks
Akash Goel
 
Recognition-of-tokens
Recognition-of-tokens
Dattatray Gandhmal
 
Feedforward neural network
Feedforward neural network
Sopheaktra YONG
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
Francesco Collova'
 
Back propagation
Back propagation
Nagarajan
 
Ensemble methods in machine learning
Ensemble methods in machine learning
SANTHOSH RAJA M G
 
Adaline madaline
Adaline madaline
Nagarajan
 
Artificial Neural Network ANN
Artificial Neural Network ANN
Abdullah al Mamun
 
Machine Learning with Decision trees
Machine Learning with Decision trees
Knoldus Inc.
 
Hetro associative memory
Hetro associative memory
DEEPENDRA KORI
 
Adaptive Resonance Theory
Adaptive Resonance Theory
Naveen Kumar
 
Artifical Neural Network and its applications
Artifical Neural Network and its applications
Sangeeta Tiwari
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
Mohammed Bennamoun
 
Decision tree induction \ Decision Tree Algorithm with Example| Data science
Decision tree induction \ Decision Tree Algorithm with Example| Data science
MaryamRehman6
 
Unsupervised learning clustering
Unsupervised learning clustering
Arshad Farhad
 
Huffman Coding
Huffman Coding
anithabalaprabhu
 
Neural Networks
Neural Networks
NikitaRuhela
 
Frames
Frames
amitp26
 
backpropagation in neural networks
backpropagation in neural networks
Akash Goel
 
Feedforward neural network
Feedforward neural network
Sopheaktra YONG
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
Francesco Collova'
 
Back propagation
Back propagation
Nagarajan
 
Ensemble methods in machine learning
Ensemble methods in machine learning
SANTHOSH RAJA M G
 
Adaline madaline
Adaline madaline
Nagarajan
 
Artificial Neural Network ANN
Artificial Neural Network ANN
Abdullah al Mamun
 
Machine Learning with Decision trees
Machine Learning with Decision trees
Knoldus Inc.
 
Hetro associative memory
Hetro associative memory
DEEPENDRA KORI
 
Adaptive Resonance Theory
Adaptive Resonance Theory
Naveen Kumar
 
Artifical Neural Network and its applications
Artifical Neural Network and its applications
Sangeeta Tiwari
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
Mohammed Bennamoun
 
Decision tree induction \ Decision Tree Algorithm with Example| Data science
Decision tree induction \ Decision Tree Algorithm with Example| Data science
MaryamRehman6
 
Unsupervised learning clustering
Unsupervised learning clustering
Arshad Farhad
 

Similar to Classification using back propagation algorithm (20)

Artificial neural network
Artificial neural network
mustafa aadel
 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...
bihira aggrey
 
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
sureshkumarece1
 
Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders
Akash Goel
 
Lec 6-bp
Lec 6-bp
Taymoor Nazmy
 
MNN
MNN
International Islamic University
 
Unit ii supervised ii
Unit ii supervised ii
Indira Priyadarsini
 
Electricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANN
Naren Chandra Kattla
 
Artificial Neural Network for machine learning
Artificial Neural Network for machine learning
2303oyxxxjdeepak
 
Multi Layer Network
Multi Layer Network
International Islamic University
 
lecture07.ppt
lecture07.ppt
butest
 
Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithms
aciijournal
 
2.5 backpropagation
2.5 backpropagation
Krish_ver2
 
Neural-Networks.ppt
Neural-Networks.ppt
RINUSATHYAN
 
CST413 KTU S7 CSE Machine Learning Neural Networks and Support Vector Machine...
CST413 KTU S7 CSE Machine Learning Neural Networks and Support Vector Machine...
resming1
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
aciijournal
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
aciijournal
 
Backpropagation.pptx
Backpropagation.pptx
VandanaVipparthi
 
22PCOAM16 _ML_ Unit 2 Full unit notes.pdf
22PCOAM16 _ML_ Unit 2 Full unit notes.pdf
Guru Nanak Technical Institutions
 
22PCOAM16 ML UNIT 2 NOTES & QB QUESTION WITH ANSWERS
22PCOAM16 ML UNIT 2 NOTES & QB QUESTION WITH ANSWERS
Guru Nanak Technical Institutions
 
Artificial neural network
Artificial neural network
mustafa aadel
 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...
bihira aggrey
 
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
sureshkumarece1
 
Intro to Deep learning - Autoencoders
Intro to Deep learning - Autoencoders
Akash Goel
 
Electricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANN
Naren Chandra Kattla
 
Artificial Neural Network for machine learning
Artificial Neural Network for machine learning
2303oyxxxjdeepak
 
lecture07.ppt
lecture07.ppt
butest
 
Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithms
aciijournal
 
2.5 backpropagation
2.5 backpropagation
Krish_ver2
 
Neural-Networks.ppt
Neural-Networks.ppt
RINUSATHYAN
 
CST413 KTU S7 CSE Machine Learning Neural Networks and Support Vector Machine...
CST413 KTU S7 CSE Machine Learning Neural Networks and Support Vector Machine...
resming1
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
aciijournal
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
aciijournal
 
Ad

Recently uploaded (20)

Rapid Prototyping for XR: Lecture 2 - Low Fidelity Prototyping.
Rapid Prototyping for XR: Lecture 2 - Low Fidelity Prototyping.
Mark Billinghurst
 
Call For Papers - 17th International Conference on Wireless & Mobile Networks...
Call For Papers - 17th International Conference on Wireless & Mobile Networks...
hosseinihamid192023
 
Deep Learning for Image Processing on 16 June 2025 MITS.pptx
Deep Learning for Image Processing on 16 June 2025 MITS.pptx
resming1
 
Tally.ERP 9 at a Glance.book - Tally Solutions .pdf
Tally.ERP 9 at a Glance.book - Tally Solutions .pdf
Shabista Imam
 
Complete guidance book of Asp.Net Web API
Complete guidance book of Asp.Net Web API
Shabista Imam
 
Bitumen Emulsion by Dr Sangita Ex CRRI Delhi
Bitumen Emulsion by Dr Sangita Ex CRRI Delhi
grilcodes
 
Structural Wonderers_new and ancient.pptx
Structural Wonderers_new and ancient.pptx
nikopapa113
 
AI_Presentation (1). Artificial intelligence
AI_Presentation (1). Artificial intelligence
RoselynKaur8thD34
 
Tesla-Stock-Analysis-and-Forecast.pptx (1).pptx
Tesla-Stock-Analysis-and-Forecast.pptx (1).pptx
moonsony54
 
Validating a Citizen Observatories enabling Platform by completing a Citizen ...
Validating a Citizen Observatories enabling Platform by completing a Citizen ...
Diego López-de-Ipiña González-de-Artaza
 
Proposal for folders structure division in projects.pdf
Proposal for folders structure division in projects.pdf
Mohamed Ahmed
 
Rapid Prototyping for XR: Lecture 4 - High Level Prototyping.
Rapid Prototyping for XR: Lecture 4 - High Level Prototyping.
Mark Billinghurst
 
DESIGN OF REINFORCED CONCRETE ELEMENTS S
DESIGN OF REINFORCED CONCRETE ELEMENTS S
prabhusp8
 
Data Structures Module 3 Binary Trees Binary Search Trees Tree Traversals AVL...
Data Structures Module 3 Binary Trees Binary Search Trees Tree Traversals AVL...
resming1
 
Complete University of Calculus :: 2nd edition
Complete University of Calculus :: 2nd edition
Shabista Imam
 
Modern multi-proposer consensus implementations
Modern multi-proposer consensus implementations
François Garillot
 
Mobile database systems 20254545645.pptx
Mobile database systems 20254545645.pptx
herosh1968
 
FSE_LLM4SE1_A Tool for In-depth Analysis of Code Execution Reasoning of Large...
FSE_LLM4SE1_A Tool for In-depth Analysis of Code Execution Reasoning of Large...
cl144
 
Rapid Prototyping for XR: Lecture 5 - Cross Platform Development
Rapid Prototyping for XR: Lecture 5 - Cross Platform Development
Mark Billinghurst
 
FUNDAMENTALS OF COMPUTER ORGANIZATION AND ARCHITECTURE
FUNDAMENTALS OF COMPUTER ORGANIZATION AND ARCHITECTURE
Shabista Imam
 
Rapid Prototyping for XR: Lecture 2 - Low Fidelity Prototyping.
Rapid Prototyping for XR: Lecture 2 - Low Fidelity Prototyping.
Mark Billinghurst
 
Call For Papers - 17th International Conference on Wireless & Mobile Networks...
Call For Papers - 17th International Conference on Wireless & Mobile Networks...
hosseinihamid192023
 
Deep Learning for Image Processing on 16 June 2025 MITS.pptx
Deep Learning for Image Processing on 16 June 2025 MITS.pptx
resming1
 
Tally.ERP 9 at a Glance.book - Tally Solutions .pdf
Tally.ERP 9 at a Glance.book - Tally Solutions .pdf
Shabista Imam
 
Complete guidance book of Asp.Net Web API
Complete guidance book of Asp.Net Web API
Shabista Imam
 
Bitumen Emulsion by Dr Sangita Ex CRRI Delhi
Bitumen Emulsion by Dr Sangita Ex CRRI Delhi
grilcodes
 
Structural Wonderers_new and ancient.pptx
Structural Wonderers_new and ancient.pptx
nikopapa113
 
AI_Presentation (1). Artificial intelligence
AI_Presentation (1). Artificial intelligence
RoselynKaur8thD34
 
Tesla-Stock-Analysis-and-Forecast.pptx (1).pptx
Tesla-Stock-Analysis-and-Forecast.pptx (1).pptx
moonsony54
 
Validating a Citizen Observatories enabling Platform by completing a Citizen ...
Validating a Citizen Observatories enabling Platform by completing a Citizen ...
Diego López-de-Ipiña González-de-Artaza
 
Proposal for folders structure division in projects.pdf
Proposal for folders structure division in projects.pdf
Mohamed Ahmed
 
Rapid Prototyping for XR: Lecture 4 - High Level Prototyping.
Rapid Prototyping for XR: Lecture 4 - High Level Prototyping.
Mark Billinghurst
 
DESIGN OF REINFORCED CONCRETE ELEMENTS S
DESIGN OF REINFORCED CONCRETE ELEMENTS S
prabhusp8
 
Data Structures Module 3 Binary Trees Binary Search Trees Tree Traversals AVL...
Data Structures Module 3 Binary Trees Binary Search Trees Tree Traversals AVL...
resming1
 
Complete University of Calculus :: 2nd edition
Complete University of Calculus :: 2nd edition
Shabista Imam
 
Modern multi-proposer consensus implementations
Modern multi-proposer consensus implementations
François Garillot
 
Mobile database systems 20254545645.pptx
Mobile database systems 20254545645.pptx
herosh1968
 
FSE_LLM4SE1_A Tool for In-depth Analysis of Code Execution Reasoning of Large...
FSE_LLM4SE1_A Tool for In-depth Analysis of Code Execution Reasoning of Large...
cl144
 
Rapid Prototyping for XR: Lecture 5 - Cross Platform Development
Rapid Prototyping for XR: Lecture 5 - Cross Platform Development
Mark Billinghurst
 
FUNDAMENTALS OF COMPUTER ORGANIZATION AND ARCHITECTURE
FUNDAMENTALS OF COMPUTER ORGANIZATION AND ARCHITECTURE
Shabista Imam
 
Ad

Classification using back propagation algorithm

  • 2. Backpropagation Algorithms  The back-propagation learning algorithm is one of the most important developments in neural networks.  Backpropagation is the generalization of the Widrow-Hoff learning rule to multiple-layer networks and nonlinear differentiable transfer functions.  This learning algorithm is applied to multilayer feed-forward networks consisting of processing elements with continuous differentiable activation functions.  The networks associated with back-propagation algorithm are also called back-propagation networks(BPNs).
  • 3. Backpropagation Algorithms  The Aim Of The Neural Network Is To Train The Net Ot Achieve A Balance Between The Net’s Ability To Respond(memorization) And Its Ability To Give Resasonable Responses To The Input That Is Similar But Not Identical To The One That Is Use In Trianing (Generalization).
  • 4. Architecture This section presents the architecture of the network that is most commonly used with the backpropagation algorithm – the multilayer feedforward network
  • 5. Architecture Feedforward Network Feedforward networks often have one or more hidden layers of sigmoid neurons followed by an output layer of linear neurons. Multiple layers of neurons with nonlinear transfer functions allow the network to learn nonlinear and linear relationships between input and output vectors. The linear output layer lets the network produce values outside the range -1 to +1. On the other hand, if you want to constrain the outputs of a network (such as between 0 and 1), then the output layer should use a sigmoid transfer function (such as logsig).
  • 6. Learning Algorithm: Backpropagation The following slides describes teaching process of multi-layer neural network employing backpropagation algorithm. To illustrate this process the three layer neural network with two inputs and one output,which is shown in the picture below, is used:
  • 7. Learning Algorithm: Backpropagation Each neuron is composed of two units. First unit adds products of weights coefficients and input signals. The second unit realise nonlinear function, called neuron transfer (activation) function. Signal e is adder output signal, and y = f(e) is output signal of nonlinear element. Signal y is also output signal of neuron.
  • 8. Learning Algorithm: Backpropagation To teach the neural network we need training data set. The training data set consists of input signals (x1 and x2 ) assigned with corresponding target (desired output) z. The network training is an iterative process. In each iteration weights coefficients of nodes are modified using new data from training data set. Modification is calculated using algorithm described below: Each teaching step starts with forcing both input signals from training set. After this stage we can determine output signals values for each neuron in each network layer.
  • 9. Learning Algorithm: Backpropagation Pictures below illustrate how signal is propagating through the network, Symbols w(xm)n represent weights of connections between network input xm and neuron n in input layer. Symbols yn represents output signal of neuron n.
  • 12. Learning Algorithm: Backpropagation Propagation of signals through the hidden layer. Symbols wmn represent weights of connections between output of neuron m and input of neuron n in the next layer.
  • 15. Learning Algorithm: Backpropagation Propagation of signals through the output layer.
  • 16. Learning Algorithm: Backpropagation In the next algorithm step the output signal of the network y is compared with the desired output value (the target), which is found in training data set. The difference is called error signal d of output layer neuron
  • 17. Learning Algorithm: Backpropagation The idea is to propagate error signal d (computed in single teaching step) back to all neurons, which output signals were input for discussed neuron.
  • 18. Learning Algorithm: Backpropagation The idea is to propagate error signal d (computed in single teaching step) back to all neurons, which output signals were input for discussed neuron.
  • 19. Learning Algorithm: Backpropagation The weights' coefficients wmn used to propagate errors back are equal to this used during computing output value. Only the direction of data flow is changed (signals are propagated from output to inputs one after the other). This technique is used for all network layers. If propagated errors came from few neurons they are added. The illustration is below:
  • 20. Learning Algorithm: Backpropagation When the error signal for each neuron is computed, the weights coefficients of each neuron input node may be modified. In formulas below df(e)/de represents derivative of neuron activation function (which weights are modified).
  • 21. Learning Algorithm: Backpropagation When the error signal for each neuron is computed, the weights coefficients of each neuron input node may be modified. In formulas below df(e)/de represents derivative of neuron activation function (which weights are modified).
  • 22. Learning Algorithm: Backpropagation When the error signal for each neuron is computed, the weights coefficients of each neuron input node may be modified. In formulas below df(e)/de represents derivative of neuron activation function (which weights are modified).
  • 23. Backpropagation applications  They have been successful on a wide array of real-world data, including  handwritten character recognition,  pathology and laboratory medicine,  and training a computer to pronounce English text.