Neural Network Cover Theory Topic :
- Neurons or Perceptron's
- Activation Function
- Cost Function
- Gradient Descent
- Backpropagation
Let's Get Started -
What is Perceptrons ?
- Neuron is also known as Perceptron
- Artificial Neural Network are based on natural biological systems.
- Artificial Neural Network (ANN) is a software based approach to replicate these biological neurons.
The Biological Neuron 
First are the dendrites :  think of these  as the terminals at which the 
neuron receives its inputs (real dendrites can do computations and have 
feedback control; this is not typically modeled an artificial neuron;
Next is the cell body; think of this as where any processing  occurs. 
Finally, the axon carries the output from the cell body to neighboring neurons and their dendrites. Basically the neuron is  a 
“simple” computational device: it receives input at its dendrites, does a
 computation at the cell body, and carries the output on its axon (I’ll 
just note here that a real neuron is much more complicated; for example 
the dendrites themselves can carry out some kinds of computation).
Biological neurons communicate across a synapse,
 which is a chemical “connection”, typically between an axon and 
dendrite. Signals flow from the axon terminal to receptors on the 
dendrite, mediated by the chemical state of both the axon and 
dendrite. 
So what does learning mean in this setting? According to Hebbian theory (named after Canadian psychologist Donald Hebb),
 learning  is related to the increase in synaptic efficacy that arises 
from the presynaptic cell’s repeated and persistent stimulation of the 
postsynaptic cell. 
 It is this increase in “communication” efficacy that
 we call learning.  One way to think about this is that the connection 
between the axon terminal and the dendrite is weighted, and the
 larger the weight (concentration of neurotransmitters in the axon 
terminal along with other chemical components including receptors in the
 dendrites) the more likely the neuron is to “fire”.  
So in this setting
 learning consists of optimizing neuron firings in certain patterns; to 
put it a different way, learning consists of optimizing the connection 
weights between axons and dendrites in a way that leads to some observed
 behavior. We will return to this idea of optimizing weights as learning
 when we talk about training Artificial Neural Networks.
The Artificial Neuron :
 
 
 
 
 
 
No comments:
Post a Comment