0 ratings0% found this document useful (0 votes) 35 views29 pagesSoft Computing
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content,
claim it here.
Available Formats
Download as PDF or read online on Scribd
UNIT-1
NEURAL NETWORKS-1
WHAT IS ARTIFICIAL NEURAL NETWORK?
An Artificial Neural Network (ANN) is a mathematical model that tries to simulate the structure and
functionalities of biological neural networks. Basic building block of every artificial neural network is
artificial neuron, that is, a simple mathematical model (function). Such a model has three simple sets of
rules; multiplication, summation and activation. At the entrance of artificial neuron the inputs are weighted
what means that every input value is multiplied with individual weight. In the middle section of artificial
neuron is sum function that sums all weighted inputs and bias. At the exit of artificial neuron the sum of
previously weighted inputs and bias is passing through activation function that is also called transfer
function.
letormaton fom >BIOLOGICAL NEURON STRUCTURE AND FUNCTIONS,
Aneuron, or nerve cell, is an electrically excitable cell that communicates with other cells via specialized
connections called synapses. It s the main component of nervous tissue, Neurons are typically classified
Into three types based on their function. Sensory neurons respond to stimuli such as touch, sound, or light
that affect the cells of the sensory organs, and they send signals to the spinal cord or brain. Motor neurons
receive signals from the brain and spinal cord to control everything from muscle contractions to glandular
output. Interneurons connect neurons to other neurons within the same region of the brain or spinal cord.
‘A group of connected neurons is called a neural circuit.
Atypical neuron consists of a cell body (soma), dendrites, and a single axon. The soma is usually compact.
‘The axon and dendrites are filaments that extrude from it. Dendrites typically branch profusely and extend
a few hundred micrometers from the soma. The axon leaves the soma at a swelling called the axon hillock,
and travels for as far as 1 meter in humans or more in other species. It branches but usually maintains a
constant diameter. At the farthest tip of the axon's branches are axon terminals, where the neuron can
transmit a signal across the synapse to another cell, Neurons may lack dendrites or have no axon. The term
neurite is used to describe either a dendrite or an axon, particularly when the cell is undifferentiated.
‘The soma is the body of the neuron. As it contains the nucleus, most protein synthesis occurs here. The
nucleus can range from 3 to 18 micrometers in diameter.
‘The dendrites of a neuron are cellular extensions with many branches. This overall shape and structure is
referred to metaphorically as a dendritic tree. This is where the majority of input to the neuron occurs via
the dendritic spine.
The axon is a finer, cable-like projection that can extend tens, hundreds, or even tens of thousands of times
the diameter of the soma in length. The axon primarily carries nerve signals away from the soma, and
‘carries some types of information back to it. Many neurons have only one axon, but this axon may—and
usually will—undergo extensive branching, enabling communication with many target cells. The part of the
‘axon where it emerges from the soma is called the axon hillock. Besides being an anatomical structure, the
axon hillock also has the greatest density of voltage-dependent sodium channels. This makes it the most
easily excited part of the neuron and the spike initiation zone for the axon. In electrophysiological terms, it
has the most negative threshold potential.
While the axon and axon hillock are generally involved in information outflow, this region can also receive
{input from other neurons.
‘The axon terminal is found at the end of the axon farthest from the soma and contains synapses. Synaptic
boutons are specialized structures where neurotransmitter chemicals are released to communicate with
target neurons. In addition to synaptic boutons at the axon terminal, a neuron may have en passant
boutons, which are located along the length of the axon.
Most neurons receive signals via the dendrites and soma and send out signals down the axon. At the
majority of synapses, signals cross from the axon of one neuron to a dendrite of another. However,
‘Synapses can connect an axon to another axon or a dendrite to another dendrite, The signaling process is
partly electrical and partly chemical. Neurons are electrically excitable, due to maintenance of voltage
‘Gradients across their membranes, If the voltage changes by a large amount over a short interval the
Neuron generates an all-or-nothing electrochemical pulse called an action potential. This potential travelsrapidly along the axon, and activates synaptic connections as it reaches them. Synaptic signals may be
excitatory or inhibitory, increasing or reducing the net voltage that reaches the soma.
In most cases, neurons are generated by neural stem cells during brain development and childhood.
Neurogenesis largely ceases during adulthood in most areas of the brain. However, strong evidence
supports generation of substantial numbers of new neurons in the hippocampus and olfactory bulb.
Colt body
‘Synaptic terminals®
4c an a mode! of boloteal neurons
modal network. THe artificial neuron
ved inhibitory postaynaptic
representing a neuron's
ely weighted, and the
er funtion or transfer function, The
My cake the frm of other non-linear
estatgnen rmonotonically Increasing
red building VOpic #Ates
has bling brain processing, For
ch log In recent times.
plteable to
Sample new devices such aS memristors nave D
atin) tin)
+1 uA
, =
in,
@ &)
step fi
p function sigmoid function
4,
2 Bengt abe
She) Co vaye #O%R) cell)
Je Ykco
era Sveti fonsin .
[pavers (Mik wt. Wee
i) Fons Eaqugy +
(babies area = BT
a rial
351. Size: Our brain contains about 86 billion neurons and more than a 100 synapses (connections). The
number of “neurons” in artificial networks is much less than that,
2, Signal transport and processing: The human brain works asynchronously, ANNs work synchronously.
3, Processing speed: Single blological neurons are slow, while standard neurons in ANNs are fast.
4, Topology: Biological neural networks have complicated topologies, while ANNs are often in a tree
structure,
5, Speed: certain biological neurons can fire around 200 times a second on average. Signals travel at
different speeds depending on the type of the nerve impulse, ranging from 0.64 m/s up to 119 m/s. Signal
travel speeds also vary from person to person depending on their sex, age, height, temperature, medical
condition, lack of sleep etc. Information in artificial neurons ts carried over by the continuous, floating point
number values of synaptic weights. There are no refractory perlods for artificial neural networks (periods
while it is impossible to send another action potential, due to the sodium channels being lock shut) and
artificial neurons do not experience “fatigue”: they are functions that can be calculated as many times and
as fast as the computer architecture would allow.
6. Fault-tolerance: biological neuron networks due to thelr topology are also fault-tolerant. Artificial
neural networks are not modeled for fault tolerance or self regeneration (similarly to fatigue, these ideas
are not applicable to matrix operations), though recovery is possible by saving the current state (weight
values) of the model and continuing the training from that save state.
7. Power consumption: the brain consumes about 20% of all the human body's energy — despite it's large
cut, an adult brain operates on about 20 watts (barely enough to dimly light a bulb) being extremely
efficient. Taking into account how humans can still operate for a while, when only given some c-vitamin
rich lemon juice and beef tallow, this is quite remarkable. For benchmark: a single Nvidia GeForce Titan X
GPU runs on 250 watts alone, and requires a power supply. Our machines are way less efficient than
biological systems. Computers also generate a lot of heat when used, with consumer GPUs operating safely
between 50-B0°Celsius instead of 36.5-37.5 °C.
8. Learning: we still do not understand how brains learn, or how redundant connections store and recall
information. By learning, we are bullding on information that is already stored in the brain. Our knowledge
deepens by repetition and during sleep, and tasks that once required a focus can be executed automatically
‘once mastered. Artifical neural networks in the other hand, have a predefined model, where no further
neurons or connections can be added or removed. Only the welghts of the connections (and biases
representing thresholds) can change during training. The networks start with random weight values and
will slowly try to reach a point where further changes In the welghts would no longer improve
performance. Biological networks usually don't stop / start learning, ANNs have different fitting (train) and
prediction (evaluate) phases.
9, Fleld of application: ANNs are specialized, They can perform one task. They might be perfect at playing
chess, but they fall at playing go (or vice versa), Blological neural networks can learn completely new tasks.
10, Training algorithm: ANNs use Gradient Descent for learning, Human brains
(put we don’t know what).
ye something differentbuilding blocks:
ids upon the following three
Processing of ANN depen
1. Network Topology
2. Adjustments of Weights
3. Activation Functions
or Learning
the arrangement of a network along with Its nodes and
1 Network Topology A network Ope ee classified asthe following kinds:
connecting lines. According to the topology,
n-recurrent network having processing units/nodes in layers and
all the nodes in a layer are connected with the nodes of the previous layers. The connection has
the signal can only flow in one
different weights upon them, There is no feedback loop means °
direction, from input to output. It may be divided into the following two types:
A. Feed forward Network: It is a not
‘© Single layer feed forward network: The concept is of feed forward ANN having only one
weighted layer. In other words, we can say the input layer Is fully connected to the output
layer.
Inputs Outputs
* _ Multilayer feed forward network:
one weighted layer. As this network has one goon
cept Is of feed forward
layer, it is called hidden layers, tn
lore layers between the i
the input and
the outpu
it®
jedback Network: As the name suggests, a feedback network has feedback paths, which means the
signal can flow In both directions using loops. This makes it a non-linear dynamic system, which
changes continuously until it reaches a state of equilibrium. It may be divided into the following
y
+ Recurrent networks: They are feedback networks with closed loops. Following are the two types
of recurrent networks,
+ Fully recurrent network: It is the simplest neural network architecture because all nodes are
connected to all other nodes and each node works as both input and output.
S Spratiad One { reas oy.
Hep pitt nyo Elvan rife
4 Pra
riot pedal | Pheri, ayes
gienpine Rowse:
We MY Bhonsle
6 ina Be
Seer cunilt
+ Jordan network - It Is a closed loop network In which the output will go to the input again as
feedback as shown in the following diagram,
LO SOe es | | ©
RZ SZ
OSBOSSBO
SOS
OBSOBSO
a
2. Adjustments of Welghts or Learning: Learning, in artificial neural network, is the method of modifying
the weights of connections between the neurons of a specified network. Learning in ANN can be classified
{nto three categorles namely supervised learning, unsupervised learning, and reinforcement learning,
femnly Prodigras ones saan (Aca oupat
U Suaferied Lamy
vs on
aoe fone
GeneratorMal mdse
Tlearning is done under the supervision of a
‘of ANN under supervised learning, the
‘This output vector Is compared
srence between the actual
Ormred-y
Supervised Learning: As the name suggests, this type of
teacher. This learning process dependent. During the training
{Input vector is presented to the network, which will give an output vector.
‘with the desired output vector. An error signal is generated, if there isa diffe
seca the desired output vector. On the basis of his errr signal, the weights arg adjusted unt the
ch satis desired ouput Sane te briny empha os oer gf ath,
Pe eee eS enn al sn, encoded ‘te Kear tert
Se z eared thing) nescal Hid WK Te take dae
“thi iSarningis done without the supervision of a \
outs
snerd es op ane gaM
UF ewig ord oi se OBE
harming Ae tie nde supiets ie Ope a
irning process is independent. During the training of ANN under unsupervised learning, the
J Grmad 6s $ opti
ita
29 4 tainty
tee ocat X input) Neural '
Network —> — ¥(Actual output)
teacher. This ea
Put vectors of similar type are combined to form cluster
ee in rs, When a new input pattern is applied, then the
hee ¥ een an output response indicating the class to which the input pattern belongs. There is
peneinnt pe ar ee ronment as to what should be the desired output and ifitis correct or incorrect.
learning the network itself must discover the patterns and features from the input
ata, andthe relation for the input data over the output.
Bante ¢ te msey
ee five oer ahve ‘or tale Wh Mer thay
Suan nbh y Ge A, fe
Greater Ron Wosbg tf Ries vee Fc
che , |
5 op tassios niente , (PTo« 2E) |
i4
1) Linear Activation Function: Its also called the Identity function as it performs no input editing, It can be
defined as: F(x) = x
1). Sigmold Activation Function: Its of two type as follows ~
+ Binary sigmoldal function: This activation function performs input editing between 0 and 1. It is
positive in nature. I is always bounded, which means its output cannot be less than 0 and more
Than 1. It is also strictly Increasing In nature, which means more the input higher would be the
output, It can be defined as
ov
F(x) =sigm(x)=11 +exp(-x)F(x)=sigm(x)=11 +exp(-»)
+ Bipolar sigmoldal function: This activation function performs input editing between -1 and 1. It
‘can be positive or negative in nature. It is always bounded, which means its output cannot be less
than -1 and more than 1, It is also strictly increasing in nature like sigmold function. It can be
defined as
F(x)=sigm(x)=21+exp(~x)-1=1-exp(x)1+exp()
WHAT IS A NEURAL NETWORK ACTIVATION FUNCTION?
Ina neural network, inputs, which are typically real values, are fed into the neurons in the network. Each
neuron has a welght, and the inputs are multiplied by the weight and fed into the activation
function, Each neuron’s output Is the input of the neurons in the next layer of the network, and so the
Inputs cascade through multiple activation functions until eventually, the output layer generates a
prediction. Neural networks rely on nonlinear activation functions—the derivative of the activation
function helps the network learn through the backpropagation process
Input | ———> f | OutputSOME COMMON ACTIVATION FUNCTIONS INCLUDE THE FOLLOWING:
n zero and one. For ve
1. The sigmoid function has a smooth gradient and outputs values betweer ni =
Nhat vauescfneipucparanecrs te newark sen bevery so Fach recon
called the vanishing gradient problem. 1
2 ‘The Tani inetion szer-cenered making easter to model Inputs thatare strongly negative
strongly positive or neutral.
2. Thee funcion sng conputtonalyeMcent buts notable to proces inputs that
approach zero or negative. -
The Leaky Rela function has a small postive slope ints negative area, enabling ito procese zero
or negative values.
The Parametric ReLu function allows the negative slope to be learned, performing
backpropagation to learn the most effective slope for zero and negative input values.
foRmax is a special activation function use for output neurons. It normalizes outputs for each class
between O and 1, and returns the probability that the input belongs to a specific class.
‘Swish Isa new activation function discovered by Google researchers. It performs better than ReLu
with a similar level of computational efficiency.
1
Nop woe
— Stemoid
os | Tnh
os.
3. Fuzzy Logic: Theory of approximate reasoning,
4. Artificial Life: Evolutionary Computation, Swarm Intelligence,
5. Artificial Immune System: A computer program based on the
6 Medical: At the moment, the research is mostly on modelling parts of the human body and
recognizing diseases from various scans (e.g cardiograms, CAT scans, ultrasonic
biological immune system,
scans, etc). Neural
Provide a specific
details of how to
networks are ideal in recognizing diseases using scans since there is no need to
algorithm on how to identify the disease. Neural networks learn by example so th
recognize the disease are not needed. What is needed is a set of examples that are Tepresentative of
all the variations of the disease, The quantity of examples is not as important as the ‘quantity’, The
examples need tobe selected very carefully f the system Isto perform reliably and efMcienty®
7. Computer Science: Researchers in quest of artificial intelligence have created spin offs lke dynamic
programming, object oriented programming symbolic programming, intelligent stories
management systems and many more such tools. The primary goal of creating an artificial
Intelligence still remains a distant dream but people are getting an idea of the ultimate path, which
could lead to it.
8, Aviation: Airlines use expert systems in planes to monitor atmospheric conditions and system
status. The plane can be put on autopilot once a course Is set for the destination.
9, Weather Forecast: Neural networks are used for predicting weather conditions. Previous data is fed
toa neural network, which learn the pattern and uses that knowledge to predict weather paterns,
10, Neural Networks in business: Business is a diverted fleld with several general areas of
specialization such as accounting or financial analysis. Almost any neural network application
‘would fit into one business area or financial analysis.
11, There is some potential for using neural networks for business purposes, including resource
allocation and scheduling,
412, There is also a strong potential for using neural networks for database mining which is, searching
for patterns implicit within the explicitly stored information in databases. Most of the funded work
In this area is cassified as proprietary. Thus it fs not possible to report on the fll extent of the
‘work going on. Most work is applying neural networks, such as the Hopfield-Tank network for
optimization and scheduling,
13, Marketing: There is a marketing application which has been integrated with neural network
system, The Airline Marketing Tactician (a trademark abbreviated as 'AMT) is a computer system
sorte of various intelligent technologies including expert systems. A feed forward neural network is
integrated with the AMT and was trained using back-propagation to assis the ‘marketing control of
vinline seat allocations. The adaptive neural approach was amenable to rule expression.
‘Additionally, the application's environment changed rapidly and constantly, which required a
continuously adaptive solution.
14, Credit Evaluation: The HNC company, founded by Robert Hecht-Nielsen, has developed several
Treural network applications. One of them is the Credit Scoring system which increases the
profitability of the existing model up to 27%. The HNC neural systems were also applied to
Trortgage sereening. A neural network automated mortgage insurance under writing SHAN Ree
developed by the Nestor Company. This system was trained with 5048 applications of ‘which 2597
aoe cortified, The data related to property and borrower qualifications. In a conservative mode
the system agreed on the under writers on 97% ofthe cases. In the liberal model the System agreed
49 ofthe cases. Tiss system run on an Apollo DN3000 and used 250K memory while processing
a case file in approximately 1 sec.
ADVANTAGES OF ANN
4. Adaptive learning: An ability to learn how todo tasks based on the data given for training or initial
experience.
Sisait Organisation: An ANN can create its own organisation or representation of the information it
recelves during learning time.
Fite Time Operation: ANN computations may be carried out in parallel, and special hardware devices are
being designed and manufactured which take advantage of this capability.
ae ies ae ition: Is a powerful technique for harnessing the information inthe data and generalizing
‘about it, Neural nets learn to recognize the patterns which exist n the data set.
5. The system is developed through learning rather than programming. Neural nets teach themselves the
patterns in the data freelng the analyst for more interesting work.Ry
ks may take some ¢
h neural networl ime
nt. iy changing information.
pting Se eeracall approaches fall. Because neu
r co
wironmer
6. Neural networks are flexible in a changing enviro
data which is too difficult to modey
tat aday
rai sudden drastic change they are excelene a 2eaP
7. Neural networks can bulld aay ees they can aially model i
networks can handle very complex inte eal tastes or programminglOgic —
with traditional approaches such as inferé sespleal statistical
3d as cl re of the data in
a Performance ee ena eae ‘are more reflective of the structu
problems, The neural netwo
significantly less time.
‘LIMITATIONS OF ANN
its i is a Limitation
Im this technological era everything has Merits and some Demerits in others ea Ey ascent
with every system which makes this ANN technology weak in some points. The vai
are
1) ANN is not a daily life general purpose problem solver.
2) There sno structured methodology available in ANN,
2) There is no single standardized paradigm for ANN development.
4 The Output Quality of an ANN may be unpredictable,
2) Many ANN Systems does not describe how they solve problems.
6) Black box Nature
7) Greater computational burden,
8) Proneness to over fitting,
9) Empirical nature of model development.
Forward Pass: The forward pass takes the JnPuts, passes them through the network and allows each
paron to react to a fraction of the Input. Neurone {generate thelr outputs and pass them on te the next
‘ayer, until eventually the network generates se output.
Error Function: Defines how far the actual wae ent Of the current model is from the correct output. When,
thane the model, the objective isto minimize the error function and bring output as close ae Possible to
lue,Backpropagation: In order to discover the optimal weights for the neurons, we perform a backward
pass, moving back from the network's prediction to the neurons that generated that prediction. This is
called backpropagation. Backpropagation tracks the derivatives of the activation functions in each
successive neuron, to find weights that bring the loss function to a minimum, which will generate the
est prediction. This is a mathematical process called gradient descent.
Blas and Variance: When training neural networks, ike in other machine learning techniques, we try to
balance between bias and variance. Bias measures how well the model fits the training set—able to
correctly predict the known outputs of the training examples. Variance measures how well the model
‘works with unknown inputs that were not available during training. Another meaning of bias is a “bias
heuron” which is used in every layer of the neural network. The bias neuron holds the number 1, and
‘makes it possible to move the activation function up, down, left and right on the number graph.
Hyperparameters: A hyper parameter is a setting that affects the structure or operation of the neural
network. In real deep learning projects, tuning hyper parameters is the primary way to build a network
that provides accurate predictions for a certain problem, Common hyper parameters include the number
‘of hidden layers, the activation function, and how many times (epochs) training should be repeated.
‘McCULLOGH-PITTS MODEL
1n 1943 two electrical engineers, Warren MeCullofgh and Walter Pitts, published the first paper describing
‘what we would call a neural network.
# § aagregs Re Um farm Fake
@ lection boelen Rt apg:
+ ip Cambe encifeory om irtibiheny
© JO ang win inhibifeig, eb
Flaten agen) > Ex
a PF GAH el '{ wre
no) alae 20 14 Fac O aif
TCO Recall Kyrahaichny Prxemety tne {Dr
His cathe Reeshstbry logre OU ifpar beng
Iemay be divided into 2 parts. The first part, g takes an input, performs an aggregation and based on the
aggregated value the second part, f makes a decision. Let us suppose that I want to predict my own.
Gecision, whether to watch a random football game or not on TV. The inputs are all boolean ie, {0,1} and
my output variable is also boolean {0: Will watch it, 1: Won't watch it).
0.1
as ‘we {0,1}
0, x1 could be ‘is Indian Premier League On’ (I like Premier League more)
X2 could be ‘is ita knockout game (I tend to care less about the league level matches)
3 could be ‘is Not Home’ (Can't watch it when I'm in College. Can I?)
4 could be ‘is my favorite team playing’ and so on.
‘These inputs can either be excitatory or inhibitory. Inhibitory inputs are those that have maximum effect
on the decision making irrespective of other inputs Le. ifX3{s 1 (not home) then my output will always be
O4e, the neuron will never fire, so X3 is an inhibitory input. Excitatory inputs are NOT the ones that will
‘make the neuron fire on their own but they might fie it when combined together, Formally, this is what Is
going on:smelt Seema Arms out tbe 2 OF ore.
‘We can see that g(x) is just doing
thresholding parameter. For exami
the theta is Zhere. This is called the
se McCulloch Pits neural models als known aslinea’ thresh orinpuns into two different
mathematically using
sensed one output'y The inear threshold gate simply
aero Thus the output y is binary-Such a function can be
x
Sum = OW,
a
Where, Wi, W2,Ws,-—, Wm are weight values normalized in the range of either (0,1) or (—1, I) and
associated with each inputline, Sum is the weighted sum, and T is a threshold constant. The function fs a
Tinea step function at threshold T’ as shown in figure 2.3. The symbolic representation of the linear
threshold gate is shown in figure below.
y
oa “Sum
Linear Threshold Function
Inputs Weights
Threshold T
Symbolic lustration of Linear Threshold Gate| ®
In any Boolean function, all Inputs are Boolean and the output is also Boolean. So essentially, the neuron is
Just trying to learn a Boolean function,
mh :
2- Cp {0,1}
z
‘This representation just denotes that, for the boolean Inputs x.1,x2 and x.3 the g(x) |, sum z theta, the
neuron will fire otherwise, it won't,
AND Function
a
| a ye {0,1}
ts
‘An AND funetion neuron would only fire when ALL the inputs are ON Le, g(x) 2 3 here.
OR Function’
cal
|
| 2 «ye {0,1}
Ts
For an OR function neuron would fire if ANY of the Inputs Is ON Le. g(x) 2 1 here.
NOR Function
dy wy
0
Go
wo ye {0,1}
% Ane 14,
ay
For a NOR neuron to fire, we want ALL the Inputs to be 0 so the thresholding parameter should also be 0
‘and we take them all as inhibitory Input. neNOT Rnetton
n—4 0 -y €{0, 1}
so we take the (nput as al inhibitory input and set the
Por aNOT neuron, 1 OutpUts and O outputs L
thresholding parameter wd,
We can summarize these rutes with the Mecutough-Pitts output rule ws:
ing potential, It also has
‘The MeCutloch-Pitts model of a neuron is siyple yo has substantial comput
ea aeratea dition, However, this mode) 6 <0 smpliste that tt only genveraes 1 binary you
pred also the weight and threshold values are roo rae neural computing algorithm has diverse feature
ged alsa applications, Thus, we need to, obra The neural modal with more flexible computational
features. .
earning rule is a method or a mathematteal owe IBIS Neural Network to learn from the existing
conditions and improve its performance, Thus learning rules updates the welghts and bias levels of a
)
‘A multilayer perceptron is quite similar to a modern neural network. By adding a few ingredients, the
perceptron architecture becomes a full-fledged deep learning system:
+ Activation functions and other hyperparameters: a full neural network uses a variety of
activation functions which output real values, not boolean values like in the classic perceptron. Itis
‘more flexible in terms of other details of the learning process, such as the number of training
iterations (iterations and epochs), weight initialization schemes, regularization, and so on. All these
can be tuned as hyperparameters.
« Backpropagation: a full neural network uses the backpropagation algorithm, to perform iterative
‘backward passes which try to find the optimal values of perceptron weights, to generate the most
accurate prediction,
+ Advanced architectures: full neural networks can have a variety of architectures that can help
solve specific problems. A few examples are Recurrent Neural Networks (RNN), Convolutional
‘Neural Networks (CNN), and Generative Adversarial Networks (GAN).
WHAT IS BACKPROPAGATION AND WHY IS IT IMPORTANT?
‘After a neural network Is defined with initial weights, and a forward pass is performed to generate the
initial prediction, there is an error function which defines how far away the model is from the true
prediction. There are many possible algorithms that can minimize the error function—for example, one
could do a brute force search to find the welghts that generate the smallest error. However, for large
neural networks, a training algorithm is needed that is very computationally efficient. Backpropagation is
that algorithm—it can discover the optimal weights relatively quickly, even for a network with millions
of weights.
HOW BACKPROPAGATION WORKS?
1. Forward pass—weights are initialized and inputs from the training set are fed into the
network, The forward pass is carried out and the model generates its initial prediction.
2. Error function—the error function is computed by checking how far away the prediction is
from the known true value.
3, Backpropagation with gradient descent—the backpropagation algorithm calculates how
much the output values are affected by each of the weights in the model. To do this, it
calculates partial derivatives, going back from the error function to a specific neuron and its
‘weight. This provides complete traceability from total errors, back to a specific weight which
contributed to that error. The result of backpropagation is a set of weights that minimize the
error function.
4. Weight update—weights can be updated after every sample in the training set, but this is
usually not practical. Typically, a batch of samples is run in one big forward pass, and then
backpropagation performed on the aggregate result. The batch size and number of batches
used in training, called iterations, are important hyperparameters that are tuned to get the
best results. Running the entire training set through the backpropagation process is called
‘an epoch,— EE
‘Training algorithm of BPNN:
the pre connected path
through Par eights W. The weights are usually randomly selected.
. Inputs X, arrive
from the input layer, to the hidden layers, to the
1
2. Inputis modeled using
3. Calculate the output for every neuron
output layer.
‘4, Calculate the error in the outputs
Errore= Actual Output - Desired Output
5, Travel back from the output layer to the hidden layer to adjust the weights such that the
erroris decreased.
Keep repeating the process until the desired output is achieved
Architecture of back propagation network:
As shown in the diagram, the architecture of BPN has three interconnected layers having weights on them.
‘the hidden layer as well asthe output layer also as bias, whose weight s always 1, on them. Asis clear
from the diagram, the working of BPN is in two phases. One phase sends the si
5 ignal from the input layer to
the output layer, and the other phase back propagates the error from the output layer to the inp layerUNIT-2
WHAT ARE THE ANN LEARNING PARADIGMS?
Learning can refer to either acquiring or enhancing knowledge. As Herbert Simon says, Machine
Learning denotes changes in the system that are adaptive in the sense that they enable the system
to do the same task or tasks drawn from the same population more efficiently and more effectively
the next time.
ANN learning paradigms can be classified as supervised, unsupervised and reinforcement learning.
Supervised learning model assumes the availability of a teacher or supervisor who classifies the
training examples into classes and utilizes the information on the class membership of each
training instance, whereas, Unsupervised learning model identify the pattern class information
heuristically and Reinforcement learning learns through trial and error interactions with its
environment (reward/penalty assignment).
Though these models address learning in different ways, learning depends on the space of
interconnection neurons. That is, supervised learning learns by adjusting its inter connection
weight combinations with the help of error signals where as unsupervised learning uses
information associated with a group of neurons and reinforcement learning uses reinforcement
function to modify local weight parameters. Thus, learning occurs in an ANN by adjusting the free
parameters of the network that are adapted where the ANN is embedded.
BRIEFLY EXPLAIN SUPERVISED LEARNING,
Supervised learning is based on training a data sample from data source with correct classification
already assigned. Such techniques are utilized in feed forward or Multi Layer Perceptron (MLP)
models. These MLP has three distinctive characteristics:
1. One or more layers of hidden neurons that are not part of the input or output layers of the
network that enable the network to learn and solve any complex problems
2. The nonlinearity reflected in the neuronal activity is differentiable and,
3. The interconnection model of the network exhibits a high degree of connectivity These
characteristics along with learning through training solve difficult and diverse problems. Learning
through training in a supervised ANN model also called as error backpropagation algorithm. The
error correction-learning algorithm trains the network based on the input-output samples and
finds error signal, which is the difference of the output calculated and the desired output and
adjusts the synaptic weights of the neurons that is proportional to the product of the error signal
and the input instance of the synaptic weight. Based on this principle, error back propagation
earning occurs in two passes:
Forward Pass: Here, input vector is presented to the network. This input signal propagates
forward, neuron by neuron through the network and emerges at the output end of the network as
output signal:
y(n) = @(v(n)), where v(n) is the induced local fleld of a neuron defined by v(n) =£ w(n)y(n).
The output that Is calculated at the output layer o{n) {s compared with the desired response d(n)
and finds the error e(n) for that neuron. The synaptic weights of the network during this pass are
remains same,Reed fom rg rot
2. Sigmar furehem + 4K )e
ee Vere
Se =
3. Fierce - whe Wear fanctir 4)
repos” i
Fone te eeye'S “
” Mp eS
os Sigel funcker
14 abet
Hn) > de
yecal neural refuerk
betraon biologi g
{ma} es
[Hicblen cutpat bype |__> culpufs
© ap pete
4. Bimry 5 _—
mea errclion a, Anak 5
ee eee, Cant a Kot valine Atl ces ichs ae
petit Fler acr £
Pals © da heo - .
Limite, | PPPS,
— 1 lowe ave om De if- Orit prrivite. mall :
onl ith The oof ite Me mrt cot
pola fomckin B27, DAK Cattect a Nivohrena om A bed
Lyin stits bee kpropey wfrim arp greotint a pes
Pamchese oy ee *! Fx Civeoe achrackin Pinckio alto finen as”!
iS (hAG Srekin', a hure % actration es pat ene
x eee Sire! bee prams 4 reat Uy lager o real On Ye.
a > olecnth oe ayy dg KB wig had Lom 9 Hs pf :
Lt PE ap 4 tales d 08 phen o, pe
Lota teen Dib if posite &
cea Ce a gO eT pn ml che
SPrckin ta coufock cota pr ith Me (oti tan 1% fometi
Bier epee 0 SS kcshyopuputonse woth Bplpap as paaslal Arey ere
Met TMs siapel iro oni a toy
Si Ft ef ty RS ts ateO besom 4 ds Urrniters po
tal? A. Ra akinte ni dacs SF afte.) a rmdel & Crab. complen wapp
G4
2, Cal :
zc ight er loss: When we pass the data instance (or one example) we will get some output
pa al at is called Predicted output (pred_out) and we have the label with the data that
real spate expected output{[Link]), Based upon these both we calculate the loss that we
¢ to backpropagate(using Backpropagation algorithm). There is various Loss Function that we
use based on our output and requirement.
4, Backward Pass; After calculating the loss, we back propagate the loss and update the weights of
the model by using gradient. This is the main step inthe training ofthe model. In this step, weights
will adjust according to the gradient flow in that direction.
Applications of MLP:
1. MLPs are useful in research for their ability to solve problems stochastically, which often
allows approximate solutions for extremely complex problems like fitness approximation.
2, MLPs are universal function approximators and they can be used to create mathematical
models by regression analysis,
3, MLPs are a popular machine learning solution in diverse fields such as speech recognition,
{mage recognition, and machine translation software,
ELECTRIC LOAD FORECASTING USING ANN
[ANNs were first applied to load forecasting n the late 1980's. ANNs have good performance in data
ntelfeation and function fiting, Some examples of utilizing ANN in Pewet system applications
fare. Load forecasting, fault classification, power system ‘assessment, real time harmonic evaluation,
power factor correction, oad scheduling, design ‘of transmission lines, and power system planning.
Powel anecast has been an attractive research topic for many decades and in many countries all
war the worl, especially in fast developing countries with higher load growth rate. Load forecast
an be generally classified into four categories based on the forecasting time as detailed in the table
© farmetrie RLV
NW.
elt
@® sev
‘An ANN for load forecasting can be trained on a training set of data that consists of time-lagged
ao aia and other nor-load parameters such as weather data, time of day, day of week, month,
toa tual load data. Some ANNS are only trained agent days with data similar to the forecast day.
aana'tne network has been trained tis tested by presenting It with predictor data inputs. The
predictor data canbe time-lagged oad data and forecasted ‘weather data (for the next 24 hours).
Prercrecasted load output from the ANN is compared tothe actual load fo ‘determine the forecast
caer Forecast error is sometimes presented in terms ofthe root mean square error (RMSE) but
QO Neo: Vino AP, Men: tdwcar aahvastin fonvtor sil fx lirns‘beion 4 Litweer fonhere -
ag ‘allow batlgropapetere Kaos. ro? Ke cbnlate Perchen rratlol neakeol &
¥
an
sa below.
ioe
5 8g Toad fore Period importance
| ‘Long term ‘One year to ~ To calculate and to allocate the: required future capacity.
| $8 re year ©] ¢ Toca t pner stations to facgtomer requirements
Pay «Plays an essential role to determine future budget.
| se Medium term She week to | Fuel allocation and maintenance schedules.
| ged Pabiors
og eae [eee eevee rane
7 week ‘« To evaluate economic dispatch, hydrothermal ‘co-ordination,
5 ‘er stn
t __ meme ae ong ter manasa non
@ > Very short term ‘One minute | Energy management systems (EMS).
a ee
R
LID ner lime ae
®
@e.
ened His powsiba. k ga ack onel conctoslore wiek Ls rm Ai tp rmorers
en ties written
lay. able 4 paps,
Be Pullman off Bees con seas
|“Actinhin fi abso ' a poi
sapnoel ape Ad San op vatan
lakes ong \
okt
Fe ke fe) ore +e),
ip op (me
BB, vaherens 50 :
; ol
spon aan wt iO Pe am af Cb
Go ‘nse FANE & preolist Ms probabil ao UY 1 t
rs cx aipforontebl ans proviets & smh
of ofp valuo: eh ex (othe th. 2-4
d Limteken mp-teo Bankr’ is s
ea: ee Op gel eT suse
me ou Saprifiert a ~o at&t; a aah Dae fot
> 1 ae
puck fale oi Mer Se ey bate
- u sonpltcs fale fr vihes gee Pon § 07 bas Har
Vey Snrell grrotenk dee (4a) 2 Prete” TO,"
Livi. gfern F% ”
. Ke greclint vi ppreaches canes & Learn Oral Suffers
Fa haw vey onbe SH OOL
achehact feck, andl om ads cope
27k Be oliffeemts «= We rR “U6 |
WB Trak, fh larger Up (arom pve) SG cleier ¢ Vale oot!
ee 19, tokerean Ur Srraller Mi fp (more Ws) fh char Ra Ip will Gh O.
Acvenless 2 7m ap a tank oblakcircflencin is taro Corttrrel, Kowms 4% Som
casily rape A of values 0 sheng ave, newhal shony'y pve
bY tbeof v0! Aiello lagers yrrer nn as Ur valor bie bf 9&4
Kiker, (hi mresen opr Ka biololn lager Comme att & & Omvery Con Rib.
HE helps ons Corton os eta anol “yradas Learning fry Ke meat dae mat
yard (olertachr< )
easier
fie) ageparateal 1
Haida he You and ale faces Ki
problem off Vani hng greeted simile ® Ms Sipraid
F Sabrrntn 7
eee -
2 aE ecureJob name: (none)
Document name: (none)
I
age number: 8
Error: XPS format eror (194,330)
O RLU > Kectifid Lincar undl-}
- a ges on imp main cy loner
fendi, keLO re
ei enol allows for prroprackons
hele Sinuffame 2eD, a fiaermen(e
Compulatte erint : each eckine 2 FE
2 lee] hee Oo 7
Denrent ap OS Sam
Achor dese D nan -bveer p00) ron (0,401) — mre HD) i
One vt cy oT HVE region
© Computetionty” geeegennve
9) converge = qpater > sipmeid onl sp terh
Diaslvew
Pe ean Ripper tiable > atn=o pees ?
beak rormalired jer (
pnsich 2 Ur ad resolve y
Cue conberen > fut Whe $19
tow Lo os bopper (hictbn Oger cel oe) bY ya) Oe
@ kw Sofiprig fe i RLU preblies eo
® 62 LU: ReLU & am Impronet
ee nes We Dying RLU
Vesim Y RLU.
mall qve Slope om Ty
probe te U-bas a
pve orta LL pe
A 1 Te adlvenian. ¢ RLV Ae Lorre ob Hey: wan (o.1 x, >)
teak 4 RLU, jnaddlitin &'Th Youd Tut ix otees ematle ftv: 30
beokpropaqate , eun fr 2—VE Vp valuex soles
————_>
bir e aig
Fionn. a. prnaltchiins trea rit be Cntit bad cfr Lay Be We (danvabve)
Megatre >/p Valins
Ti Gveoknt ave Values 4 moll Mah rakes earns! motel
forthe. ak gi Sree ee eT