SlideShare una empresa de Scribd logo
1 de 26
Neural Network Application in System Design
CHAPTER 1
Introduction to Neural Networks
Work on artificial neural networks commonly referred to as "neural networks"
(NN) has been motivated right from its origin by the recognition that the human brain
computes in an entirely different way than the conventional computer. The brain is a
highly complex, nonlinear and parallel computer (information processing system). It has
the capability to organize its structural constituents, known as neurons, so as to perform
certain computations (e.g. pattern recognition, perception, and motor control) many times
faster than the fastest digital computer in existence today. Consider for example, human
vision, which is an information-processing task. It is the function of the visual system to
provide a representation of the environment around us and, more important, to supply the
information we need to interact with the environment. To be specific, the brain routinely
accomplish perceptual recognition task (e.g. recognizing a familiar face embedded in an
un-familiar scene) in approximately 100-200 ms, where as tasks of much lesser
complexity may take days on a conventional computer.
How, then, does a human brain do it? At birth, a brain has great structure and the
ability to built-up its own rules through what we usually refer to as "experience". Indeed,
experience is built up over time, with the most dramatic development (i.e. hard wiring) of
the human brain taking place during the first two years from birth: but the development
continues well beyond that stage.
A "developing" neuron is synonymous with a plastic brain: Plasticity permits the
developing nervous system to adapt to its surrounding environment. Just as plasticity
appears to be essential to the functioning of neurons as information-processing units in
the human brain, so it is with neural networks made up of artificial neurons. In its most
general form, a neural network is a machine that is designed to model the way in which
the brain performs a particular task or function of interest; the network is usually
Dept of E&C, SJCE Mysore Page 1
Neural Network Application in System Design
implemented by electronic components or is simulated in software on a digital computer.
The interest is confined to an important class of neural networks that perform useful
computations through a process of learning. To achieve good performance, neural
networks employ a massive interconnection of simple computing definition of a neural
network viewed as an adaptive machine.
A neural network is a massively equivalent distributed process or made up of simple
processing units, which has a natural propensity for storing experiential knowledge and
making it available for use. It resembles the brain in two respects:
 Knowledge is acquired by the network from its environment through a learning
process.
 Inter neuron connection strengths, known as synaptic weights, are used to store
the acquired knowledge.
1.1Neuron & Artificial Neuron
The two figures 1 & 2 compare the human neuron and the artificial neuron. For the
human neuron the main functioning parts are:
Figure 1: Human Neuron
Dept of E&C, SJCE Mysore Page 2
Neural Network Application in System Design
 Dendrites: These act as the input points to the main body of the neuron.
 Synapse: This is the storage area of the past experience.
 Soma: It receives synaptic information and performs further processing on the
information.
 Axon: This is the output line for the neuron.
Figure 2: Artificial Neuron
In artificial neural networks, the synaptic and somatic operations are emulated as
follows:
 Synaptic Operation: The input weights act as storage for knowledge (and
therefore, as memory for previous experiences).
 Somatic Operation: The somatic operation is provided by various mathematical
operations such as aggregation, thresholding, nonlinear activation and dynamic
processing to the synaptic inputs.
Dept of E&C, SJCE Mysore Page 3
Neural Network Application in System Design
1.2Adaptation in NN’s
The procedure that is used to perform the learning process is called a learning
algorithm (fig. 3), the function of which is to modify the synaptic weights of the network
in an orderly fashion to attain a desired design objective.
Figure 3: Adaptation in NN’s
The modification of synaptic weights provides the traditional method for the design of
neural networks. Such an approach is the closest to linear adaptive filter theory, which is
already well established and successfully applied in many diverse fields. However, it is
also possible for a neural network to modify its own topology, which is motivated by the
fact that neurons in the human brain can die and then new synaptic connections can grow.
1.3 Common Neural Network Architectures
The manner in which the neurons of a neural network are structured is intimately
linked with the learning algorithm used to train the network. We may therefore speak of
algorithms (rules) used in the design of neural networks as being structured. In general
we may identify three fundamentally different classes of network architectures:
Dept of E&C, SJCE Mysore Page 4
Neural Network Application in System Design
1.3.1 Single-Layer Feed-forward Networks
In a layered neural network the neurons are organized in the form, of layers. In
the simplest form of layered network, we have an input layer of source nodes that
projects onto an output layer of neurons (computation nodes), but not vise versa. In other
words, this network is strictly a feed-forward or acyclic type. It is illustrated in the figure
for the case if four nodes in both the input and output layers. Such a network is called a
single-layered network, with the name "single-layer" referring to the output layer of
computation nodes (neurons). We do not count the input layer of source nodes because
no computation is performed there.
Figure 4: Single Layer Feedforward NN
1.3.2. Multilayer Feed-Forward Networks
The second class of a feed-forward neural network distinguishes its self by the
presence of one or more hidden layers, whose computation nodes are correspondingly
called hidden neurons or hidden units. The function of the hidden neuron is to interfere
between the external input and the network output in some useful manner.
Multilayer feed forward networks are an important class of neural networks.
Typically, the network consists of a set of sensory units (source nodes) that constitute the
input layer, one or more hidden layers of computation nodes, and an out-put layer of
computation nodes. The input signal propagates through the network in a forward
Dept of E&C, SJCE Mysore Page 5
Neural Network Application in System Design
direction, on a layer-by-layer basis. These neural networks are commonly referred to as
multilayer perceptions (MLP’s), which represent a generalization of the single-layer
perception.
The source nodes in the input layer of the network supply respective elements of
the activation pattern (input vector), which constitute the input signals applied to the
neurons (computation nodes) in the second layer (i.e., the first hidden layer). The output
signals of the second layer are used as an input to the third layer, and so on for the rest of
the network. Typically the neurons at each layer of the network have as there inputs the
outputs of the preceding layers only. The set of output signals of the neurons in the output
(final layer) constitutes the overall response of the network to the activation pattern
supplied by the source nodes in the input (first) layer.
Multilayer perceptrons have been applied successfully to solve some difficult and
diverse problems by training them in a supervised manner with a highly popular algo-
rithm known as the error back-propagation algorithm. This algorithm is based on the
error-correction learning rule. As such, it may be viewed as a generalization of an equally
popular adaptive filtering algorithm: the least-mean-square (LMS) algorithm.
Basically, error back-propagation learning consists of two passes through the dif-
ferent layers of the network: a forward pass and a backward pass. In the forward pass, an
activity pattern (input vector) is applied to the sensory nodes of the network, and its effect
propagates through the network layer by layer. Finally, a set of outputs is produced as the
actual response of the network. During the forward pass the synaptic weights of the
networks are all fixed. During the backward pass, on the other hand, the synaptic weights
are all adjusted in accordance with an error-correction rule. Specifically, the actual
response of the network is subtracted from a desired (target) response to produce an error
signal. This error signal is then propagated backward through the network, against the
direction of synaptic connections, hence the name "error back-propagation." The synaptic
weights are adjusted to make the actual response of the network move closer to the
desired response in a statistical sense. The error back-propagation algorithm is also
referred to in the literature as the back-propagation algorithm.
Dept of E&C, SJCE Mysore Page 6
Neural Network Application in System Design
Figure 5 shows the architectural graph of a multilayer perceptron with two hidden
layers and an output layer. Signal flow through the network progresses in a forward
direction, from left to right and on a layer-by-layer basis.
Figure 5: Multi-layer Feedforward NN’s
The neural network in the figure is said to be fully connected in the sense that
every node in each layer of the network is connected to every other node in the adjacent
forward layer. If, however, some of the communication links are missing from the
network, we say that the network is partially connected.
1.3.3. Recurrent Networks
A recurrent neural network distinguishes itself from the feed-forward network in
that it has at least one feedback loop. For example, a recurrent network may consist of a
single layer of neurons with each neuron feeding its output signal back to the input of all
input neurons.
The presence of feedback loops has a profound impact on the learning capability
of the network and on its performance. Moreover, the feedback loops involve the use of
Dept of E&C, SJCE Mysore Page 7
Neural Network Application in System Design
particular branches composed of unit-delay elements which result in a nonlinear
dynamical behavior, assuming that the neural network contains nonlinear units.
1.4 Applications of Neural Networks
Neural networks are applicable in virtually every situation in which a relationship
between the predictor variables (independents, inputs) and predicted variables
(dependents, outputs) exists, even when that relationship is very complex and not easy to
articulate in the usual terms of "correlations" or "differences between groups." A few
representative examples of problems to which neural network analysis has been applied
successfully are:
 Detection of medical phenomena: A variety of health-related indices (e.g., a
combination of heart rate, levels of various substances in the blood, respiration
rate) can be monitored. The onset of a particular medical condition could be
associated with a very complex (e.g., nonlinear and interactive) combination of
changes on a subset of the variables being monitored. Neural networks have been
used to recognize this predictive pattern so that the appropriate treatment can be
prescribed.
 Stock market prediction: Fluctuations of stock prices and stock indices are
another example of a complex, multidimensional, but in some circumstances at
least partially-deterministic phenomenon. Neural networks are being used by
many technical analysts to make predictions about stock prices based upon a large
number of factors such as past performance of other stocks and various economic
indicators.
 Credit assignment: A variety of pieces of information are usually known about
an applicant for a loan. For instance, the applicant's age, education, occupation,
and many other facts may be available. After training a neural network on
historical data, neural network analysis can identify the most relevant
characteristics and use those to classify applicants as good or bad credit risks.
Dept of E&C, SJCE Mysore Page 8
Neural Network Application in System Design
 Condition Monitoring: Neural networks can be instrumental in cutting costs by
bringing additional expertise to scheduling the preventive maintenance of
machines. A neural network can be trained to distinguish between the sounds a
machine makes when it is running normally ("false alarms") versus when it is on
the verge of a problem. After this training period, the expertise of the network can
be used to warn a technician of an upcoming breakdown, before it occurs and
causes costly unforeseen "downtime."
 Engine management: Neural networks have been used to analyze the input of
sensors from an engine. The neural network controls the various parameters
within which the engine functions, in order to achieve a particular goal, such as
minimizing fuel consumption.
 Signature analysis: as a mechanism for comparing signatures made (e.g. in a
bank) with those stored. This is one of the first large-scale applications of neural
networks in the USA, and is also one of the first to use a neural network chip.
 Process control: most processes cannot be determined as computable algorithms.
Neural Networks can be used to adaptively control the process
 Nonlinear Identification & Adaptive Control: This is one of the main areas of
application of the neural networks. Neural Networks find applications in
situations where the plant dynamics are uncertain or un-modeled.
 Application of Neural Network on Flight Control:
Dept of E&C, SJCE Mysore Page 9
Neural Network Application in System Design
Fig. 1. Application of neural network in flight control
Fig. 2. Intelligent flight control system generation II system
Dept of E&C, SJCE Mysore Page 10
Neural Network Application in System Design
CHAPTER 2
LMS and RBF-NN’s
2.1 The Least Mean Square (LMS) Adaptation Algorithm
As discussed in section 1, the learning process in the neurons involves updating of
certain “weights”. A number of adaptation algorithms are available in literature. The
criteria/cost functions used for adaptation and the methods are usually derived from the
richly developed field of adaptive filter theory. In the following we present one of the
most commonly used adaptation algorithms, the LMS.
Let:
][nw : time varying neuron tap weights
][nv : input to neuron
][nd : desired response
[ ]ns~
: actual output of neuron
J : cost function (the mean square error)
The estimation error is the difference between the desired response and the
estimated output:
][~][][ nsndne −=
using ][][~ nvwns H
= gives,
][][][ nvwndne H
−=
The mean square error (cost function) is defined as:
Dept of E&C, SJCE Mysore Page 11
Neural Network Application in System Design
{ } { }22
][][][ nvwndEneEJ H
−==
Thus the cost function J is a function of vector w. Minimizing J with respect to
the complex tap weights w leads to the set of equations called the Wiener-Hopf
equations. If we limit the number of taps to M then we obtain the matrix formulation of
the Wiener- Hopf equations from which the solution is obtained as
pRwo
1−
=
where
{ }][][ nvnvER H
= and { }][][ *
ndnvEp =
This method of solution requires inversion of a matrix. An alternative adaptive
method of solution is the Steepest Descent algorithm. It can be shown that the cost
function J has the shape of an M-dimensional bowl whose minimum is at the optimal
solution of w. The steepest descent algorithm moves the tap weights towards the
minimum of the J bowl at every iteration by moving them in the direction opposite to the
gradient vector:
pRwJw 22 −=∇
Thus we have an iterative definition for the tap weight updates:
Jnwnw w∇−=+
2
][]1[
µ
]][[][ nRwpnw −+= µ
µ is known as the adaptation step. (µ > 0)
In practice, although we do not know R and p, we can use their instantaneous estimates:
][][][ˆ nvnvnR H
=
][][ˆ *
ndnvp =
Dept of E&C, SJCE Mysore Page 12
Neural Network Application in System Design
This approach is known as the Least Mean Squares (LMS) method which is a
member of a particular class of algorithms called the stochastic gradient algorithms. Thus
the adaptation equation is given as:
][][][ˆ]1[ˆ *
nenvnwnw µ+=+
As we shall see, this equation is used in a variety of adaptation algorithms each having its
own definition for the error functions.
2.2. RBF Neural Networks
Among the vast variety of neural networks, the RBF-NN is a quire commonly used
structure. The design of a RBF-NN in its most basic form consists of three separate
layers. The input layer is the set of source nodes (sensory units). The second layer is a
hidden layer of high dimension. The output layer gives the response of the network to the
activation patterns applied to the input layer. The transformation from the input space to
the hidden-unit space is nonlinear. On the other hand, the transformation from the hidden
space to the output space is linear.
Figure 6: RBF-NN Basic Structure
With reference to the figure above, the output y(t) is a weighted sum of the outputs of the
hidden layer, given by
Dept of E&C, SJCE Mysore Page 13
Neural Network Application in System Design
1
ˆ( ) ( ( ) ),
n
i i
i
y t w u t cφ
=
= −∑
(3.3)
where
( )u t is the input
( ).φ
is an arbitrary nonlinear radial basis function
.
denotes the norm that is usually assumed to be Euclidean
ic
are the known centers of the radial basis functions
iw
are the weights
Radial functions are a special class of functions. Their characteristic feature is that their
response decreases (or increases) monotonically with distance from a central point and
they are radially symmetric. The centre, the distance scale, and the precise shape of the
radial function are parameters of the model. There are a variety of radial functions
available in literature. The most commonly used one is the Gaussian radial filter, which
in case of a scalar input is
2
2
( )
( ) exp( )
x c
h x
β
−
= −
Its parameters are its centre c and its radius β (width), Figure 1.3 illustrates a Gaussian
RBF with centre c = 0 and radius β = 1. A Gaussian RBF monotonically decreases with
Dept of E&C, SJCE Mysore Page 14
Neural Network Application in System Design
distance from the centre
Figure 7: Gaussian function Profile
A summary of the characteristics of the RBF-NN’s is given below
 They are two-layer feed-forward networks.
 The hidden nodes implement a set of radial basis functions (e.g. Gaussian
functions).
 The output nodes implement linear summation functions as in an MLP.
 The network training is divided into two stages: first the weights from the input to
hidden layer are determined, and then the weights from the hidden to output layer.
 The training/learning is very fast.
 The networks are very good at interpolation.
2.3 Detailed Learning Algorithm for Rbf-Nn’s
The whole algorithm of RBF network learning may be split into two phases:
 Hidden layer learning or basis function selection (unsupervised learning).
 Fitting of outputs in a transformed feature space (supervised learning).
2.3.1 Unsupervised learning
The first phase is an unsupervised learning. It does not use any information on
target outputs and deals only with a set of inputs. At this stage we have to:
Dept of E&C, SJCE Mysore Page 15
Neural Network Application in System Design
1. Select a number of radial basis functions.
2. Select a center for each basis function and
3. Select a value for the parameter β (width), which characterizes the basis function
range of definition (the range of its influence). A too large value of β forms too narrow
basis functions.
In step1, usually all the RBF’s are chosen to be the same. There are many algorithms
available for the selection of the centers (step-2), one of the more popular ones is the K-
means clustering algorithm which goes as follows
Given m data points, select l as the number of clusters such that l < m
Take the first l learning data as the center vectors for the l clusters
, 1,2,...., ;j jc x j l= =
Assign the remaining data points to one of the clusters with the least distance criterion.
Recomputed the center vectors using the new mean, that is
1
; 1
j
j j
i cj
c x j l
m ∈
= ≤ ≤∑
where mj is the number of data points belonging to the jth cluster.
As soon as the clustering algorithm is complete we may move to the selection of the
variance or width parameter β (step-3). These parameters control the amount of overlap
of the radial basis functions as well as the network generalizations. A small value yields a
rapidly decreasing function, whereas a large value results in a more gently varying
function. The mostly commonly used method for the selection of the width parameter for
a cluster is to take it equal to the average distance between the data in the cluster and
center of the cluster.
Dept of E&C, SJCE Mysore Page 16
Neural Network Application in System Design
2.3.2 Supervised Learning
The second phase is a supervised learning. The goal is to fit outputs with a linear
function of nonlinear transformed inputs. Any gradient optimization method may be
used, but the LMS (discussed above) is used most often.
2.3.3 Relative Advantages of RBF-NN’s
 Many pattern recognition experiments show that the RBF-NN’s are superior over
other neural network approaches in the following senses.
 RBF-NN’s are capable of approximating nonlinear mappings effectively.
 The training time of the RBF-NN’s is quite low compared to that of other neural
network approaches such as the multi-layer perceptron.
 The RBF-NN’s produce classification accuracies from 5% to 10% higher than
accuracies produced by the back propagation algorithm.
 The RBF-NN’s are quite successful for identifying regions of sample data not in
any known class because they use a non-monotonic transfer function based on the
Gaussian density function.
Dept of E&C, SJCE Mysore Page 17
Neural Network Application in System Design
CHAPTER 3
Neural Network Applications
3.1 Nonlinear Plant Identification
One of the major areas of application of the neural networks is in the identification
of nonlinear plants. The RBF-NN was introduced in section 2. Here we make use of the
Gaussian RBF-NN to identify the nonlinear system known as the continuous stirred tank
reactor. The nonlinear model of the continuous stirred tank reactor when the sampling
time is chosen as 0.05 seconds is as follows.
2 3 4
2 3 2
2 2 2 3 3
( 1) 0.8606 ( ) 0.0401 ( ) 0.0017 ( ) 0.000125 ( ) 0.0464 ( )
0.045 ( ) ( ) 0.0034 ( ) ( ) 0.00025 ( ) ( ) 0.0012 ( )
0.0013 ( ) ( ) 0.0001458 ( ) ( ) 0.00002083 ( ) 0.00002083 ( ) ( )
y t y t y t y t y t u t
y t u t y t u t y t u t u t
y t u t y t u t u t y t u t
+ = − + − +
− + − − +
− + −
Figure 8: Identification Structure
The RBF-NN assumes no prior knowledge of the system parameters and tries to identify
the system online. Simulations were carried out using SIMULINK. The number of linear
combiner weights was chosen to be ten. The neural network worked well and was able to
identify the nonlinear plant online. The simulation results are given in the figures 9 & 10.
Dept of E&C, SJCE Mysore Page 18
Neural Network Application in System Design
0 20 40 60 80 100 120 140 160 180 200
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
0.6
Time
Plant Output
NN Output
Figure 9: Identification simulation
0 20 40 60 80 100 120 140 160 180 200
-1
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
0.6
0.8
1
Time
Error
Figure 10: Mismatch error
Dept of E&C, SJCE Mysore Page 19
Neural Network Application in System Design
3.2 Adaptive Tracking of Nonlinear Dynamic Plants
The adaptive control of nonlinear dynamic plants is an extremely important area of
research. It involves the online identification of the plant and development of a controller
based on this identified plant. The identification part is generally carried out using
powerful neural networks. A number of techniques have been suggested in literature.
Here we utilize one of the more recent approaches. The main structure is depicted below.
The basic structure is that of an IMC (Internal Model Control). The identification task is
carried out utilizing the Gaussian RBF-NN. A control law is synthesized which is based
on the identified system parameters.
Figure 11: Adaptive Tracking
3.2.1 The Plant
We assume a stable nonlinear dynamic plant whose functional parameters or the
functional structure need not be known.
3.2.2 The Identifying Model
We identify the plant online using the radial basis function with the following
structure:
1 21( ) ( 1) ( ( 1)) ( ( 2)) ......... ( ( ))nMy t a u t b u t b u t b u t n
∧ ∧ ∧
= − + Φ − + Φ − + + Φ −
(3.1)
Dept of E&C, SJCE Mysore Page 20
Neural Network Application in System Design
where the parameter 1a is selected in advance and the parameters nbbb
∧∧∧
,......, 21
are
estimated using the normalized least mean square algorithm. Φ can be any function used
in neural networks. Here we use the Gaussian radial basis function.
3.2.3 The Control Law
To simplify the synthesis of the control law we use the equivalent U-model for the
RBF of equation (3.1):
0 1( ) ( ) ( ) ( 1)My t t t u tα α= + −
(3.2)
where
1 20 ( ) ( ( 1)) ( ( 2)) ........ ( ( ))nt b u t b u t b u t nα
∧ ∧ ∧
= Φ − + Φ − + + Φ −
( ) 11 at =α
Using the U-model of equation (3.2) which is linear with respect to the control term
)1( −tu , the controller has the simplified form as follows:
)(
)()(
)1(
1
0
t
ttU
tu
α
α−
=−
(3.3)
This controller is clearly an inverse of the identified plant.
3.2.4. Simulation Results
We carried out simulations on the following nonlinear Hammerstein model
)(2.0)()(1)(
)2(1.0)1()1(5.0)(
32
tutututx
txtxtyty
+−+=
−+−+−=
Dept of E&C, SJCE Mysore Page 21
Neural Network Application in System Design
The system was modeled according to equation 3.1, and then its equivalent U-model
(equation 3.2) was used to synthesize the control law (equation 3.3). The first parameter
1a was selected as 5, while the number of linear combination weights was four (
1 2 3 4, , &b b b b
∧ ∧ ∧ ∧
). All weights were initialized to 0 and the step size was chosen to be 0.1
The results are depicted in figures 12,13 & 14.
0 10 20 30 40 50 60 70 80 90 100
-2
-1.5
-1
-0.5
0
0.5
1
1.5
2
Time
Reference
Plant Output
Model Output
Figure 12: Tracking simulation
Dept of E&C, SJCE Mysore Page 22
Neural Network Application in System Design
0 10 20 30 40 50 60 70 80 90 100
-1
0
1
2
3
4
Time
Error
Figure 13: Tracking error
0 10 20 30 40 50 60 70 80 90 100
-1
-0.8
-0.6
-0.4
-0.2
0
0.2
0.4
0.6
0.8
1
Time
ControlSignal
Figure 14: Control Input
Dept of E&C, SJCE Mysore Page 23
Neural Network Application in System Design
CONCLUSION
The performance of aircraft systems is highly dependent on the capabilities of the
navigation and control systems. In order to maximize the performance, the control system
needs to be intelligent. There are numerous opportunities envisioned for transitioning
neural network technology to applications within the aerospace industry, besides the tilt
rotor application described in a previous section. The potential payoffs in aircraft
applications include:
• Reduced flight control system design/development costs.
• Reduced costs associated with the need to develop a large aerodynamic data base.
• Reduction of control related accidents, and maintenance of handling qualities
immediately following failures and damage.
Payoffs in missile and guided munitions applications include several of the above
mentioned factors, and in Addition:
Dept of E&C, SJCE Mysore Page 24
Neural Network Application in System Design
• Accommodation of variants within a class of munitions with a single control
design.
• Improve and more predictable weapon performance and the potential for
eliminating the need for wind-tunnel testing.
References
The following sources were consulted in the making of this report
 Gupta, M. M., Jin, L. and Homma, N., “Static and dynamic Neural Networks”,
IEEE Press, 2003.
 Shafiq M. and Riyaz S.H., “Internal Model Control Structure Using Adaptive
Inverse Control Strategy.”, The 4th Int. Conf. on Control and Automation (ICCA),
pp.59-59, 2003.
 Spooner, J. T. , Maggiore, M. , Ordonez, R. and Passino, K. M., “Stable adaptive
control and estimation for nonlinear systems”, Wiley-Interscience, NY , 2002.
Dept of E&C, SJCE Mysore Page 25
Neural Network Application in System Design
 Zhu, Q. M. and Guo. L. Z., “A pole placement controller for nonlinear dynamic
plants”, J. Systems and Control Engineering, Vol. 216 (part I), pp. 467 – 476 ,
2002.
Dept of E&C, SJCE Mysore Page 26

Más contenido relacionado

La actualidad más candente

Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural NetworkPrakash K
 
Computer Networking: A Top-Down Approach
Computer Networking: A Top-Down Approach Computer Networking: A Top-Down Approach
Computer Networking: A Top-Down Approach PolRobinson
 
interferences entre-symboles
interferences entre-symbolesinterferences entre-symboles
interferences entre-symbolesBAKKOURY Jamila
 
Handwritten Digit Recognition using Convolutional Neural Networks
Handwritten Digit Recognition using Convolutional Neural  NetworksHandwritten Digit Recognition using Convolutional Neural  Networks
Handwritten Digit Recognition using Convolutional Neural NetworksIRJET Journal
 
Deep Learning - RNN and CNN
Deep Learning - RNN and CNNDeep Learning - RNN and CNN
Deep Learning - RNN and CNNPradnya Saval
 
3 g signaling_trace_and_analysis
3 g signaling_trace_and_analysis3 g signaling_trace_and_analysis
3 g signaling_trace_and_analysisamakRF
 
Back propagation
Back propagationBack propagation
Back propagationNagarajan
 
CNN and its applications by ketaki
CNN and its applications by ketakiCNN and its applications by ketaki
CNN and its applications by ketakiKetaki Patwari
 
Huawei - Making the World Smaller with Small Cell
Huawei  - Making the World Smaller with Small CellHuawei  - Making the World Smaller with Small Cell
Huawei - Making the World Smaller with Small CellSmall Cell Forum
 
Soft Computing
Soft ComputingSoft Computing
Soft ComputingMANISH T I
 
Artificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKSArtificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKSREHMAT ULLAH
 
Gsm frequency-planning-issue2
Gsm frequency-planning-issue2Gsm frequency-planning-issue2
Gsm frequency-planning-issue2swatisabnis87
 
Introduction to soft computing
Introduction to soft computingIntroduction to soft computing
Introduction to soft computingAnkush Kumar
 
Artificial Intelligence
Artificial Intelligence Artificial Intelligence
Artificial Intelligence Muhammad Ahad
 
FINAL REPORT ON ENTERPRISE NETWORK
FINAL REPORT ON ENTERPRISE NETWORKFINAL REPORT ON ENTERPRISE NETWORK
FINAL REPORT ON ENTERPRISE NETWORKKulsumKhan13
 
Counter propagation Network
Counter propagation NetworkCounter propagation Network
Counter propagation NetworkAkshay Dhole
 

La actualidad más candente (20)

Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Computer Networking: A Top-Down Approach
Computer Networking: A Top-Down Approach Computer Networking: A Top-Down Approach
Computer Networking: A Top-Down Approach
 
interferences entre-symboles
interferences entre-symbolesinterferences entre-symboles
interferences entre-symboles
 
Handwritten Digit Recognition using Convolutional Neural Networks
Handwritten Digit Recognition using Convolutional Neural  NetworksHandwritten Digit Recognition using Convolutional Neural  Networks
Handwritten Digit Recognition using Convolutional Neural Networks
 
Deep Learning - RNN and CNN
Deep Learning - RNN and CNNDeep Learning - RNN and CNN
Deep Learning - RNN and CNN
 
Fuzzy logic
Fuzzy logicFuzzy logic
Fuzzy logic
 
HIGH SPEED NETWORKS
HIGH SPEED NETWORKSHIGH SPEED NETWORKS
HIGH SPEED NETWORKS
 
3 g signaling_trace_and_analysis
3 g signaling_trace_and_analysis3 g signaling_trace_and_analysis
3 g signaling_trace_and_analysis
 
Back propagation
Back propagationBack propagation
Back propagation
 
CNN and its applications by ketaki
CNN and its applications by ketakiCNN and its applications by ketaki
CNN and its applications by ketaki
 
Soft computing
Soft computingSoft computing
Soft computing
 
MobileNet V3
MobileNet V3MobileNet V3
MobileNet V3
 
Huawei - Making the World Smaller with Small Cell
Huawei  - Making the World Smaller with Small CellHuawei  - Making the World Smaller with Small Cell
Huawei - Making the World Smaller with Small Cell
 
Soft Computing
Soft ComputingSoft Computing
Soft Computing
 
Artificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKSArtificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKS
 
Gsm frequency-planning-issue2
Gsm frequency-planning-issue2Gsm frequency-planning-issue2
Gsm frequency-planning-issue2
 
Introduction to soft computing
Introduction to soft computingIntroduction to soft computing
Introduction to soft computing
 
Artificial Intelligence
Artificial Intelligence Artificial Intelligence
Artificial Intelligence
 
FINAL REPORT ON ENTERPRISE NETWORK
FINAL REPORT ON ENTERPRISE NETWORKFINAL REPORT ON ENTERPRISE NETWORK
FINAL REPORT ON ENTERPRISE NETWORK
 
Counter propagation Network
Counter propagation NetworkCounter propagation Network
Counter propagation Network
 

Destacado

MTech - AI_NeuralNetworks_Assignment
MTech - AI_NeuralNetworks_AssignmentMTech - AI_NeuralNetworks_Assignment
MTech - AI_NeuralNetworks_AssignmentVijayananda Mohire
 
Artificial Neural Networks
Artificial Neural NetworksArtificial Neural Networks
Artificial Neural Networksguestac67362
 
Paper Reading : Learning to compose neural networks for question answering
Paper Reading : Learning to compose neural networks for question answeringPaper Reading : Learning to compose neural networks for question answering
Paper Reading : Learning to compose neural networks for question answeringSean Park
 
NLP_Project_Paper_up276_vec241
NLP_Project_Paper_up276_vec241NLP_Project_Paper_up276_vec241
NLP_Project_Paper_up276_vec241Urjit Patel
 
Basic Electrical Engineering
Basic Electrical EngineeringBasic Electrical Engineering
Basic Electrical EngineeringMathankumar S
 
Neural network for machine learning
Neural network for machine learningNeural network for machine learning
Neural network for machine learningUjjawal
 
RAIN WATER HARVESTING
RAIN WATER HARVESTING RAIN WATER HARVESTING
RAIN WATER HARVESTING Mathankumar S
 
Digital image processing - Image Enhancement (MATERIAL)
Digital image processing  - Image Enhancement (MATERIAL)Digital image processing  - Image Enhancement (MATERIAL)
Digital image processing - Image Enhancement (MATERIAL)Mathankumar S
 
Calculating the hamming code
Calculating the hamming codeCalculating the hamming code
Calculating the hamming codeUmesh Gupta
 
Recurrent Neural Network tutorial (2nd)
Recurrent Neural Network tutorial (2nd) Recurrent Neural Network tutorial (2nd)
Recurrent Neural Network tutorial (2nd) 신동 강
 
FISH SEED PRODUCTION & CULTIVABLE FISH SPECIES WITH FISH CUM DUCK FORMING
FISH SEED PRODUCTION & CULTIVABLE FISH SPECIES WITH FISH CUM DUCK FORMINGFISH SEED PRODUCTION & CULTIVABLE FISH SPECIES WITH FISH CUM DUCK FORMING
FISH SEED PRODUCTION & CULTIVABLE FISH SPECIES WITH FISH CUM DUCK FORMINGMathankumar S
 
Backpropagation
BackpropagationBackpropagation
Backpropagationariffast
 
Biological control systems - Time Response Analysis - S.Mathankumar-VMKVEC
Biological control systems - Time Response Analysis - S.Mathankumar-VMKVECBiological control systems - Time Response Analysis - S.Mathankumar-VMKVEC
Biological control systems - Time Response Analysis - S.Mathankumar-VMKVECMathankumar S
 
Back propagation network
Back propagation networkBack propagation network
Back propagation networkHIRA Zaidi
 
Convolution Neural Networks
Convolution Neural NetworksConvolution Neural Networks
Convolution Neural NetworksAhmedMahany
 

Destacado (20)

Pattern recognition
Pattern recognitionPattern recognition
Pattern recognition
 
MTech - AI_NeuralNetworks_Assignment
MTech - AI_NeuralNetworks_AssignmentMTech - AI_NeuralNetworks_Assignment
MTech - AI_NeuralNetworks_Assignment
 
Artificial Neural Networks
Artificial Neural NetworksArtificial Neural Networks
Artificial Neural Networks
 
Hamming
HammingHamming
Hamming
 
Max net
Max netMax net
Max net
 
Paper Reading : Learning to compose neural networks for question answering
Paper Reading : Learning to compose neural networks for question answeringPaper Reading : Learning to compose neural networks for question answering
Paper Reading : Learning to compose neural networks for question answering
 
NLP_Project_Paper_up276_vec241
NLP_Project_Paper_up276_vec241NLP_Project_Paper_up276_vec241
NLP_Project_Paper_up276_vec241
 
Basic Electrical Engineering
Basic Electrical EngineeringBasic Electrical Engineering
Basic Electrical Engineering
 
Neural network for machine learning
Neural network for machine learningNeural network for machine learning
Neural network for machine learning
 
RAIN WATER HARVESTING
RAIN WATER HARVESTING RAIN WATER HARVESTING
RAIN WATER HARVESTING
 
Digital image processing - Image Enhancement (MATERIAL)
Digital image processing  - Image Enhancement (MATERIAL)Digital image processing  - Image Enhancement (MATERIAL)
Digital image processing - Image Enhancement (MATERIAL)
 
Calculating the hamming code
Calculating the hamming codeCalculating the hamming code
Calculating the hamming code
 
Recurrent Neural Network tutorial (2nd)
Recurrent Neural Network tutorial (2nd) Recurrent Neural Network tutorial (2nd)
Recurrent Neural Network tutorial (2nd)
 
FISH SEED PRODUCTION & CULTIVABLE FISH SPECIES WITH FISH CUM DUCK FORMING
FISH SEED PRODUCTION & CULTIVABLE FISH SPECIES WITH FISH CUM DUCK FORMINGFISH SEED PRODUCTION & CULTIVABLE FISH SPECIES WITH FISH CUM DUCK FORMING
FISH SEED PRODUCTION & CULTIVABLE FISH SPECIES WITH FISH CUM DUCK FORMING
 
Backpropagation
BackpropagationBackpropagation
Backpropagation
 
Neural Networks: Introducton
Neural Networks: IntroductonNeural Networks: Introducton
Neural Networks: Introducton
 
Biological control systems - Time Response Analysis - S.Mathankumar-VMKVEC
Biological control systems - Time Response Analysis - S.Mathankumar-VMKVECBiological control systems - Time Response Analysis - S.Mathankumar-VMKVEC
Biological control systems - Time Response Analysis - S.Mathankumar-VMKVEC
 
Back propagation network
Back propagation networkBack propagation network
Back propagation network
 
FISH FARMING
FISH FARMING FISH FARMING
FISH FARMING
 
Convolution Neural Networks
Convolution Neural NetworksConvolution Neural Networks
Convolution Neural Networks
 

Similar a Neural Network Application in System Design Title

Fuzzy Logic Final Report
Fuzzy Logic Final ReportFuzzy Logic Final Report
Fuzzy Logic Final ReportShikhar Agarwal
 
EXPERT SYSTEMS AND ARTIFICIAL INTELLIGENCE_ Neural Networks.pptx
EXPERT SYSTEMS AND ARTIFICIAL INTELLIGENCE_ Neural Networks.pptxEXPERT SYSTEMS AND ARTIFICIAL INTELLIGENCE_ Neural Networks.pptx
EXPERT SYSTEMS AND ARTIFICIAL INTELLIGENCE_ Neural Networks.pptxJavier Daza
 
Nature Inspired Reasoning Applied in Semantic Web
Nature Inspired Reasoning Applied in Semantic WebNature Inspired Reasoning Applied in Semantic Web
Nature Inspired Reasoning Applied in Semantic Webguestecf0af
 
Neural networks are parallel computing devices.docx.pdf
Neural networks are parallel computing devices.docx.pdfNeural networks are parallel computing devices.docx.pdf
Neural networks are parallel computing devices.docx.pdfneelamsanjeevkumar
 
Artificial Neural Networks: Applications In Management
Artificial Neural Networks: Applications In ManagementArtificial Neural Networks: Applications In Management
Artificial Neural Networks: Applications In ManagementIOSR Journals
 
Artificial Neural Network and its Applications
Artificial Neural Network and its ApplicationsArtificial Neural Network and its Applications
Artificial Neural Network and its Applicationsshritosh kumar
 
Artificial neural network for machine learning
Artificial neural network for machine learningArtificial neural network for machine learning
Artificial neural network for machine learninggrinu
 
Artificial Neural Networks ppt.pptx for final sem cse
Artificial Neural Networks  ppt.pptx for final sem cseArtificial Neural Networks  ppt.pptx for final sem cse
Artificial Neural Networks ppt.pptx for final sem cseNaveenBhajantri1
 
Artificial Neural Network in Medical Diagnosis
Artificial Neural Network in Medical DiagnosisArtificial Neural Network in Medical Diagnosis
Artificial Neural Network in Medical DiagnosisAdityendra Kumar Singh
 
Acem neuralnetworks
Acem neuralnetworksAcem neuralnetworks
Acem neuralnetworksAastha Kohli
 
Neuralnetwork 101222074552-phpapp02
Neuralnetwork 101222074552-phpapp02Neuralnetwork 101222074552-phpapp02
Neuralnetwork 101222074552-phpapp02Deepu Gupta
 

Similar a Neural Network Application in System Design Title (20)

Lesson 37
Lesson 37Lesson 37
Lesson 37
 
AI Lesson 37
AI Lesson 37AI Lesson 37
AI Lesson 37
 
Fuzzy Logic Final Report
Fuzzy Logic Final ReportFuzzy Logic Final Report
Fuzzy Logic Final Report
 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
Project Report -Vaibhav
Project Report -VaibhavProject Report -Vaibhav
Project Report -Vaibhav
 
EXPERT SYSTEMS AND ARTIFICIAL INTELLIGENCE_ Neural Networks.pptx
EXPERT SYSTEMS AND ARTIFICIAL INTELLIGENCE_ Neural Networks.pptxEXPERT SYSTEMS AND ARTIFICIAL INTELLIGENCE_ Neural Networks.pptx
EXPERT SYSTEMS AND ARTIFICIAL INTELLIGENCE_ Neural Networks.pptx
 
Nature Inspired Reasoning Applied in Semantic Web
Nature Inspired Reasoning Applied in Semantic WebNature Inspired Reasoning Applied in Semantic Web
Nature Inspired Reasoning Applied in Semantic Web
 
Neural networks are parallel computing devices.docx.pdf
Neural networks are parallel computing devices.docx.pdfNeural networks are parallel computing devices.docx.pdf
Neural networks are parallel computing devices.docx.pdf
 
Jack
JackJack
Jack
 
Artificial Neural Networks: Applications In Management
Artificial Neural Networks: Applications In ManagementArtificial Neural Networks: Applications In Management
Artificial Neural Networks: Applications In Management
 
Artificial Neural Network and its Applications
Artificial Neural Network and its ApplicationsArtificial Neural Network and its Applications
Artificial Neural Network and its Applications
 
Artificial neural network for machine learning
Artificial neural network for machine learningArtificial neural network for machine learning
Artificial neural network for machine learning
 
Artificial Neural Networks ppt.pptx for final sem cse
Artificial Neural Networks  ppt.pptx for final sem cseArtificial Neural Networks  ppt.pptx for final sem cse
Artificial Neural Networks ppt.pptx for final sem cse
 
Deep Learning Survey
Deep Learning SurveyDeep Learning Survey
Deep Learning Survey
 
A Study On Deep Learning
A Study On Deep LearningA Study On Deep Learning
A Study On Deep Learning
 
ANN - UNIT 1.pptx
ANN - UNIT 1.pptxANN - UNIT 1.pptx
ANN - UNIT 1.pptx
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
Artificial Neural Network in Medical Diagnosis
Artificial Neural Network in Medical DiagnosisArtificial Neural Network in Medical Diagnosis
Artificial Neural Network in Medical Diagnosis
 
Acem neuralnetworks
Acem neuralnetworksAcem neuralnetworks
Acem neuralnetworks
 
Neuralnetwork 101222074552-phpapp02
Neuralnetwork 101222074552-phpapp02Neuralnetwork 101222074552-phpapp02
Neuralnetwork 101222074552-phpapp02
 

Último

So einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfSo einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfpanagenda
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxLoriGlavin3
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfMounikaPolabathina
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxLoriGlavin3
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfLoriGlavin3
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxLoriGlavin3
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024Lonnie McRorey
 
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...AliaaTarek5
 
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...panagenda
 
Scale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL RouterScale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL RouterMydbops
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPathCommunity
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentPim van der Noll
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationKnoldus Inc.
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.Curtis Poe
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Mark Goldstein
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesThousandEyes
 
Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Hiroshi SHIBATA
 
Genislab builds better products and faster go-to-market with Lean project man...
Genislab builds better products and faster go-to-market with Lean project man...Genislab builds better products and faster go-to-market with Lean project man...
Genislab builds better products and faster go-to-market with Lean project man...Farhan Tariq
 

Último (20)

So einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdfSo einfach geht modernes Roaming fuer Notes und Nomad.pdf
So einfach geht modernes Roaming fuer Notes und Nomad.pdf
 
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptxDigital Identity is Under Attack: FIDO Paris Seminar.pptx
Digital Identity is Under Attack: FIDO Paris Seminar.pptx
 
What is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdfWhat is DBT - The Ultimate Data Build Tool.pdf
What is DBT - The Ultimate Data Build Tool.pdf
 
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptxA Deep Dive on Passkeys: FIDO Paris Seminar.pptx
A Deep Dive on Passkeys: FIDO Paris Seminar.pptx
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdf
 
The State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptxThe State of Passkeys with FIDO Alliance.pptx
The State of Passkeys with FIDO Alliance.pptx
 
TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024TeamStation AI System Report LATAM IT Salaries 2024
TeamStation AI System Report LATAM IT Salaries 2024
 
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
(How to Program) Paul Deitel, Harvey Deitel-Java How to Program, Early Object...
 
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
Why device, WIFI, and ISP insights are crucial to supporting remote Microsoft...
 
Scale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL RouterScale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL Router
 
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data PrivacyTrustArc Webinar - How to Build Consumer Trust Through Data Privacy
TrustArc Webinar - How to Build Consumer Trust Through Data Privacy
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to Hero
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog Presentation
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.How AI, OpenAI, and ChatGPT impact business and software.
How AI, OpenAI, and ChatGPT impact business and software.
 
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
Arizona Broadband Policy Past, Present, and Future Presentation 3/25/24
 
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyesAssure Ecommerce and Retail Operations Uptime with ThousandEyes
Assure Ecommerce and Retail Operations Uptime with ThousandEyes
 
Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024Long journey of Ruby standard library at RubyConf AU 2024
Long journey of Ruby standard library at RubyConf AU 2024
 
Genislab builds better products and faster go-to-market with Lean project man...
Genislab builds better products and faster go-to-market with Lean project man...Genislab builds better products and faster go-to-market with Lean project man...
Genislab builds better products and faster go-to-market with Lean project man...
 

Neural Network Application in System Design Title

  • 1. Neural Network Application in System Design CHAPTER 1 Introduction to Neural Networks Work on artificial neural networks commonly referred to as "neural networks" (NN) has been motivated right from its origin by the recognition that the human brain computes in an entirely different way than the conventional computer. The brain is a highly complex, nonlinear and parallel computer (information processing system). It has the capability to organize its structural constituents, known as neurons, so as to perform certain computations (e.g. pattern recognition, perception, and motor control) many times faster than the fastest digital computer in existence today. Consider for example, human vision, which is an information-processing task. It is the function of the visual system to provide a representation of the environment around us and, more important, to supply the information we need to interact with the environment. To be specific, the brain routinely accomplish perceptual recognition task (e.g. recognizing a familiar face embedded in an un-familiar scene) in approximately 100-200 ms, where as tasks of much lesser complexity may take days on a conventional computer. How, then, does a human brain do it? At birth, a brain has great structure and the ability to built-up its own rules through what we usually refer to as "experience". Indeed, experience is built up over time, with the most dramatic development (i.e. hard wiring) of the human brain taking place during the first two years from birth: but the development continues well beyond that stage. A "developing" neuron is synonymous with a plastic brain: Plasticity permits the developing nervous system to adapt to its surrounding environment. Just as plasticity appears to be essential to the functioning of neurons as information-processing units in the human brain, so it is with neural networks made up of artificial neurons. In its most general form, a neural network is a machine that is designed to model the way in which the brain performs a particular task or function of interest; the network is usually Dept of E&C, SJCE Mysore Page 1
  • 2. Neural Network Application in System Design implemented by electronic components or is simulated in software on a digital computer. The interest is confined to an important class of neural networks that perform useful computations through a process of learning. To achieve good performance, neural networks employ a massive interconnection of simple computing definition of a neural network viewed as an adaptive machine. A neural network is a massively equivalent distributed process or made up of simple processing units, which has a natural propensity for storing experiential knowledge and making it available for use. It resembles the brain in two respects:  Knowledge is acquired by the network from its environment through a learning process.  Inter neuron connection strengths, known as synaptic weights, are used to store the acquired knowledge. 1.1Neuron & Artificial Neuron The two figures 1 & 2 compare the human neuron and the artificial neuron. For the human neuron the main functioning parts are: Figure 1: Human Neuron Dept of E&C, SJCE Mysore Page 2
  • 3. Neural Network Application in System Design  Dendrites: These act as the input points to the main body of the neuron.  Synapse: This is the storage area of the past experience.  Soma: It receives synaptic information and performs further processing on the information.  Axon: This is the output line for the neuron. Figure 2: Artificial Neuron In artificial neural networks, the synaptic and somatic operations are emulated as follows:  Synaptic Operation: The input weights act as storage for knowledge (and therefore, as memory for previous experiences).  Somatic Operation: The somatic operation is provided by various mathematical operations such as aggregation, thresholding, nonlinear activation and dynamic processing to the synaptic inputs. Dept of E&C, SJCE Mysore Page 3
  • 4. Neural Network Application in System Design 1.2Adaptation in NN’s The procedure that is used to perform the learning process is called a learning algorithm (fig. 3), the function of which is to modify the synaptic weights of the network in an orderly fashion to attain a desired design objective. Figure 3: Adaptation in NN’s The modification of synaptic weights provides the traditional method for the design of neural networks. Such an approach is the closest to linear adaptive filter theory, which is already well established and successfully applied in many diverse fields. However, it is also possible for a neural network to modify its own topology, which is motivated by the fact that neurons in the human brain can die and then new synaptic connections can grow. 1.3 Common Neural Network Architectures The manner in which the neurons of a neural network are structured is intimately linked with the learning algorithm used to train the network. We may therefore speak of algorithms (rules) used in the design of neural networks as being structured. In general we may identify three fundamentally different classes of network architectures: Dept of E&C, SJCE Mysore Page 4
  • 5. Neural Network Application in System Design 1.3.1 Single-Layer Feed-forward Networks In a layered neural network the neurons are organized in the form, of layers. In the simplest form of layered network, we have an input layer of source nodes that projects onto an output layer of neurons (computation nodes), but not vise versa. In other words, this network is strictly a feed-forward or acyclic type. It is illustrated in the figure for the case if four nodes in both the input and output layers. Such a network is called a single-layered network, with the name "single-layer" referring to the output layer of computation nodes (neurons). We do not count the input layer of source nodes because no computation is performed there. Figure 4: Single Layer Feedforward NN 1.3.2. Multilayer Feed-Forward Networks The second class of a feed-forward neural network distinguishes its self by the presence of one or more hidden layers, whose computation nodes are correspondingly called hidden neurons or hidden units. The function of the hidden neuron is to interfere between the external input and the network output in some useful manner. Multilayer feed forward networks are an important class of neural networks. Typically, the network consists of a set of sensory units (source nodes) that constitute the input layer, one or more hidden layers of computation nodes, and an out-put layer of computation nodes. The input signal propagates through the network in a forward Dept of E&C, SJCE Mysore Page 5
  • 6. Neural Network Application in System Design direction, on a layer-by-layer basis. These neural networks are commonly referred to as multilayer perceptions (MLP’s), which represent a generalization of the single-layer perception. The source nodes in the input layer of the network supply respective elements of the activation pattern (input vector), which constitute the input signals applied to the neurons (computation nodes) in the second layer (i.e., the first hidden layer). The output signals of the second layer are used as an input to the third layer, and so on for the rest of the network. Typically the neurons at each layer of the network have as there inputs the outputs of the preceding layers only. The set of output signals of the neurons in the output (final layer) constitutes the overall response of the network to the activation pattern supplied by the source nodes in the input (first) layer. Multilayer perceptrons have been applied successfully to solve some difficult and diverse problems by training them in a supervised manner with a highly popular algo- rithm known as the error back-propagation algorithm. This algorithm is based on the error-correction learning rule. As such, it may be viewed as a generalization of an equally popular adaptive filtering algorithm: the least-mean-square (LMS) algorithm. Basically, error back-propagation learning consists of two passes through the dif- ferent layers of the network: a forward pass and a backward pass. In the forward pass, an activity pattern (input vector) is applied to the sensory nodes of the network, and its effect propagates through the network layer by layer. Finally, a set of outputs is produced as the actual response of the network. During the forward pass the synaptic weights of the networks are all fixed. During the backward pass, on the other hand, the synaptic weights are all adjusted in accordance with an error-correction rule. Specifically, the actual response of the network is subtracted from a desired (target) response to produce an error signal. This error signal is then propagated backward through the network, against the direction of synaptic connections, hence the name "error back-propagation." The synaptic weights are adjusted to make the actual response of the network move closer to the desired response in a statistical sense. The error back-propagation algorithm is also referred to in the literature as the back-propagation algorithm. Dept of E&C, SJCE Mysore Page 6
  • 7. Neural Network Application in System Design Figure 5 shows the architectural graph of a multilayer perceptron with two hidden layers and an output layer. Signal flow through the network progresses in a forward direction, from left to right and on a layer-by-layer basis. Figure 5: Multi-layer Feedforward NN’s The neural network in the figure is said to be fully connected in the sense that every node in each layer of the network is connected to every other node in the adjacent forward layer. If, however, some of the communication links are missing from the network, we say that the network is partially connected. 1.3.3. Recurrent Networks A recurrent neural network distinguishes itself from the feed-forward network in that it has at least one feedback loop. For example, a recurrent network may consist of a single layer of neurons with each neuron feeding its output signal back to the input of all input neurons. The presence of feedback loops has a profound impact on the learning capability of the network and on its performance. Moreover, the feedback loops involve the use of Dept of E&C, SJCE Mysore Page 7
  • 8. Neural Network Application in System Design particular branches composed of unit-delay elements which result in a nonlinear dynamical behavior, assuming that the neural network contains nonlinear units. 1.4 Applications of Neural Networks Neural networks are applicable in virtually every situation in which a relationship between the predictor variables (independents, inputs) and predicted variables (dependents, outputs) exists, even when that relationship is very complex and not easy to articulate in the usual terms of "correlations" or "differences between groups." A few representative examples of problems to which neural network analysis has been applied successfully are:  Detection of medical phenomena: A variety of health-related indices (e.g., a combination of heart rate, levels of various substances in the blood, respiration rate) can be monitored. The onset of a particular medical condition could be associated with a very complex (e.g., nonlinear and interactive) combination of changes on a subset of the variables being monitored. Neural networks have been used to recognize this predictive pattern so that the appropriate treatment can be prescribed.  Stock market prediction: Fluctuations of stock prices and stock indices are another example of a complex, multidimensional, but in some circumstances at least partially-deterministic phenomenon. Neural networks are being used by many technical analysts to make predictions about stock prices based upon a large number of factors such as past performance of other stocks and various economic indicators.  Credit assignment: A variety of pieces of information are usually known about an applicant for a loan. For instance, the applicant's age, education, occupation, and many other facts may be available. After training a neural network on historical data, neural network analysis can identify the most relevant characteristics and use those to classify applicants as good or bad credit risks. Dept of E&C, SJCE Mysore Page 8
  • 9. Neural Network Application in System Design  Condition Monitoring: Neural networks can be instrumental in cutting costs by bringing additional expertise to scheduling the preventive maintenance of machines. A neural network can be trained to distinguish between the sounds a machine makes when it is running normally ("false alarms") versus when it is on the verge of a problem. After this training period, the expertise of the network can be used to warn a technician of an upcoming breakdown, before it occurs and causes costly unforeseen "downtime."  Engine management: Neural networks have been used to analyze the input of sensors from an engine. The neural network controls the various parameters within which the engine functions, in order to achieve a particular goal, such as minimizing fuel consumption.  Signature analysis: as a mechanism for comparing signatures made (e.g. in a bank) with those stored. This is one of the first large-scale applications of neural networks in the USA, and is also one of the first to use a neural network chip.  Process control: most processes cannot be determined as computable algorithms. Neural Networks can be used to adaptively control the process  Nonlinear Identification & Adaptive Control: This is one of the main areas of application of the neural networks. Neural Networks find applications in situations where the plant dynamics are uncertain or un-modeled.  Application of Neural Network on Flight Control: Dept of E&C, SJCE Mysore Page 9
  • 10. Neural Network Application in System Design Fig. 1. Application of neural network in flight control Fig. 2. Intelligent flight control system generation II system Dept of E&C, SJCE Mysore Page 10
  • 11. Neural Network Application in System Design CHAPTER 2 LMS and RBF-NN’s 2.1 The Least Mean Square (LMS) Adaptation Algorithm As discussed in section 1, the learning process in the neurons involves updating of certain “weights”. A number of adaptation algorithms are available in literature. The criteria/cost functions used for adaptation and the methods are usually derived from the richly developed field of adaptive filter theory. In the following we present one of the most commonly used adaptation algorithms, the LMS. Let: ][nw : time varying neuron tap weights ][nv : input to neuron ][nd : desired response [ ]ns~ : actual output of neuron J : cost function (the mean square error) The estimation error is the difference between the desired response and the estimated output: ][~][][ nsndne −= using ][][~ nvwns H = gives, ][][][ nvwndne H −= The mean square error (cost function) is defined as: Dept of E&C, SJCE Mysore Page 11
  • 12. Neural Network Application in System Design { } { }22 ][][][ nvwndEneEJ H −== Thus the cost function J is a function of vector w. Minimizing J with respect to the complex tap weights w leads to the set of equations called the Wiener-Hopf equations. If we limit the number of taps to M then we obtain the matrix formulation of the Wiener- Hopf equations from which the solution is obtained as pRwo 1− = where { }][][ nvnvER H = and { }][][ * ndnvEp = This method of solution requires inversion of a matrix. An alternative adaptive method of solution is the Steepest Descent algorithm. It can be shown that the cost function J has the shape of an M-dimensional bowl whose minimum is at the optimal solution of w. The steepest descent algorithm moves the tap weights towards the minimum of the J bowl at every iteration by moving them in the direction opposite to the gradient vector: pRwJw 22 −=∇ Thus we have an iterative definition for the tap weight updates: Jnwnw w∇−=+ 2 ][]1[ µ ]][[][ nRwpnw −+= µ µ is known as the adaptation step. (µ > 0) In practice, although we do not know R and p, we can use their instantaneous estimates: ][][][ˆ nvnvnR H = ][][ˆ * ndnvp = Dept of E&C, SJCE Mysore Page 12
  • 13. Neural Network Application in System Design This approach is known as the Least Mean Squares (LMS) method which is a member of a particular class of algorithms called the stochastic gradient algorithms. Thus the adaptation equation is given as: ][][][ˆ]1[ˆ * nenvnwnw µ+=+ As we shall see, this equation is used in a variety of adaptation algorithms each having its own definition for the error functions. 2.2. RBF Neural Networks Among the vast variety of neural networks, the RBF-NN is a quire commonly used structure. The design of a RBF-NN in its most basic form consists of three separate layers. The input layer is the set of source nodes (sensory units). The second layer is a hidden layer of high dimension. The output layer gives the response of the network to the activation patterns applied to the input layer. The transformation from the input space to the hidden-unit space is nonlinear. On the other hand, the transformation from the hidden space to the output space is linear. Figure 6: RBF-NN Basic Structure With reference to the figure above, the output y(t) is a weighted sum of the outputs of the hidden layer, given by Dept of E&C, SJCE Mysore Page 13
  • 14. Neural Network Application in System Design 1 ˆ( ) ( ( ) ), n i i i y t w u t cφ = = −∑ (3.3) where ( )u t is the input ( ).φ is an arbitrary nonlinear radial basis function . denotes the norm that is usually assumed to be Euclidean ic are the known centers of the radial basis functions iw are the weights Radial functions are a special class of functions. Their characteristic feature is that their response decreases (or increases) monotonically with distance from a central point and they are radially symmetric. The centre, the distance scale, and the precise shape of the radial function are parameters of the model. There are a variety of radial functions available in literature. The most commonly used one is the Gaussian radial filter, which in case of a scalar input is 2 2 ( ) ( ) exp( ) x c h x β − = − Its parameters are its centre c and its radius β (width), Figure 1.3 illustrates a Gaussian RBF with centre c = 0 and radius β = 1. A Gaussian RBF monotonically decreases with Dept of E&C, SJCE Mysore Page 14
  • 15. Neural Network Application in System Design distance from the centre Figure 7: Gaussian function Profile A summary of the characteristics of the RBF-NN’s is given below  They are two-layer feed-forward networks.  The hidden nodes implement a set of radial basis functions (e.g. Gaussian functions).  The output nodes implement linear summation functions as in an MLP.  The network training is divided into two stages: first the weights from the input to hidden layer are determined, and then the weights from the hidden to output layer.  The training/learning is very fast.  The networks are very good at interpolation. 2.3 Detailed Learning Algorithm for Rbf-Nn’s The whole algorithm of RBF network learning may be split into two phases:  Hidden layer learning or basis function selection (unsupervised learning).  Fitting of outputs in a transformed feature space (supervised learning). 2.3.1 Unsupervised learning The first phase is an unsupervised learning. It does not use any information on target outputs and deals only with a set of inputs. At this stage we have to: Dept of E&C, SJCE Mysore Page 15
  • 16. Neural Network Application in System Design 1. Select a number of radial basis functions. 2. Select a center for each basis function and 3. Select a value for the parameter β (width), which characterizes the basis function range of definition (the range of its influence). A too large value of β forms too narrow basis functions. In step1, usually all the RBF’s are chosen to be the same. There are many algorithms available for the selection of the centers (step-2), one of the more popular ones is the K- means clustering algorithm which goes as follows Given m data points, select l as the number of clusters such that l < m Take the first l learning data as the center vectors for the l clusters , 1,2,...., ;j jc x j l= = Assign the remaining data points to one of the clusters with the least distance criterion. Recomputed the center vectors using the new mean, that is 1 ; 1 j j j i cj c x j l m ∈ = ≤ ≤∑ where mj is the number of data points belonging to the jth cluster. As soon as the clustering algorithm is complete we may move to the selection of the variance or width parameter β (step-3). These parameters control the amount of overlap of the radial basis functions as well as the network generalizations. A small value yields a rapidly decreasing function, whereas a large value results in a more gently varying function. The mostly commonly used method for the selection of the width parameter for a cluster is to take it equal to the average distance between the data in the cluster and center of the cluster. Dept of E&C, SJCE Mysore Page 16
  • 17. Neural Network Application in System Design 2.3.2 Supervised Learning The second phase is a supervised learning. The goal is to fit outputs with a linear function of nonlinear transformed inputs. Any gradient optimization method may be used, but the LMS (discussed above) is used most often. 2.3.3 Relative Advantages of RBF-NN’s  Many pattern recognition experiments show that the RBF-NN’s are superior over other neural network approaches in the following senses.  RBF-NN’s are capable of approximating nonlinear mappings effectively.  The training time of the RBF-NN’s is quite low compared to that of other neural network approaches such as the multi-layer perceptron.  The RBF-NN’s produce classification accuracies from 5% to 10% higher than accuracies produced by the back propagation algorithm.  The RBF-NN’s are quite successful for identifying regions of sample data not in any known class because they use a non-monotonic transfer function based on the Gaussian density function. Dept of E&C, SJCE Mysore Page 17
  • 18. Neural Network Application in System Design CHAPTER 3 Neural Network Applications 3.1 Nonlinear Plant Identification One of the major areas of application of the neural networks is in the identification of nonlinear plants. The RBF-NN was introduced in section 2. Here we make use of the Gaussian RBF-NN to identify the nonlinear system known as the continuous stirred tank reactor. The nonlinear model of the continuous stirred tank reactor when the sampling time is chosen as 0.05 seconds is as follows. 2 3 4 2 3 2 2 2 2 3 3 ( 1) 0.8606 ( ) 0.0401 ( ) 0.0017 ( ) 0.000125 ( ) 0.0464 ( ) 0.045 ( ) ( ) 0.0034 ( ) ( ) 0.00025 ( ) ( ) 0.0012 ( ) 0.0013 ( ) ( ) 0.0001458 ( ) ( ) 0.00002083 ( ) 0.00002083 ( ) ( ) y t y t y t y t y t u t y t u t y t u t y t u t u t y t u t y t u t u t y t u t + = − + − + − + − − + − + − Figure 8: Identification Structure The RBF-NN assumes no prior knowledge of the system parameters and tries to identify the system online. Simulations were carried out using SIMULINK. The number of linear combiner weights was chosen to be ten. The neural network worked well and was able to identify the nonlinear plant online. The simulation results are given in the figures 9 & 10. Dept of E&C, SJCE Mysore Page 18
  • 19. Neural Network Application in System Design 0 20 40 60 80 100 120 140 160 180 200 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 Time Plant Output NN Output Figure 9: Identification simulation 0 20 40 60 80 100 120 140 160 180 200 -1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1 Time Error Figure 10: Mismatch error Dept of E&C, SJCE Mysore Page 19
  • 20. Neural Network Application in System Design 3.2 Adaptive Tracking of Nonlinear Dynamic Plants The adaptive control of nonlinear dynamic plants is an extremely important area of research. It involves the online identification of the plant and development of a controller based on this identified plant. The identification part is generally carried out using powerful neural networks. A number of techniques have been suggested in literature. Here we utilize one of the more recent approaches. The main structure is depicted below. The basic structure is that of an IMC (Internal Model Control). The identification task is carried out utilizing the Gaussian RBF-NN. A control law is synthesized which is based on the identified system parameters. Figure 11: Adaptive Tracking 3.2.1 The Plant We assume a stable nonlinear dynamic plant whose functional parameters or the functional structure need not be known. 3.2.2 The Identifying Model We identify the plant online using the radial basis function with the following structure: 1 21( ) ( 1) ( ( 1)) ( ( 2)) ......... ( ( ))nMy t a u t b u t b u t b u t n ∧ ∧ ∧ = − + Φ − + Φ − + + Φ − (3.1) Dept of E&C, SJCE Mysore Page 20
  • 21. Neural Network Application in System Design where the parameter 1a is selected in advance and the parameters nbbb ∧∧∧ ,......, 21 are estimated using the normalized least mean square algorithm. Φ can be any function used in neural networks. Here we use the Gaussian radial basis function. 3.2.3 The Control Law To simplify the synthesis of the control law we use the equivalent U-model for the RBF of equation (3.1): 0 1( ) ( ) ( ) ( 1)My t t t u tα α= + − (3.2) where 1 20 ( ) ( ( 1)) ( ( 2)) ........ ( ( ))nt b u t b u t b u t nα ∧ ∧ ∧ = Φ − + Φ − + + Φ − ( ) 11 at =α Using the U-model of equation (3.2) which is linear with respect to the control term )1( −tu , the controller has the simplified form as follows: )( )()( )1( 1 0 t ttU tu α α− =− (3.3) This controller is clearly an inverse of the identified plant. 3.2.4. Simulation Results We carried out simulations on the following nonlinear Hammerstein model )(2.0)()(1)( )2(1.0)1()1(5.0)( 32 tutututx txtxtyty +−+= −+−+−= Dept of E&C, SJCE Mysore Page 21
  • 22. Neural Network Application in System Design The system was modeled according to equation 3.1, and then its equivalent U-model (equation 3.2) was used to synthesize the control law (equation 3.3). The first parameter 1a was selected as 5, while the number of linear combination weights was four ( 1 2 3 4, , &b b b b ∧ ∧ ∧ ∧ ). All weights were initialized to 0 and the step size was chosen to be 0.1 The results are depicted in figures 12,13 & 14. 0 10 20 30 40 50 60 70 80 90 100 -2 -1.5 -1 -0.5 0 0.5 1 1.5 2 Time Reference Plant Output Model Output Figure 12: Tracking simulation Dept of E&C, SJCE Mysore Page 22
  • 23. Neural Network Application in System Design 0 10 20 30 40 50 60 70 80 90 100 -1 0 1 2 3 4 Time Error Figure 13: Tracking error 0 10 20 30 40 50 60 70 80 90 100 -1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1 Time ControlSignal Figure 14: Control Input Dept of E&C, SJCE Mysore Page 23
  • 24. Neural Network Application in System Design CONCLUSION The performance of aircraft systems is highly dependent on the capabilities of the navigation and control systems. In order to maximize the performance, the control system needs to be intelligent. There are numerous opportunities envisioned for transitioning neural network technology to applications within the aerospace industry, besides the tilt rotor application described in a previous section. The potential payoffs in aircraft applications include: • Reduced flight control system design/development costs. • Reduced costs associated with the need to develop a large aerodynamic data base. • Reduction of control related accidents, and maintenance of handling qualities immediately following failures and damage. Payoffs in missile and guided munitions applications include several of the above mentioned factors, and in Addition: Dept of E&C, SJCE Mysore Page 24
  • 25. Neural Network Application in System Design • Accommodation of variants within a class of munitions with a single control design. • Improve and more predictable weapon performance and the potential for eliminating the need for wind-tunnel testing. References The following sources were consulted in the making of this report  Gupta, M. M., Jin, L. and Homma, N., “Static and dynamic Neural Networks”, IEEE Press, 2003.  Shafiq M. and Riyaz S.H., “Internal Model Control Structure Using Adaptive Inverse Control Strategy.”, The 4th Int. Conf. on Control and Automation (ICCA), pp.59-59, 2003.  Spooner, J. T. , Maggiore, M. , Ordonez, R. and Passino, K. M., “Stable adaptive control and estimation for nonlinear systems”, Wiley-Interscience, NY , 2002. Dept of E&C, SJCE Mysore Page 25
  • 26. Neural Network Application in System Design  Zhu, Q. M. and Guo. L. Z., “A pole placement controller for nonlinear dynamic plants”, J. Systems and Control Engineering, Vol. 216 (part I), pp. 467 – 476 , 2002. Dept of E&C, SJCE Mysore Page 26