seminar report ann

31
A Seminar Presentation Report On “ARTIFICIAL NEURAL NETWORK” Submitted to: Submitted by: Er. Rashi Gupta Balveer Singh (CSE ) Final Year (CSE)

Upload: kartik143

Post on 18-Nov-2014

142 views

Category:

Documents


7 download

DESCRIPTION

Report on Artificial Neural Network

TRANSCRIPT

Page 1: Seminar Report ANN

A

Seminar Presentation Report

On

“ARTIFICIAL NEURAL NETWORK”

Submitted to: Submitted by:

Er. Rashi Gupta Balveer Singh

(CSE ) Final Year (CSE)

DEPARTMENT OF COMPUTER SCIENCE & ENGINEERING

MARWAR ENGINEERING COLLEGE AND RESEARCH CENTER JODHPUR (RAJ.)

Page 2: Seminar Report ANN

MARWAR ENGINEERING COLLEGE & RESEARCH CENTRE JODHPUR

CERTIFICATE

This is to certify that the project entitled “ARTIFICIAL NEURAL NETWORK “has been

carried out by BALVEER SINGH under my guidance in partial fulfillment of the degree of

Bachelor of Engineering in Computer Engineering Rajasthan Technical University, Kota during

the academic year 2009-2010. To the best of my knowledge and belief this work has not been

submitted elsewhere for the award of any other degree.

Head of the Department

Page 3: Seminar Report ANN

AFTIFICIAL NEURAL NETWORK

Definition:

An Artificial Neural Network (ANN) is an information-processing paradigm that is inspired by the way biological nervous systems, such as the brain, process information. The key element of this paradigm is the novel structure of the information processing system. It is composed of a large number of highly interconnected processing elements (neurons) working in unison to solve specific problems. ANNs, like people, learn by example. An ANN is configured for a specific application, such as pattern recognition or data classification, through a learning process. Learning in biological systems involves adjustments to the synaptic connections that exist between the neurons.

Introduction:

ANNS are used as a solution to various problems; however, their success as an intelligent pattern recognition methodology has been most prominently advertised. The most important, and attractive, feature of ANNs is their capability of learning (generalizing) from example (extracting knowledge from data). ANNs can do this without any pre-specified rules that define intelligence or represent an expert’s knowledge. This feature makes ANNs a very popular choice for gene expression analysis and sequencing. Due to their power and flexibility, ANNS have even been used as tools for relevant variable selection, which can in turn greatly increase the expert’s knowledge and understanding of the problem.

A neural network is characterized by its pattern of connections between the neurons referred to as network architecture and its method of determining the weights on the connections called training or learning algorithm. The weights are adjusted on the basis of data. In other words, neural networks learn from examples and exhibit some capability for generalization beyond the training data. This feature makes such computational models very appealing in application domains where one has little or incomplete understanding of the problem to be solved, but where training data is readily available. Neural networks normally have great potential for parallelism, since the computations of the components are largely interdependent of each other.

Artificial neural networks are viable computational models for a wide variety of problems. Already, useful applications have been designed, built, and commercialized for various areas in engineering, business and biology. These include pattern classification, speech synthesis and recognition, adaptive interfaces between human and complex physical systems, function approximation, image compression, associative memory, clustering, forecasting and prediction, combinatorial optimization, nonlinear system modeling, and control. Although they may have been inspired by neuroscience, the majority of the networks have close relevance or counterparts to traditional statistical methods such as non-parametric pattern classifiers, clustering algorithm, nonlinear filters, and statistical regression models

Page 4: Seminar Report ANN

NEURAL NETWORKS

Artificial neural networks have emerged from the studies of how brain performs. The human brain consists of many millions of individual processing elements, called neurons that are highly interconnected.

Information from the outputs of the neurons, in the form of electric pulses is received by the cells at connections called synapses. The synapses connect to the cell inputs, or dendrites and the single output pf the neuron appears at the axon. An electric pulse is sent down the axon when the total input stimuli for all of the dendrites exceed a certain threshold.

Artificial neural networks are made up of simplified individual models of the biological neuron that are connected together to form a network. Information is stored in the network in the form of weights or different connections strengths associated with synapses in the artificial neuron models.

Many different types of neural networks are available and multi layer neural networks are the most popular which are extremely successful in pattern reorganization problems. An artificial neuron model is shown below. Each neuron input is weighted by W. changing the weights of an element will alter the behavior of the whole network. The output y is obtained by summing the weighted inputs to the neuron and passing the result through a non-linear activation function, f ().

Page 5: Seminar Report ANN

Multi layer networks consists of an input layer, a hidden layer are made up of no. of nodes. Data flows through the network in one direction only, from input to output; hence this type of network is called a feed-forwarded network. A two-layered network is shown below.

NEURAL NETWORKS IN PROCESS CONTROL

Artificial neural networks are implemented as software packages in computers and being used to incorporate of artificial intelligence in control system. ANN is basically mathematical tools which are being designed to employ principles similar to neurons networks of biological system. ANN is able to emulate the information processing capabilities of biological neural system. ANN has overcome many of the difficulties that t conventional adaptive control systems suffer while dealing with non linear behavior of process.

PROCEDURES FOR ANN SYSTEM ENGINEERING

In realistic application the design of ANN system is complex, usually iterative and interactive task. Although it is impossible to provide an all inclusive algorithmic procedure, the following highly interrelated, skeletal steps reflect typical efforts and concerns. The plethora of possible ANN design parameters include:

Page 6: Seminar Report ANN

• The interconnection strategy/network topology/network structure.

• Unit characteristics (may vary within the network and within subdivisions within the network such as layers).

• Training procedures.

• Training and test sets.

• Input/output representation and pre- and post-processing.

FEATURES OF ANN

• Their ability to represent nonlinear relations makes them well suited for non linear modeling in control systems.

• Adaptation and learning in uncertain system through off line and on line weight adaptation

• Parallel processing architecture allows fast processing for large-scale dynamic system.

• Neural network can handle large number of inputs and can have many outputs.

Neural network architecture have learning algorithm associated with them. The most popular network architecture used for control purpose is multi layered neural network [MLNN] with error propagation [EBP] algorithm.

BENEFITS OF NEURAL NETWORKS

It is apparent that a neural network derives its computational power through, first, its massively parallel distributed structure and, second its ability to learn and therefore generalize. Generalization refers to the neural network producing reasonable outputs for inputs not encountered during training. These two information-processing capabilities make it possible for neural networks to solve complex problems that are currently intractable. A complex problem of interest is decomposed into a number of relatively simple tasks, and neural networks are assigned a subset of tasks that match their inherent capabilities. The use of neural networks offers the following useful properties and capabilities.

Page 7: Seminar Report ANN

• Nonlinearity, An artificial neuron can be linear or nonlinear. A neural network made up of an interconnection of nonlinear neurons, is itself nonlinear. Nonlinearity is a highly important property, particularly if the underlying physical mechanism responsible for generation of the input signal (e.g., speech signal) is inherently nonlinear.

• Input-Output Mapping, A popular paradigm of learning called learning with a teacher or supervised learning involves modification of the synaptic weights of a neural network by applying a set of labeled training samples or task examples. Each example consists of a unique input signal and a corresponding desired response. The network is presented with an example picked at random from the set, and the synaptic weights of the network are modified to minimize the difference between the desired response and the actual response of the network produced by the input signal in accordance with an appropriate statistical criterion. Thus the network learns from the examples by constructing an input-output mapping for the problem at hand.

• Adaptivity, Neural networks have a built-in capability to adapt their synaptic weights to changes in the surrounding environment. In particular, a neural network trained to operate in a specific environment can be easily restrained to deal with minor changes in the operating environmental conditions. Moreover, when it is operating in a non-stationary environment, a neural network can be designed to change its synaptic weights in real time.

• Evidential Response, In the context of pattern classification, a neural network can be designed to provide information not only about which pattern to select, but also about the confidence in the decision made. This latter information may be used to reject ambiguous patterns, should they rise, and thereby improving the classification performance of the network.

• Contextual Information, Knowledge is represented by the very structure and activation state of a neural network. Every neuron in the network is potentially affected by the global activity of all other neurons in the network. Consequently, contextual information is dealt with naturally by a neural network.

• Fault Tolerance, A neural network implemented in hardware form, has the potential to be inherently fault tolerant, or capable of robust computation, in the sense that its performance degrades gracefully under adverse operating conditions. This can be attributed to its massively distributed nature of information stored in the network. Thus, in principle, a neural network exhibits a graceful degradation in performance rather than catastrophic failure.

• VLSI Implement ability, The massively parallel nature of a neural network makes it potentially fast for the computation of certain tasks. This same feature makes a neural network well suited for implementation using very-large-scale-integrated technology.

Page 8: Seminar Report ANN

One particular beneficial virtue of VLSI is that it provides a means of capturing truly complex behavior in a highly hierarchical fashion.

• Neurobiological Analogy, The design of a neural network is motivated by analogy with the brain, which is living proof that fault tolerant parallel processing is not only physically possible but also fast and powerful.

NEURAL NETWORK FOUNDATIONS

Artificial Neural networks (ANNs) belong to the adaptive class of techniques in the

machine learning arena. ANNS are used as a solution to various problems; however, their

success as an intelligent pattern recognition methodology has been most prominently

advertised. Most models of ANNs are organized in the form of a number of processing units

called artificial neurons, or simply neurons, and a number of weighted connections referred to

as artificial synapses between the neurons. The process of building an ANN, similar to its

biological inspiration, involves a learning episode. During learning episode, the network

observes a sequence of recorded data, and adjusts the strength of its synapses according to a

learning algorithm and based on the observed data. The process of adjusting the synaptic

strengths in order to be able to accomplish a certain task, much like the brain, is called

“learning”. Learning algorithms are generally divided into two types, supervised and

unsupervised. The supervised algorithms require labeled training data. In other words, they

require more a priori knowledge about the training set.

Page 9: Seminar Report ANN

MODELS OF A NEURON

A neuron is an information-processing unit that is fundamental to the operation of a

neural network. Fig.2.1a shows the model of a neuron, which forms the basis for designing

artificial neural networks. The three basic elements of the neuronal model are

1. A set of synapses or connecting links, each of which is characterized by a weight or

strength of its own. Specifically, a signal xj at which the input of synapse j connected

to neuron k is multiplied by the synaptic weight wjk.

2. An adder for summing the input signals, weighted by the respective synapses of the

neuron; the operations described here constitutes a linear combiner.

3. An activation function for limiting the amplitude of the output of a neuron. The

activation is also referred to as a squashing function as it squashes the permissible

amplitude range of the output signal to some finite value.

NETWORK ARCHITECTURES

The manner in which the neurons of a neuron network are structured is intimately

linked with the learning algorithm used to train the network. Network structure can be broadly

divided into three classes of network architectures.

Single-Layer Feed forward Networks

In a layered neural network the neurons are organized in the form of layers. The

simplest form of a layered network consists of an input layer of source nodes those projects

onto an output layer of neurons, but not vice-versa. In other words, this network is strictly a

feed forward or acyclic type.

Such a network is called a single-layer network, with the designation “single layer”

referring to the output layer of computational nodes. The input layer of source nodes are not

counted as no computation is performed here.

Page 10: Seminar Report ANN

Multilayer Feed forward Networks

The second class of a feed forward neural network distinguishes itself by the

presence of one or more hidden layers, whose computation nodes are correspondingly called

hidden neurons or hidden units. The function of hidden neurons is to intervene between the

external input and the network output in some useful manner. By adding one or more hidden

layers, the network is enabled to extract higher-order statistics.

In a rather loose sense the network acquires a global perspective despite its local

connections due to the extra set of synaptic connections and extra dimension of neural

interactions. The ability of hidden neurons to extract higher-order statistics is particularly

valuable when the size of the input is large.

Recurrent Networks

A recurrent network distinguishes itself from a forward neural network in that it has

at least one feedback loop. For example, a recurrent network may consist of a single layer of

neurons with each neuron feeding its output signal back to the inputs of all the other neurons,

as illustrated in the architectural graph. The presence of feedback loops has a profound impact

on the learning compatibility of the network and its performance. Moreover, the feedback

loops involve the use of particular branches composed of unit-delay elements (denoted by z -1),

which result in a nonlinear dynamical behavior, assuming that the neural network contains

nonlinear units.

Page 11: Seminar Report ANN

TRAINING OF NEURAL NETWORKS

Neural networks are models that may be used to approximate summaries, classify,

generalize or otherwise represent real situations. Before models can be used they have to be

trained or made to ‘fit’ into the representative data. The model parameters, e.g., number of

layers, number of units in each layer and weights of the connections between them, must be

determined. In ordinary statistical terms this is called regression. There are two fundamental

types of training with neural networks: supervised and unsupervised learning. For supervised

training, as in regression, data used for the training consists of independent variables (also

referred to as feature variables or predictor variables) and dependent variables (target values).

The independent variables (input to the neural network) are used to predict the dependent

variables (output from the network). Unsupervised training does not have dependent (target)

values supplied: the network is supposed to cluster the data automatically into meaningful sets.

The fundamental idea behind training, for all neural networks, is in picking a set of

weights and then applying the inputs to the network and gauging the network’s performance

with this set of weights. If the network does not perform well, then the weights are modified

by an algorithm specific to each architecture and the procedure is then repeated. This iterative

process continues until some pre-specified criterion has been achieved. A training pass

through all vectors of the input data is called an epoch. Iterative changes can be made to the

weights with each input vector, or changes can be made after all input vectors have been

processed. Typically, weights are iteratively modified by epochs.

LEARNING TECHNIQUES

Page 12: Seminar Report ANN

Learning rules are algorithm for slowly alerting the connections weighs to achieve a desirable

goal such a minimization of an error function. The generalized step for any neural network

leaning algorithm is follows are the commonly used learning algorithm for neural networks.

Multi layer neural net (MLNN)

Error back propagation (EBB)

Radial basis functions (RBF

Reinforcement learning

Temporal deference learning

Adaptive resonance theory (ART)

Genetic algorithm

Selection of a particular learning algorithm depends on the network and network

topology. As MLNN with EBP is most extensively used and widely accepted network for

process application, namely for identification and control of the process.

MLNN IN SYSTEM IDENTIFICATION:

There has been an “explosion” of neural network application in the areas

of process control engineering in the last few years. Since it become very difficult to obtain the

model of complex non-linear system due its unknown dynamics and a noise environment. it

necessitates the requirement for a non classic technique which has the ability to model the

physical process accurately. since nonlinear governing relationships can be handled very

contendly by neutral network, these networks offer a cost effective solution to modeling of time

varying chemical process.

Using ANN carries out the modeling of the process by using ANN by any one of the following

two ways:

Forward modeling

Page 13: Seminar Report ANN

Direct inverse modeling

FORWARD MODELING

The basic configuration used for non-linear system modeling and identification using

neural network. The number of input nodes specifies the dimensions of the network input. In

system identification context, the assignment of network input and output to network input

vector.

DIRECT INVERSE MODELING:

This approach employs a generalized model suggested by Psaltis et al.to learn the inverse

dynamic model of the plant as a feed forward controller. Here, during the training stage, the

control input are chosen randomly within there working range. And the corresponding plant

output values are stored, as a training of the controller cannot guarantee the inclusion of all

possible situations that may occur in future. Thus, the developed model has take of robustness

The design of the identification experiment used to guarantee data for training the neural

network models is crucial, particularly, in-linear problem. The training data must contain process

input-output information over the entire operating range. In such experiment, the types of

manipulated variables used are very important.

The traditional pseudo binary sequence (PRBS) is inadequate because the training data

set contains most of its steady state information at only two levels, allowing only to fit linear

model in over to overcome the problem with binary signal and to provide data points throughout

the range of manipulated variables. The PRBS must be a multilevel sequence. This kind of

modeling of the process play a vital role in ANN based direct inverse control configuration.

ANN BASED CONTROL CONFIGURATION:

Page 14: Seminar Report ANN

Direct inverse control

Direct adaptive control

Indirect adaptive control

Internal model control

Model reference adaptive control

DIRECT INVERSE CONTROL

This control configuration used the inverse planet model. Fro the direct inverse control.

The network is required to be trained offline to learn the inverse dynamics of the plant. The

networks are usually trained using the output errors of the networks and not that of the plant. The

output error of the networks is defined.

En=1/2(ud-on)2

Where En is the networks output error ud is the actual controls signal required to get desired

process output and on is the networks output. When the networks is to be trained as a controller.

The output errors of the networks are unknown. Once the network is trained using direct inverse

modeling learns the inverse system model. It is directly placed in series with the plant to be

controlled and the configuration shown in figure. Since inverse model of the plant is in off line

trained model, it tacks robustness.

Page 15: Seminar Report ANN

DIRECT ADAPTIVE CONTROL (DAC)

In the direct adaptive control. The network is trained on line. And the weights of

connections are updated during each sampling interval. In this case’ the cost function is the plant

output error rather than the networks output error. The configuration of DAC is shown in figure.

The limitation of using this configuration is that one must have the some knowledge

about the plant dynamics i.e. Jacobin matrix of the plant. To solve the problems; initially, Psaltis

D.et al proposed a technique for determining the partial derivatives of the plant at its operating

point Xianzhang et al and Yao Zhang et al presented a simple approach, in which by using the

sign of the plant Jacobin. The modifications of the weights are carried out.

INDIRECT ADAPTIVE CONTROL

Page 16: Seminar Report ANN

Narendra K S et al. Presented an indirect adaptive control strategy.

In this approach, two neutral networks for controller purpose and another for plant modeling and

is called plant emulator decides the performance of the controller. The configuration of indirect

adaptive control scheme becomes as shown FIG.3.

In direct learning the neutral controller, it is well known that the partial derivatives of the

controlled plant output with respect to the plant input (plant Jacobin) is required. One method to

overcome this problem is the use NN to identify the plant, and to calculate its Jacobin. Since the

plant emilator learning converges before the neutral controllers learning begins, an effective

neutral control system is achieved.

INTERNAL MODEL CONTROL (IMC)

Page 17: Seminar Report ANN

The IMC uses two neutral networks for implementation. I n this configrurations,one

neutral networks is placed in parallel with the plant and other neutral network in series the plant.

The structure of nonlinear IMC is shown in FIG.4.

The IMC provides a direct method for the design of nonlinear feedback controllers. If a

good model of the plant is savable, the close loop system gives an exact set point tracking despite

immeasurable disturbance acting on the plant.

For the development of NN based IMC, the following two steps are required:

Plant identification

Plant inverse model

Page 18: Seminar Report ANN

Plant identification is carried out using the forward modeling techniques. Once the network is

trained, it represents the perfect dynamics of the plant the error signal used to adjust the networks

weights is the difference between the plant output and the model output.

The neutral networks used to represent the inverse of the plant (NCC) are trained using

the plant itself. The error signal used to train the plant inverse model is the difference between

the desired plant and model outputs.

DIRECT NEURAL NETWORK MODEL REFERENCE ADAPTIVE

CONTROL:

The neural network approximates a wide variety of nonlinear control laws by adjusting

the weights in training to achieve the desired approximate accuracy. One possible MRAC

structure based on neural network is shown

In this configuration, control systems attempted to make the plant output YP(t) to follow

the reference model output asymptotically. The error signal

Used to train the neural network controller is the difference between the model and the plant

outputs, principally; this network works like the direct adaptive neural control system.

Page 19: Seminar Report ANN
Page 20: Seminar Report ANN

APPLICATION’S:

Aerospace

– High performance aircraft autopilots, flight path simulations, aircraft control

systems, autopilot enhancements, aircraft component simulations, aircraft

component fault detectors

Automotive

– Automobile automatic guidance systems, warranty activity analyzers

Banking

– Check and other document readers, credit application evaluators

Defense

– Weapon steering, target tracking, object discrimination, facial recognition, new

kinds of sensors, sonar, radar and image signal processing including data

compression, feature extraction and noise suppression, signal/image

identification

Electronics

– Code sequence prediction, integrated circuit chip layout, process control, chip

failure analysis, machine vision, voice synthesis, nonlinear modeling

Financial

– Real estate appraisal, loan advisor, mortgage screening, corporate bond rating,

credit line use analysis, portfolio trading program, corporate financial analysis,

currency price prediction

Page 21: Seminar Report ANN

Manufacturing

– Manufacturing process control, product design and analysis, process and

machine diagnosis, real-time particle identification, visual quality inspection

systems, beer testing, welding quality analysis, paper quality prediction,

computer chip quality analysis, analysis of grinding operations, chemical product

design analysis, machine maintenance analysis, project bidding, planning and

management, dynamic modeling of chemical process systems

Medical

– Breast cancer cell analysis, EEG and ECG analysis, prosthesis design,

optimization of transplant times, hospital expense reduction, hospital quality

improvement, emergency room test advisement

Robotics

– Trajectory control, forklift robot, manipulator controllers, vision systems

Speech

– Speech recognition, speech compression, vowel classification, text to speech

synthesis

Securities

– Market analysis, automatic bond rating, stock trading advisory systems

Telecommunications

– Image and data compression, automated information services, real-time

translation of spoken language, customer payment processing systems

Transportation

– Truck brake diagnosis systems, vehicle scheduling, routing systems

Page 22: Seminar Report ANN

CONCLUSION

Artificial neural networks thus have paved the way for automatic analysis of

biological data. The simultaneous analysis of millions of genes at a time has driven the new

field of computational biology—Genome informatics.

Application of artificial neural networks in Genome informatics has great

significance in this arena. The artificial neural network attains its knowledge through the

process of learning, which is similar to its inspiration---the Human brain.

But like all clouds with a silver lining Genomic engineering can be misused too, which can be a real threat to the mankind. Let us hope that man puts his brain for the welfare of his fellow beings and not against them.