Professional Documents
Culture Documents
A SEMINAR REPORT ON
SUBMITTED BY -
SUBMITTED TO -
ACKNOWLEDGEMENT
I pose my copious gratitude and like to thank the entire staff of G.L.A.I.T.M, Mathura for their help and kind cooperation during my entire seminar preparation. I am extremely thankful to them for providing me with vital information about the topic. I rejoice in expressing my prodigious gratification to Department of Electronics &Communication Engineering Department, G.L.A.I.T.M, Mathura for his indispensable guidance, generous help, perpetual encouragement, constant attention offered throughout in preparing the seminar. I take this opportunity to pay my sincere thanks to Mr. MANISH KASHYAP, seminar in charge & lecturer, Electronics & Communication Engineering Department, G.L.A. Institute of Technology & Management, Mathura, for giving me the golden opportunity to present the seminar. At last but not the least, I would like to thank my parents and all my peers who have been a constant source of encouragement and inspiration in every walk of life. VIVEK YADAV
CERTIFICATE
This is to certify that the work which is being successfully presented in the seminar report entitled NEURAL NETWORKS & THEIR APPLICATIONS by me in the partial fulfillment of the requirement for the award of Bachelor Of Technology Degree in Electronics & Communication Engineering Department at G.L.A. Institute Of Technology & Management, Mathura from Uttar Pradesh Technical University, Lucknow. The matter embodied in this dissertation has not been submitted by me for award of any other degree. DATE-: 21 ,APRIL , 2011 This is to certify that the above statement made by the candidate is correct to the best of my knowledge. SUBMITTED BY -: SUBMITTED TO-: VIVEKYADAV KASHYAP) B.TECH. III YEAR (EC) INCHARGE ROLL NO.: 0806331112 SEMINAR (MR. MANISH
CONTENTS
Topic Page no.
Abstract 5
1. Introduction
6 Biological neuron. 8
2. Artificial
neuron .9
3. Different
models
of
neuron.10 4. Classical
5. Artificial
function.14 network..16
6. Qualities
of architecture
neural of
network17
7. Different
ANN19
8. Learning
ofANN 20
9. Advantages&Disadvantages
ofANN22
10. Applications
of advances in field of
ANN23
11. Recent
ANN.25
12. Conclusion
27
13. Bibliography
28
ABSTRACT
Neural network are inspired by biological nervous system and re composed of many simple computational elements operating in parallel. In this study basic component of neural network are introduced and brief on their working. Concept of activation function is also discussed. Different learning algorithm are also enlisted.Topics on advantage ,disadvantage and recent development in the field of ANNs are mentioned in this study.
Introduction
Neural networks have seen an explosion of interest over the last few years and are being successfully applied across an extraordinary range of problem domains, in areas as diverse as finance, medicine, engineering, geology, physics and biology. The excitement stems from the fact that these networks are attempts to model the capabilities of the human brain. From a statistical perspective neural networks are interesting because of their potential use inprediction and classification problems. Artificial neural networks (ANNs) are non-linear data driven self adaptive approach as opposed to the traditional model based methods. They are powerful tools for modelling, especially when the underlying data relationship is unknown. ANNs can identify and learn correlated patterns between input data sets and corresponding target values. After training,ANNs can be used to predict the outcome of new independent input data. ANNs imitate the learning process of the human brain and can process problems involving non-linear and complex data even if the data are imprecise and noisy. Thus they are ideally suited for the modeling of
agricultural data which are known to be complex and often non-linear. These networks are neural in the sense that they may have been inspired by neuroscience but not necessarily because they are faithful models of biological neural or cognitive phenomena. In fact majority of the network are more closely related to traditional mathematical and/or statistical models such as nonparametric pattern classifiers, clustering algorithms, nonlinear filters, and statistical regression models than they are to neurobiology models. Neural networks (NNs) have been used for a wide variety of applications where statistical methods are traditionally employed. They have been used in classification problems, such as identifying underwater sonar currents, recognizing speech, and predicting the secondary structure of globular proteins. In time-series applications, NNs have been used in predicting stock market classical performance. statistical and As statisticians such as or users of statistics, these problems are normally solved through methods, discriminant It is, analysis, logistic regression, Bayes analysis, multiple regression, ARIMA time-series models. therefore, time to recognize neural networks as a powerful tool for data analysis.
Input (xi)
Typically, these values are external stimuli from the environment or come from the outputs of other artificial neurons. They can be discrete values from a set, such as {0,1}, or real-valued numbers.
Weights (wi)
These are real-valued numbers that determine the contribution of each input to the neuron's weighted sum and eventually its output. The goal of neural network training algorithms is to determine the best possible set of weight values for the problem under consideration. Finding the optimal set is often a tradeoff between computation time and minimizing the network error.
Threshold (u)
The threshold is referred to as a bias value. In this case, the real number is added to the weighted sum.
For simplicity, the threshold can be regarded as another input / weight pair, where w0 = u and x0 = -1.
this device consists of a set of controllable resistors connected to a circuit which can sum up currents caused by the input voltage signals. Usually the central block, the summer, is also followed by a quantiser which outputs either +1 of 1,depending on the polarity of the sum. Although the adaptive process is here exemplified in a case when there is only one output, it may be clear that a system with many parallel outputs is directly implementable by multiple units of the above kind. If the input conductances are denoted by wi, i = 0; 1; : : : ; n, and the input and output signals by xi and y, respectively, then the output of the central block is defined to be:
where = w0.
(2.2)
are weight values normalized in the range of either input line, or and associated with each is a threshold
Different types of activation functions can be used and three of them are described in [activation functions] . The most commonly used is nonlinear sigmoid activation functions such as the logistic function . A logistic function assumes a continuous range of values form 0 and 1 in contrary to the discrete threshold function. A binary threshold function was used in the first model of an artificial neuron back in 1943, the so-called McCulloch-Pitts model . Threshold functions goes by many names, e.g. step-function , heavyside function , hard-limiter etc. Common for all is that they produce one of two scalar output values (usually 1 and -1 or 0 and 1) depending on the value of the threshold. Another type of activation function is the linear function or some times called the identity function since the activation is just the input. In general if the task is to approximate some function then the output nodes are linear and if the task is classification then sigmoidal output nodes are used
2. Parallel processing 3. Fault tolerance 4. Self organising 5. Ability to generalize 6. Complete computability 7. Continuous adaptability
Fig-:Hopfield network
Fig-:Kehonen network
Learning of ANN
An ANN learns from its experience. The usual process of learning involves three tasks: 1.Compute output(s). 2.Compare outputs with desired patterns and feed-back the error.
3.Adjust the weights and repeat the process 4.The learning process starts by setting the weights by some rules . The difference between the actual output (y) and the desired output(z) is called error (delta). 5.The objective is to minimize delta (error)to zero. The reduction in error is done by changing the weights 1.Supervised learning-: or Associative learning in which the network is trained by providing it with input and matching output patterns. These input-output pairs can be provided by an external teacher, or by the system which contains the neural network (selfsupervised).
2.Unsupervised learning -:or Self-organisation in which an (output) unit is trained to respond to clusters of pattern within the input. In this paradigm the system is supposed to discover statistically salient features of
the input population. Unlike the supervised learning paradigm, there is no a priori set of categories into which the patterns are to be classified; rather the system must develop its own representation of the input stimuli. 3.Reinforcement Learning -:This type of learning may be considered as an intermediate form of the above two types of learning. Here the learning machine does some action on the environment and gets a feedback response from the environment. The learning system grades its action good (rewarding) or bad (punishable) based on the environmental response and accordingly adjusts its parameters. Generally, parameter adjustment is continued until an equilibrium state occurs, following which there will be no more changes in its parameters. The selforganizing neural learning may be categorized under this type of learning.
Advantages&Disadvantages of ANN
A.)Advantages
1.Adapt to unknown situations 2.Robustness: fault tolerance due to network redundancy 3.Autonomous learning and generalization
B.)Disadvantages
1.Not exact 2.Large complexity of the network structure.
APPLICATIONS OF ANN
Fuzzy logic is a type of logic that recognizes more than simple true and false values, hence better simulating the real world. For example, the statement today is sunny might be 100% true if there are no clouds, 80% true if there are a few clouds, 50% true if it's hazy, and 0% true if rains all day. Hence, it takes into account concepts like -usually, somewhat, and sometimes.
2.
Fuzzy logic and neural networks have been integrated for uses as diverse as automotive engineering,
applicant screening for jobs, the control of a crane, and the monitoring of glaucoma.
"Most practical applications of artificial neural networks are based on a computational model involving the propagation of continuous variables from one processing unit to the next. In recent years, data from neurobiological experiments have made it increasingly clear that biological neural networks, which communicate through pulses, use the timing of the pulses to transmit information and perform computation. This realization has stimulated significant research on pulsed neural networks, including theoretical analyses and model development, neurobiological modeling, and hardware implementation."
Some networks have been hardcoded into chips or analog devices ? this technology will become more useful as the networks we use become more complex.
2.
The primary benefit of directly encoding neural networks onto chips or specialized analog devices is SPEED!
3.
NN hardware currently runs in a few niche areas, such as those areas where very high performance is required (e.g. high energy physics) and in embedded
Many NNs today use less than 100 neurons and only need occasional training. In these situations, software simulation is usually found sufficient
When NN algorithms develop to the point where useful things can be done with 1000's of neurons and 10000's of synapses, high performance NN hardware will become essential for practical operation.
CONCLUSION
1.
All current NN technologies will most likely be vastly improved upon in the future. Everything from handwriting and speech recognition to stock market prediction will become more sophisticated as researchers develop better training methods and network architectures
2.
Although neural networks do seem to be able to solve many problems, we must put our exuberance in check sometimes ? they are not magic! Overconfidence in neural networks can result in costly mistakes: see for a rather funny story about the government and neural networks.
.
the under standing of the data compiled by the Human Genome Project
g. Self-diagnosis of medical problems using
neural networks
h. And much more!
REFRENCES
1.
Hertz, J., Palmer, R.G., Krogh. A.S. (1990) Introduction to the theory of
neural
computation,
Peruses
Books.
ISBN 0-201-51560-1
2.
B.Yegnarayana(2010) 203-1253-1
Artificial neural
3.