Learning algorithm for multi-layered networks, Computer Engineering

Learning algorithm for multi-layered networks:

Furthermore details we see that  if S is too high, the contribution from wi * xi is reduced. It means that t(E) - o(E) is multiplied by xi after then if xi is a big value as positive or negative so the change to the weight will be greater. Here to get a better feel for why this direction  correction works so it's a good idea to do some simple calculations by hand. 

Here η simply controls how far the correction should go at one time that is usually set to be a fairly low value, e.g., 0.1. However the weight learning problem can be seen as finding the global minimum error which calculated as the proportion of mis-categorised training examples or over a space when all the input values can vary. Means it is possible to move too far in a direction and improve one particular weight to the detriment of the overall sum: whereas the sum may work for the training example being looked at and it may no longer be a good value for categorising all the examples correctly. Conversely for this reason here η restricts the amount of movement possible. Whether large movement is in reality required for a weight then this will happen over a series of iterations by the example set. But there sometimes η is set to decay as the number of that iterations through the entire set of training examples increases it means, can move more slowly towards the global minimum in order not to overshoot in one direction.

However this kind of gradient descent is at the heart of the learning algorithm for multi-layered networks that are discussed in the next lecture. 

Further Perceptrons with step functions have limited abilities where it comes to the range of concepts that can be learned and as discussed in a later section. The other one way to improve matters is to replace the threshold function into a linear unit through which the network outputs a real value, before than a 1 or -1. Conversely this enables us to use another rule that called the delta rule where it is also based on gradient descent.

Posted Date: 1/11/2013 7:02:52 AM | Location : United States







Related Discussions:- Learning algorithm for multi-layered networks, Assignment Help, Ask Question on Learning algorithm for multi-layered networks, Get Answer, Expert's Help, Learning algorithm for multi-layered networks Discussions

Write discussion on Learning algorithm for multi-layered networks
Your posts are moderated
Related Questions
E-brokerage. An e-brokerage is an investment house that permits you to buy and sell stocks and get investment information from its Web site.

8085 the interrupts are classified as Software and Hardware interrupts.

Choose one area of rapid technological change in IT or Computer Science and research and report on recent developments and the outlook for the future in the area that you have chos

Normal 0 false false false EN-IN X-NONE X-NONE MicrosoftInternetExplorer4 Select a range of a

MX is conceptually easy, yet bears the fruit of years of domain experience and research. In a nutshell, JMX describes a standard means for applications to expose management functio

Rectifier output with fitters: When half-wave and full-wave rectification suffice to deliver a type of DC output, neither produces constant-voltage DC (direct current). To gen

What are the requirements a dialog program must fulfill? A dialog program must fulfil the following requirements A user friendly user interface. - Format and consistency

Q. What is the impact of overflow for binary numbers? An overflow is said to have happened when sum of two n digits number takes n+ 1 digits. This definition is perfectly appli

Suppose you have to develop an error recovery protocol for a link that is unreliable and delay sensitive, which of the following protocol would you choose? (i) Stop & wait.

In this assignment, you are required to develop part of a B2C e-Commerce site which is an online movie reservation system for a multiplex theatre. Through the developed site, the u