Over fitting considerations - artificial intelligence, Computer Engineering

Over fitting Considerations - artificial intelligence

Left  unexamined ,  back  propagation  in  multi-layer  networks  may  be very susceptible  to over fitting itself to the training examples. The following graph plots the error on the training and test set as the number of weight updates increases. It is error prone of networks left to train unchecked.

810_Over fitting Considerations.png

Alarmingly, even though the error on the training set continues to slowly decrease, the error on the test set essentially begins to increase towards the end. It is clearly over fitting, and it relates to the network starting to find and fine-tune to idiosyncrasies in the data, rather than to general properties. Given this phenomena, it would not be wise to use some sort of threshold for the error as the termination condition for back propagation.

In the cases where the number of training examples is high, one antidote to over fitting is to crack the training examples into a set to use to train the weight and a set to hold back as an internal validation set. This is a mini-test set, which may be used to keep the network in check: if the error on the validation set reaches minima and then start to increase, then it could be over fitting in beginning to occur.

Note that (time permitting) it is good giving the training algorithm the advantage of the doubt as much as possible. That is, in the validation set, the error may also go through local minima, and it is unwise to stop training as soon as the validation set error begin to increase, as a better minima can be achieved later on. Of course, if the minima are never bettered, then the network which is in final presented by the learning algorithm should be re-wound to be the 1 which produced the minimum on the validation set.

Another way around over fitting is to decrease each weight by a little weight decay factor during each epoch. Learned networks with large (negative or positive) weights tend to have over fitted the data, because larger weights are needed to accommodate outliers in the data. Thus, keeping the weights low with a weight decay factor can help to steer the network from over fitting.

Posted Date: 10/2/2012 6:39:54 AM | Location : United States







Related Discussions:- Over fitting considerations - artificial intelligence, Assignment Help, Ask Question on Over fitting considerations - artificial intelligence, Get Answer, Expert's Help, Over fitting considerations - artificial intelligence Discussions

Write discussion on Over fitting considerations - artificial intelligence
Your posts are moderated
Related Questions
Determine the primary memory of the server The keys used for accessing the server are held at a secret location in the primary memory of the server. This area or location is hi

How does the Xml Serializer work?  What ACL permissions does a process using it require?   Xml Serializer needs write permission to the system's TEMP directory.

Visual basic is useful if you are planning to make the programs from scratch. This language helps you in producing Active x controls, exe files, etc. Visual script is a powerful t

List the factors that determine the storage device performance. i.Access time ii.Cycle time iii.Transfer Rate.

Explain Classification Based On Grain Size This categorization is based on distinguishing the parallelism in a program to be executed on a multiprocessor system. The concept is

Explain File scope File scope: The variables and functions with file scope appear outside any block or list of parameters and are accessible from any place in the translation

Basic Software: The basic software is also referred to as utilities. Basic software packages are available for performing operations such as data entry and validation, sorting

What is the purpose of DOTNET? Dot Net Is a collection of products like C#, ASp.Net, Vb.Net, XML. The purpose of .Net is that we can simply upgrade and degrade our programs whi