Improvement in The Speed of Training A Neural Network Using Sampled Training
Improvement in The Speed of Training A Neural Network Using Sampled Training
Improvement in The Speed of Training A Neural Network Using Sampled Training
ABSTRACT
Today the most important use of neural network is in classifying the data. The neural network is trained for classification by giving some input-output pairs which is also known as supervised learning. As training a network is a part of the whole classification process, the time required for training should be as small as possible. In this paper, we propose a scheme where sampling of the training set is done for reducing the time required for training the network and hence the overall time required for classification is reduced.
Keywords: Supervised Learning, Learning rate, Classification, Training Time, Neural Network, Sampled Training.
1. INTRODUCTION
Classification is basically a set of activities responsible for deriving a model that categorizes and describes classes of data and concepts, whose sole purpose is to determine and predict the classes of objects who have no label as [1]. The model derived can be shown in various forms, such as classification (IF-THEN) rules, decision trees, mathematical formulae, or neural network. A neural network, when used for classification, is typically a collection of neuron-like processing units with weighted connections between the units. Fig 1 shown below is the example of a Feed-Forward Artificial Neural Network. Every neural network has one input and one output unit along with zero or more number of hidden units. The input unit receives the input, the output unit generates the classification result and the hidden unit performs the processing. There are some very important components those make up an artificial neuron. These include weighting factor, summation function, transfer function, error function, error and back-propagated value and learning function. The most important among these is the learning function. Its purpose is to modify the variable connection weights when inputs are given to each processing element based on some neural based algorithm.
Figure 1 A Feed-Forward Neural Network There are a large number of neural network algorithms that are implemented for different applications. In each type of algorithms, some type of learning technique is always used. The time required for training a network should be very less so that the actual data is classified in less time. Further this paper is divided into following sections, section II is about training a neural network. Section III is about the proposed scheme of sampled training in a neural network, section IV is about the advantages of the proposed scheme and section V is the conclusion so derived after the study of the two schemes.
Page 241
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com Volume 2, Issue 4, April 2013 ISSN 2319 - 4847
In a feed-forward neural network, training of the network is done by providing the input-output pairs to the network and then generating the model which can then be used for classification of the test data. In case of the existing system, the whole of the training set is given as input for training the network. So the total time required for performing the classification includes the time required for training and the time required for classifying the test data based on the model generated. It has found that for better error gradient estimations, more number of input vectors for each gradient calculation should be used [2]. But the down side of using more input vectors per weight update is that the gradient has to be calculated for each input vector used [3]. In the training process it is the gradient calculations that take the longest. Using multiple input vectors quickly lengthens the training times to intolerable levels.
Figure 2 Flow-Chart of Batch Training in Neural Networks Figure 2 shows the flowchart of the batch mode training scheme that is being used in the existing system. As it can seen from the figure above, in batch mode of training a neural network, the complete training set is used for training. After the training has been accomplished and the model generated, the test data is given as input for classifying the unknown class labels. This is how the existing system works. But the problem with this training is that the time required to train a neural network is more as the whole training set is used [4]. So, in order to reduce the training time, we propose a new scheme where the training set is sampled into smaller chunks.
Page 242
Web Site: www.ijaiem.org Email: editor@ijaiem.org, editorijaiem@gmail.com Volume 2, Issue 4, April 2013 ISSN 2319 - 4847
With re-sampling taking place, the training set is reduced and so is the training time. As the training time is reduced, the overall time of classification is also reduced [5].
5. CONCLUSION
In Feed-Forward neural networks used with Gradient Descent Technique, training time is the most important parameter to be considered. As the gradient calculations require or consume more time than actual training. So the technique of using the training set should be checked to reduce the training time. In batch mode training, the numbers of weight updates are more with increase in the training data which in turn results in the increase of gradient calculations. Due to more gradient calculations, the training time also increases. In sampled training, as the original training data is sampled into fixed size samples and then given as input to the network for training, the number of weight updates is less as compared to number of weight updates in batch mode training. And so, as a result gradient calculation also reduces which results in lesser training times.
6. ACKNOWLEDGEMENTS
Advait Bhatt wishes to thank Prof. H.B.Jethva, for his guidance and help for doing this work. He also acknowledges Prof. D. A. Parikh, Head of computer department, and to all staff of computer department for full support for completion of this work. Prof. Harikrishna Jethva wishes to acknowledge his family and staff of computer department at L.D.College of engineering.
References
[1] Jiaweu Han and Micheline Kamber, Data Mining Concepts and Techniques , second edition, Elsevier press [2] Nortje William Daniel, Comparison of Bayesian Learning and Conjugate Gradient Descent Training Of Neural Networks, University of Pretoria, October 2001. [3] Er. Shegal Parveen, Dr. Gupta Sangeeta and Prof. Kumar Dharminder, Minimization of Error in Training A Neural Network Using Gradient Descent Technique, International Journal of Technical Research (IJTR), Vol-1, Issue-1, Mar-Apr 2012. [4] Gong Liang, Liu Chengliang and Yuan Fuqing, Training Feed-forward Neural Network Using the Gradient Descent Method With The Optimal Step Size, Journal of Computational Information System 8:4 (2012). [5] Nandy Sudarshan, Sarkar Prartha Prathim and Das Achintya, An Improved Gauss-Newtons Method Based BackPropagation Algorithm For Fast Convergence, International Journal for Computer Applications, Vol 39-No. 8, February 2012. AUTHOR Advait S. Bhatt, pursuing his Master Degree in Computer Science & Technology from Gujarat Technological University (L D College of Engg., Ahmedabad), received his Bachelor Degree in Information Technology from Dharmsinh Desai University, Nadiad in 2010. His area of interest include Database Systems, Data Mining, C/C++/Java/Advanced java Programming. Prof. Harikrishna B. Jethva, received his post graduate degree in Computer engineering from Dharmsinh Desai University in 2009 and Bachelor Degree in Computer Engineering from Saurashtra University, Rajkot in 2001. He has worked as a Assistant Professor for 10 Years and presently working as a Associate Professor in L.D.College of Engineering. His area of interest is in Neural Network, Theory of Computation, Computer Network, Compiler Design, Soft Computing and Distributed Computing.
Page 243