کد مقاله کد نشریه سال انتشار مقاله انگلیسی نسخه تمام متن
476533 699859 2011 13 صفحه PDF دانلود رایگان
عنوان انگلیسی مقاله ISI
Shannon Entropy and Mean Square Errors for speeding the convergence of Multilayer Neural Networks: A comparative approach
موضوعات مرتبط
مهندسی و علوم پایه مهندسی کامپیوتر علوم کامپیوتر (عمومی)
پیش نمایش صفحه اول مقاله
Shannon Entropy and Mean Square Errors for speeding the convergence of Multilayer Neural Networks: A comparative approach
چکیده انگلیسی

Improving the efficiency and convergence rate of the Multilayer Backpropagation Neural Network Algorithms is an active area of research. The last years have witnessed an increasing attention to entropy based criteria in adaptive systems. Several principles were proposed based on the maximization or minimization of entropic cost functions. One way of entropy criteria in learning systems is to minimize the entropy of the error between two variables: typically one is the output of the learning system and the other is the target. In this paper, improving the efficiency and convergence rate of Multilayer Backpropagation (BP) Neural Networks was proposed. The usual Mean Square Error (MSE) minimization principle is substituted by the minimization of Shannon Entropy (SE) of the differences between the multilayer perceptions output and the desired target. These two cost functions are studied, analyzed and tested with two different activation functions namely, the Cauchy and the hyperbolic tangent activation functions. The comparative approach indicates that the Degree of convergence using Shannon Entropy cost function is higher than its counterpart using MSE and that MSE speeds the convergence than Shannon Entropy.

ناشر
Database: Elsevier - ScienceDirect (ساینس دایرکت)
Journal: Egyptian Informatics Journal - Volume 12, Issue 3, November 2011, Pages 197–209
نویسندگان
,