|کد مقاله||کد نشریه||سال انتشار||مقاله انگلیسی||ترجمه فارسی||نسخه تمام متن|
|4969161||1365259||2018||12 صفحه PDF||ندارد||دانلود رایگان|
â¢Explores how binarization permits/improves diversification in deep machines.â¢Shows the effectiveness of pre-emphasizing samples for deep classification.â¢Combines the above with data augmentation to reach record results.â¢Opens further research lines in deep learning.
To aggregate diverse learners and to train deep architectures are the two principal avenues towards increasing the expressive capabilities of neural networks. Therefore, their combinations merit attention. In this contribution, we study how to apply some conventional diversity methods âbagging and label switchingâ to a general deep machine, the stacked denoising auto-encoding classifier, in order to solve a number of appropriately selected image recognition problems. The main conclusion of our work is that binarizing multi-class problems is the key to obtain benefit from those diversity methods.Additionally, we check that adding other kinds of performance improvement procedures, such as pre-emphasizing training samples and elastic distortion mechanisms, further increases the quality of the results. In particular, an appropriate combination of all the above methods leads us to reach a new absolute record in classifying MNIST handwritten digits.These facts reveal that there are clear opportunities for designing more powerful classifiers by means of combining different improvement techniques.
Journal: Information Fusion - Volume 39, January 2018, Pages 41-52