2 sonuçlar
Arama Sonuçları
Listeleniyor 1 - 2 / 2
Yayın Subset selection for tuning of hyper-parameters in artificial neural networks(IEEE, 2017) Aki, K.K.Emre; Erkoç, Tuğba; Eskil, Mustafa TanerHyper-parameters of a machine learning architecture define its design. Tuning of hyper-parameters is costly and for large data sets outright impractical, whether it is performed manually or algorithmically. In this study we propose a Neocognitron based method for reducing the training set to a fraction, while keeping the dynamics and complexity of the domain. Our approach does not require processing of the entire training set, making it feasible for larger data sets. In our experiments we could successfully reduce the MNIST training data set to less than 2.5% (1,489 images) by processing less than 10% of the 60K images. We showed that the reduced data set can be used for tuning of number of hidden neurons in a multi-layer perceptron.Yayın El yazısı rakam sınıflandırması için gözetimsiz benzerlik tabanlı evrişimler(Institute of Electrical and Electronics Engineers Inc., 2022) Erkoç, Tuğba; Eskil, Mustafa TanerEffective training of filters in Convolutional Neural Networks (CNN) ensures their success. In order to achieve good classification results in CNNs, filters must be carefully initialized, trained and fine-tuned. We propose an unsupervised method that allows the discovery of filters from the given dataset in a single epoch without specifying the number of filters hyper-parameter in convolutional layers. Our proposed method gradually builds the convolutional layers by a discovery routine that extracts a number of features that adequately represent the complexity of the input domain. The discovered filters represent the patterns in the domain, so they do not require any initialization method or backpropagation training for fine tuning purposes. Our method achieves 99.03% accuracy on MNIST dataset without applying any data augmentation techniques.












