Arama Sonuçları

Listeleniyor 1 - 4 / 4
  • Yayın
    VC-dimension of univariate decision trees
    (IEEE-INST Electrical Electronics Engineers Inc, 2015-02-25) Yıldız, Olcay Taner
    In this paper, we give and prove the lower bounds of the Vapnik-Chervonenkis (VC)-dimension of the univariate decision tree hypothesis class. The VC-dimension of the univariate decision tree depends on the VC-dimension values of its subtrees and the number of inputs. Via a search algorithm that calculates the VC-dimension of univariate decision trees exhaustively, we show that our VC-dimension bounds are tight for simple trees. To verify that the VC-dimension bounds are useful, we also use them to get VC-generalization bounds for complexity control using structural risk minimization in decision trees, i.e., pruning. Our simulation results show that structural risk minimization pruning using the VC-dimension bounds finds trees that are more accurate as those pruned using cross validation.
  • Yayın
    Incremental construction of classifier and discriminant ensembles
    (Elsevier Science Inc, 2009-04-15) Ulaş, Aydın; Semerci, Murat; Yıldız, Olcay Taner; Alpaydın, Ahmet İbrahim Ethem
    We discuss approaches to incrementally construct an ensemble. The first constructs an ensemble of classifiers choosing a subset from a larger set, and the second constructs an ensemble of discriminants, where a classifier is used for some classes only. We investigate criteria including accuracy, significant improvement, diversity, correlation, and the role of search direction. For discriminant ensembles, we test subset selection and trees. Fusion is by voting or by a linear model. Using 14 classifiers on 38 data sets. incremental search finds small, accurate ensembles in polynomial time. The discriminant ensemble uses a subset of discriminants and is simpler, interpretable, and accurate. We see that an incremental ensemble has higher accuracy than bagging and random subspace method; and it has a comparable accuracy to AdaBoost. but fewer classifiers.
  • Yayın
    Cost-conscious comparison of supervised learning algorithms over multiple data sets
    (Elsevier Sci Ltd, 2012-04) Ulaş, Aydın; Yıldız, Olcay Taner; Alpaydın, Ahmet İbrahim Ethem
    In the literature, there exist statistical tests to compare supervised learning algorithms on multiple data sets in terms of accuracy but they do not always generate an ordering. We propose Multi(2)Test, a generalization of our previous work, for ordering multiple learning algorithms on multiple data sets from "best" to "worst" where our goodness measure is composed of a prior cost term additional to generalization error. Our simulations show that Multi2Test generates orderings using pairwise tests on error and different types of cost using time and space complexity of the learning algorithms.
  • Yayın
    Theta and Beta1 frequency band values predict dyslexia classification
    (John Wiley and Sons Ltd, 2025-12-29) Eroğlu, Günet; Harb, Mhd Raja Abou
    Dyslexia, impacting children's reading skills, prompts families to seek cost-effective neurofeedback therapy solutions. Utilising machine learning, we identified predictive factors for dyslexia classification. Employing advanced techniques, we gathered 14-channel Quantitative Electroencephalography (QEEG) data from 200 participants, achieving 99.6% dyslexic classification accuracy through cross-validation. During validation, 48% of dyslexic children's sessions were consistently classified as normal, with a 95% confidence interval of 47.31 to 48.68. Focusing on individuals consistently diagnosed with dyslexia during therapy, we found that dyslexic individuals exhibited higher theta values and lower beta1 values compared to typically developing children. This study pioneers machine learning in predicting dyslexia classification factors, offering valuable insights for families considering neurofeedback therapy investment.