9 sonuçlar
Arama Sonuçları
Listeleniyor 1 - 9 / 9
Yayın Identification of metabolic correlates of mild cognitive impairment in Parkinson's disease using magnetic resonance spectroscopic imaging and machine learning(Springer Science and Business Media Deutschland GmbH, 2022-12) Cengiz, Sevim; Arslan, Dilek Betül; Kıçik, Ani; Erdoğdu, Emel; Yıldırım, Muhammed; Hatay, Gökçe Hale; Tüfekçioğlu, Zeynep; Uluğ, Aziz Müfit; Bilgiç, Başar; Hanagasi, Haşmet; Demiralp, Tamer; Gürvit, Hakan; Öztürk Işıkk, EsinObjective: To investigate metabolic changes of mild cognitive impairment in Parkinson’s disease (PD-MCI) using proton magnetic resonance spectroscopic imaging (1H-MRSI). Methods: Sixteen healthy controls (HC), 26 cognitively normal Parkinson’s disease (PD-CN) patients, and 34 PD-MCI patients were scanned in this prospective study. Neuropsychological tests were performed, and three-dimensional 1H-MRSI was obtained at 3 T. Metabolic parameters and neuropsychological test scores were compared between PD-MCI, PD-CN, and HC. The correlations between neuropsychological test scores and metabolic intensities were also assessed. Supervised machine learning algorithms were applied to classify HC, PD-CN, and PD-MCI groups based on metabolite levels. Results: PD-MCI had a lower corrected total N-acetylaspartate over total creatine ratio (tNAA/tCr) in the right precentral gyrus, corresponding to the sensorimotor network (p = 0.01), and a lower tNAA over myoinositol ratio (tNAA/mI) at a part of the default mode network, corresponding to the retrosplenial cortex (p = 0.04) than PD-CN. The HC and PD-MCI patients were classified with an accuracy of 86.4% (sensitivity = 72.7% and specificity = 81.8%) using bagged trees. Conclusion: 1H-MRSI revealed metabolic changes in the default mode, ventral attention/salience, and sensorimotor networks of PD-MCI patients, which could be summarized mainly as ‘posterior cortical metabolic changes’ related with cognitive dysfunction.Yayın Effective semi-supervised learning strategies for automatic sentence segmentation(Elsevier Science BV, 2018-04-01) Dalva, Doğan; Güz, Ümit; Gürkan, HakanThe primary objective of sentence segmentation process is to determine the sentence boundaries of a stream of words output by the automatic speech recognizers. Statistical methods developed for sentence segmentation requires a significant amount of labeled data which is time-consuming, labor intensive and expensive. In this work, we propose new multi-view semi-supervised learning strategies for sentence boundary classification problem using lexical, prosodic, and morphological information. The aim is to find effective semi-supervised machine learning strategies when only small sets of sentence boundary labeled data are available. We primarily investigate two semi-supervised learning approaches, called self-training and co-training. Different example selection strategies were also used for co-training, namely, agreement, disagreement and self-combined. Furthermore, we propose three-view and committee-based algorithms incorporating with agreement, disagreement and self-combined strategies using three disjoint feature sets. We present comparative results of different learning strategies on the sentence segmentation task. The experimental results show that the sentence segmentation performance can be highly improved using multi-view learning strategies that we proposed since data sets can be represented by three redundantly sufficient and disjoint feature sets. We show that the proposed strategies substantially improve the average baseline F-measure of 67.66% to 75.15% and 64.84% to 66.32% when only a small set of manually labeled data is available for Turkish and English spoken languages, respectively.Yayın VC-dimension of univariate decision trees(IEEE-INST Electrical Electronics Engineers Inc, 2015-02-25) Yıldız, Olcay TanerIn this paper, we give and prove the lower bounds of the Vapnik-Chervonenkis (VC)-dimension of the univariate decision tree hypothesis class. The VC-dimension of the univariate decision tree depends on the VC-dimension values of its subtrees and the number of inputs. Via a search algorithm that calculates the VC-dimension of univariate decision trees exhaustively, we show that our VC-dimension bounds are tight for simple trees. To verify that the VC-dimension bounds are useful, we also use them to get VC-generalization bounds for complexity control using structural risk minimization in decision trees, i.e., pruning. Our simulation results show that structural risk minimization pruning using the VC-dimension bounds finds trees that are more accurate as those pruned using cross validation.Yayın A cooperative neural network control structure and its application for systems having dead-zone nonlinearities(Springer International Publishing Ag, 2022-03) Dinçmen, ErkinAn adaptive control structure utilizing two feed-forward neural networks (NN) is proposed to deal with systems having unknown nonlinearities. One of the networks is trained to mimic the nonlinear system dynamics. Its training will be repeated with periods in order to keep it an updated valid model of the system all the times since the parameters and/or nonlinearities of the system may change during time. The other network, which is the Controller NN, adapts itself continuously by collaborating with the Model NN. The stability-convergence analysis of both networks is performed via Lyapunov method. An example system is chosen to show the applicability of the control algorithm. This example system is created by combining a linear dynamics model with a dead-zone function to represent a nonlinear system to be controlled. It should be noted that the proposed control structure can be used in any nonlinear system without knowing the system dynamics. The only information required by Model NN is the training set consisting input-output data pairs of the system. The Model NN is trained offline with this training set, and afterward the Controller NN adapts its weights online continuously during the control task with the help of Model NN. The performances of PD and PID controllers are also given for comparison purposes.Yayın Incremental construction of classifier and discriminant ensembles(Elsevier Science Inc, 2009-04-15) Ulaş, Aydın; Semerci, Murat; Yıldız, Olcay Taner; Alpaydın, Ahmet İbrahim EthemWe discuss approaches to incrementally construct an ensemble. The first constructs an ensemble of classifiers choosing a subset from a larger set, and the second constructs an ensemble of discriminants, where a classifier is used for some classes only. We investigate criteria including accuracy, significant improvement, diversity, correlation, and the role of search direction. For discriminant ensembles, we test subset selection and trees. Fusion is by voting or by a linear model. Using 14 classifiers on 38 data sets. incremental search finds small, accurate ensembles in polynomial time. The discriminant ensemble uses a subset of discriminants and is simpler, interpretable, and accurate. We see that an incremental ensemble has higher accuracy than bagging and random subspace method; and it has a comparable accuracy to AdaBoost. but fewer classifiers.Yayın Cost-conscious comparison of supervised learning algorithms over multiple data sets(Elsevier Sci Ltd, 2012-04) Ulaş, Aydın; Yıldız, Olcay Taner; Alpaydın, Ahmet İbrahim EthemIn the literature, there exist statistical tests to compare supervised learning algorithms on multiple data sets in terms of accuracy but they do not always generate an ordering. We propose Multi(2)Test, a generalization of our previous work, for ordering multiple learning algorithms on multiple data sets from "best" to "worst" where our goodness measure is composed of a prior cost term additional to generalization error. Our simulations show that Multi2Test generates orderings using pairwise tests on error and different types of cost using time and space complexity of the learning algorithms.Yayın Eigenclassifiers for combining correlated classifiers(Elsevier Science Inc, 2012-03-15) Ulaş, Aydın; Yıldız, Olcay Taner; Alpaydın, Ahmet İbrahim EthemIn practice, classifiers in an ensemble are not independent. This paper is the continuation of our previous work on ensemble subset selection [A. Ulas, M. Semerci, O.T. Yildiz, E. Alpaydin, Incremental construction of classifier and discriminant ensembles, Information Sciences, 179 (9) (2009) 1298-1318] and has two parts: first, we investigate the effect of four factors on correlation: (i) algorithms used for training, (ii) hyperparameters of the algorithms, (iii) resampled training sets, (iv) input feature subsets. Simulations using 14 classifiers on 38 data sets indicate that hyperparameters and overlapping training sets have higher effect on positive correlation than features and algorithms. Second, we propose postprocessing before fusing using principal component analysis (PCA) to form uncorrelated eigenclassifiers from a set of correlated experts. Combining the information from all classifiers may be better than subset selection where some base classifiers are pruned before combination, because using all allows redundancy.Yayın Model selection in omnivariate decision trees using Structural Risk Minimization(Elsevier Science Inc, 2011-12-01) Yıldız, Olcay TanerAs opposed to trees that use a single type of decision node, an omnivariate decision tree contains nodes of different types. We propose to use Structural Risk Minimization (SRM) to choose between node types in omnivariate decision tree construction to match the complexity of a node to the complexity of the data reaching that node. In order to apply SRM for model selection, one needs the VC-dimension of the candidate models. In this paper, we first derive the VC-dimension of the univariate model, and estimate the VC-dimension of all three models (univariate, linear multivariate or quadratic multivariate) experimentally. Second, we compare SRM with other model selection techniques including Akaike's Information Criterion (AIC), Bayesian Information Criterion (BIC) and cross-validation (CV) on standard datasets from the UCI and Delve repositories. We see that SRM induces omnivariate trees that have a small percentage of multivariate nodes close to the root and they generalize more or at least as accurately as those constructed using other model selection techniques.Yayın ANN activation function estimators for homomorphic encrypted inference(Institute of Electrical and Electronics Engineers Inc., 2025-06-13) Harb, Mhd Raja Abou; Çeliktaş, BarışHomomorphic Encryption (HE) enables secure computations on encrypted data, facilitating machine learning inference in sensitive environments such as healthcare and finance. However, efficiently handling non-linear activation functions, specifically Sigmoid and Tanh, remains a significant computational challenge for encrypted inference using Artificial Neural Networks (ANNs). This study introduces a lightweight, ANN-based estimator designed to accurately approximate activation functions under homomorphic encryption. Unlike traditional polynomial and piecewise linear approximations, the proposed ANN estimators achieve superior accuracy with lower computational overhead associated with bootstrapping or high-degree polynomial techniques. These estimators are trained on plaintext data and seamlessly integrated into encrypted inference pipelines, significantly outperforming conventional methods. Experimental evaluations demonstrate notable improvements, with ANN estimators enhancing accuracy by approximately 2% for Sigmoid and up to 73% for Tanh functions, improving F1-scores by approximately 2% for Sigmoid and up to 88% for Tanh, and markedly reducing Mean Square Error (MSE) by up to 96% compared to polynomial approximations. The ANN estimator achieves an accuracy of 97.70% and an AUC of 0.9997 when integrated into a CNN architecture on the MNIST dataset, and an accuracy of 85.25% with an AUC of 0.9459 on the UCI Heart Disease dataset during ciphertext inference. These results underscore the estimator’s practical effectiveness and computational feasibility, making it suitable for secure and efficient ANN inference in encrypted environments.












