5 sonuçlar
Arama Sonuçları
Listeleniyor 1 - 5 / 5
Yayın Convolutional attention network for MRI-based Alzheimer's disease classification and its interpretability analysis(IEEE, 2021-09-17) Türkan, Yasemin; Tek, Faik BorayNeuroimaging techniques, such as Magnetic Resonance Imaging (MRI) and Positron Emission Tomography (PET), help to identify Alzheimer's disease (AD). These techniques generate large-scale, high-dimensional, multimodal neuroimaging data, which is time-consuming and difficult to interpret and classify. Therefore, interest in deep learning approaches for the classification of 3D structural MRI brain scans has grown rapidly. In this research study, we improved the 3D VGG model proposed by Korolev et al. [2]. We increased the filters in the 3D convolutional layers and then added an attention mechanism for better classification. We compared the performance of the proposed approaches for the classification of Alzheimer's disease versus mild cognitive impairments and normal cohorts on the Alzheimer's Disease Neuroimaging Initiative (ADNI) dataset. We observed that both the accuracy and area under curve results improved with the proposed models. However, deep neural networks are black boxes that produce predictions that require further explanation for medical usage. We compared the 3D-data interpretation capabilities of the proposed models using four different interpretability methods: Occlusion, 3D Ultrametric Contour Map, 3D Gradient-Weighted Class Activation Mapping, and SHapley Additive explanations (SHAP). We observed that explanation results differed in different network models and data classes.Yayın Uyarlanır yerel bağlı katman kullanan dikkat tabanlı derin ağ ile sesli komut tanıma(Institute of Electrical and Electronics Engineers Inc., 2020-10-05) Turkan, Yasemin; Tek, Faik BoraySesli komut tanıma insan-makine ara yüzüyle ilişkili aktif bir araştırma konusudur. Dikkat tabanlı derin ağlar ile bu tür problemler başarılı bir şekilde çözülebilmektedir. Bu çalışmada, var olan bir dikkat tabanlı derin ağ yöntemi, uyarlanır yerel bağlı (odaklanan) katman kullanılarak daha da geliştirilmiştir. Orijinal yönteminde sınandığı Google ve Kaggle sesli komut veri setlerinde karşılaştırmalı olarak yapılan deneylerde önerdiğimiz uyarlanır yerel bağlı katman kullanan dikkat tabanlı ağın tanıma doğruluğunu %2.6 oranında iyileştirdiği gözlemledik.Yayın Adaptive convolution kernel for artificial neural networks(Academic Press Inc., 2021-02) Tek, Faik Boray; Çam, İlker; Karlı, DenizMany deep neural networks are built by using stacked convolutional layers of fixed and single size (often 3 × 3) kernels. This paper describes a method for learning the size of convolutional kernels to provide varying size kernels in a single layer. The method utilizes a differentiable, and therefore backpropagation-trainable Gaussian envelope which can grow or shrink in a base grid. Our experiments compared the proposed adaptive layers to ordinary convolution layers in a simple two-layer network, a deeper residual network, and a U-Net architecture. The results in the popular image classification datasets such as MNIST, MNIST-CLUTTERED, CIFAR-10, Fashion, and ‘‘Faces in the Wild’’ showed that the adaptive kernels can provide statistically significant improvements on ordinary convolution kernels. A segmentation experiment in the Oxford-Pets dataset demonstrated that replacing ordinary convolution layers in a U-shaped network with 7 × 7 adaptive layers can improve its learning performance and ability to generalize.Yayın Retinal disease diagnosis in OCT scans using a foundational model(Springer Science and Business Media Deutschland GmbH, 2025) Nazlı, Muhammet Serdar; Turkan, Yasemin; Tek, Faik Boray; Toslak, Devrim; Bulut, Mehmet; Arpacı, Fatih; Öcal, Mevlüt CelalThis study examines the feasibility and performance of using single OCT slices from the OCTA-500 dataset to classify DR (Diabetic Retinopathy) and AMD (Age-Related Macular Degeneration) with a pre-trained transformer-based model (RETFound). The experiments revealed the effective adaptation capability of the pretrained model to the retinal disease classification problem. We further explored the impact of using different slices from the OCT volume, assessing the sensitivity of the results to the choice of a single slice (e.g., “middle slice”) and whether analyzing both horizontal and vertical cross-sectional slices could improve outcomes. However, deep neural networks are complex systems that do not indicate directly whether they have learned and generalized the disease appearance as human experts do. The original dataset lacked disease localization annotations. Therefore, we collected new disease classification and localization annotations from independent experts for a subset of OCTA-500 images. We compared RETFound’s explainability-based localization outputs with these newly collected annotations and found that the region attributions aligned well with the expert annotations. Additionally, we assessed the agreement and variability between experts and RETFound in classifying disease conditions. The Kappa values, ranging from 0.35 to 0.69, indicated moderate agreement among experts and between the experts and the model. The transformer-based RETFound model using single or multiple OCT slices, is an efficient approach to diagnosing AMD and DR.Yayın Retinal disease classification from bimodal OCT and OCTA using a CNN-ViT hybrid architecture(Institute of Electrical and Electronics Engineers Inc., 2025-09-21) Aydın, Ömer Faruk; Tek, Faik Boray; Turkan, YaseminRetinal diseases are the leading cause of vision impairment and blindness worldwide. Early and accurate diagnosis is critical for effective treatment, and recent advances in imaging technologies such as Optical Coherence Tomography (OCT) and OCT Angiography (OCTA), have enabled detailed visualization of the retinal structure and vasculature. By leveraging these modalities, this study proposes an advanced deep learning architecture called MultiModalNet for automated multi-class retinal disease classification. MultiModalNet employs a dual-branch design, where OCTA projection maps are processed through a ResNet101 encoder, and cross-sectional slices from the OCT volume (B-scans) are analyzed using a Vision Transformer (ViT-Large). The extracted features from both branches were fused and passed through the fully connected layers for the final classification. Evaluated on the 3-class OCTA-500 dataset, which includes Age-related Macular Degeneration (AMD), Diabetic Retinopathy (DR), and Normal cases, the proposed model achieved state-of-the-art classification accuracy of 94.59 percent, significantly o utperforming single-modality baselines. This result highlights the effectiveness of integrating vascular and structural information to improve the diagnostic performance. The findings suggest that hybrid multi-modal deep learning approaches can play a transformative role in computer-aided ophthalmology, enhancing both clinical decision-making and screening workflows.












