Arama Sonuçları

Listeleniyor 1 - 3 / 3
  • Yayın
    Malaria parasite detection with deep transfer learning
    (IEEE, 2018-12-06) Var, Esra; Tek, Faik Boray
    This study aims to automatically detect malaria parasites (Plasmodium sp) on images taken from Giemsa stained blood smears. Deep learning methods provide limited performance when sample size is low. In transfer learning, visual features are learned from large general data sets, and problem-specific classification problem can be solved successfully in restricted problem specific data sets. In this study, we apply transfer learning method to detect and classify malaria parasites. We use a popular pre-trained CNN model VGG19. We trained the model for 20 epoch on 1428 P Vivax, 1425 P Ovule, 1446 E Falciparum, 1450 P Malariae and 1440 non-parasite samples. The transfer learning model achieves %80, %83, %86, %75 precision and 83%, 86%, 86%, 79% f-measure on 19 test images.
  • Yayın
    Animal sound classification using a convolutional neural network
    (IEEE, 2018-12-06) Şaşmaz, Emre; Tek, Faik Boray
    In this paper, we investigate the problem of animal sound classification using deep learning and propose a system based on convolutional neural network architecture. As the input to the network, sound files were preprocessed to extract Mel Frequency Cepstral Coefficients (MFCC) using LibROSA library. To train and test the system we have collected 875 animal sound samples from an online sound source site for 10 different animal types. We report classification confusion matrices and the results obtained by different gradient descent optimizers. The best accuracy of 75% was obtained by Nesterov-accelerated Adaptive Moment Estimation (Nadam).
  • Yayın
    Retinal disease classification from bimodal OCT and OCTA using a CNN-ViT hybrid architecture
    (Institute of Electrical and Electronics Engineers Inc., 2025-09-21) Aydın, Ömer Faruk; Tek, Faik Boray; Turkan, Yasemin
    Retinal diseases are the leading cause of vision impairment and blindness worldwide. Early and accurate diagnosis is critical for effective treatment, and recent advances in imaging technologies such as Optical Coherence Tomography (OCT) and OCT Angiography (OCTA), have enabled detailed visualization of the retinal structure and vasculature. By leveraging these modalities, this study proposes an advanced deep learning architecture called MultiModalNet for automated multi-class retinal disease classification. MultiModalNet employs a dual-branch design, where OCTA projection maps are processed through a ResNet101 encoder, and cross-sectional slices from the OCT volume (B-scans) are analyzed using a Vision Transformer (ViT-Large). The extracted features from both branches were fused and passed through the fully connected layers for the final classification. Evaluated on the 3-class OCTA-500 dataset, which includes Age-related Macular Degeneration (AMD), Diabetic Retinopathy (DR), and Normal cases, the proposed model achieved state-of-the-art classification accuracy of 94.59 percent, significantly o utperforming single-modality baselines. This result highlights the effectiveness of integrating vascular and structural information to improve the diagnostic performance. The findings suggest that hybrid multi-modal deep learning approaches can play a transformative role in computer-aided ophthalmology, enhancing both clinical decision-making and screening workflows.