Arama Sonuçları

Listeleniyor 1 - 3 / 3
  • Yayın
    Anatomy based animation of facial expressions
    (Işık Üniversitesi, 2012-12-24) Erkoç, Tuğba; Eskil, Mustafa Taner; Işık Üniversitesi, Fen Bilimleri Enstitüsü, Bilgisayar Mühendisliği Yüksek Lisans Programı
    This study presents a new physics based facial expression animation system based on anatomic data. Proposed system consists a face model and a new facial expression animation generation algorithm. The proposed system is a Mass-Spring-Damper (MSD) system. It consists single layer face model HIGEM and a set of facial muscles which are placed anatomically correct places. Non-linear viscoelastic characteristics of the human skin is approximated with non-linear springs. The set of muscles triggers facial expressions. HIGEM does not include a skull to support the facial mesh, so it tends to collapse under muscle forces. A new algorithm proposed to prevent this. This algorithm uses back and forward projections for re-defining the muscle forces. Another problem of MSD systems is individual element collapse under large forces. This is addressed with Edge Repulsion (ER) approach. Dynamics of the face is modelled with a quasi-implicit Ordinary Differential Equation (ODE). The regions of the face affected by the facial muscles were computed and marked offline in previous works to speed up the animations. It requires offline re-calculation whenever the activation level of the muscles or facial mesh topology changes. Thus, a new generic stack based approach which works at runtime is proposed. The proposed system is tested with eight facial expressions; happy, sad, angry, feared, surprised, disgusted, disgusted anger and happy surprise. Elapsed times for the facial expressions vary with number of contracting muscles, their region of influence, and the time value for each timestep. Keywords: Facial expression animation, physics based approach, generic wireframe, mass-spring-damper system.
  • Yayın
    A novel similarity based unsupervised technique for training convolutional filters
    (IEEE, 2023-05-17) Erkoç, Tuğba; Eskil, Mustata Taner
    Achieving satisfactory results with Convolutional Neural Networks (CNNs) depends on how effectively the filters are trained. Conventionally, an appropriate number of filters is carefully selected, the filters are initialized with a proper initialization method and trained with backpropagation over several epochs. This training scheme requires a large labeled dataset, which is costly and time-consuming to obtain. In this study, we propose an unsupervised approach that extracts convolutional filters from a given dataset in a self-organized manner by processing the training set only once without using backpropagation training. The proposed method allows for the extraction of filters from a given dataset in the absence of labels. In contrast to previous studies, we no longer need to select the best number of filters and a suitable filter weight initialization scheme. Applying this method to the MNIST, EMNIST-Digits, Kuzushiji-MNIST, and Fashion-MNIST datasets yields high test performances of 99.19%, 99.39%, 95.03%, and 90.11%, respectively, without applying backpropagation training or using any preprocessed and augmented data.
  • Yayın
    Object recognition with competitive convolutional neural networks
    (Işık Üniversitesi, 2023-06-12) Erkoç, Tuğba; Eskil, M. Taner; Işık Üniversitesi, Lisansüstü Eğitim Enstitüsü, Bilgisayar Mühendisliği Doktora Programı; Işık University, School of Graduate Studies, Ph.D. in Computer Engineering
    In recent years, Artificial Intelligence (AI) has achieved impressive results, often surpassing human capabilities in tasks involving language comprehension and visual recognition. Among these, computer vision has experienced remarkable progress, largely due to the introduction of Convolutional Neural Networks (CNNs). CNNs are inspired by the hierarchical structure of the visual cortex and are designed to detect patterns, objects, and complex relationships within visual data. One key advantage is their ability to learn directly from pixel values without the need for domain expertise, which has contributed to their popularity. These networks are trained using supervised backpropagation, a process that calculates gradients of the network’s parameters (weights and biases) with respect to the loss function. While backpropagation enables impressive performance with CNNs, it also presents certain drawbacks. One such drawback is the requirement for large amounts of labeled data. When the available data samples are limited, the gradients estimated from this limited information may not accurately capture the overall data behavior, leading to suboptimal parameter updates. However, obtaining a sufficient quantity of labeled data poses a challenge. Another drawback is the requirement of careful configuration of hyperparameters, including the number of neurons, learning rate, and network architecture. Finding optimal values for these hyperparameters can be a time-consuming process. Furthermore, as the complexity of the task increases, the network architecture becomes deeper and more complex. To effectively train the shallow layers of the network, one must increase the number of epochs and experiment with solutions to prevent vanishing gradients. Complex problems often require a greater number of epochs to learn the intricate patterns and features present in the data. It’s important to note that while CNNs aim to mimic the structure of the visual cortex, the brain’s learning mechanism does not necessarily involve back-propagation. Although CNNs incorporate the layered architecture of the visual cortex, the reliance on backpropagation introduces an artificial learning procedure that may not align with the brain’s actual learning process. Therefore, it is crucial to explore alternative learning paradigms that do not rely on backpropagation. In this dissertation study, a unique approach to unsupervised training for CNNs is explored, setting it apart from previous research. Unlike other unsupervised methods, the proposed approach eliminates the reliance on backpropagation for training the filters. Instead, we introduce a filter extraction algorithm capable of extracting dataset features by processing images only once, without requiring data labels or backward error updates. This approach operates on individual convolutional layers, gradually constructing them by discovering filters. To evaluate the effectiveness of this backpropagation-free algorithm, we design four distinct CNN architectures and conduct experiments. The results demonstrate the promising performance of training without backpropagation, achieving impressive classification accuracies on different datasets. Notably, these outcomes are attained using a single network setup without any data augmentation. Additionally, our study reveals that the proposed algorithm eliminates the need to predefine the number of filters per convolutional layer, as the algorithm automatically determines this value. Furthermore, we demonstrate that filter initialization from a random distribution is unnecessary when backpropagation is not employed during training.