Arama Sonuçları

Listeleniyor 1 - 10 / 12
  • Yayın
    EEG signal compression based on classified signature and envelope vector sets
    (Wiley, 2009-03) Gürkan, Hakan; Güz, Ümit; Yarman, Bekir Sıddık Binboğa
    In this paper, a novel method to compress electroencephalogram (EEG) signal is proposed. The proposed method is based on the generation process of the classified signature and envelope vector sets (CSEVS), which employs an effective k-means clustering algorithm. It is assumed that both the transmitter and the receiver units have the same CSEVS. In this work, on a frame basis, EEG signals are modeled by multiplying only three factors called as classified signature vector, classified envelope vector, and gain coefficient (GC), respectively. In other words, every frame of an EEG signal is represented by two indices R and K of CSEVS and the GC. EEG signals are reconstructed frame by frame using these numbers in the receiver unit by employing the CSEVS. The proposed method is evaluated by using some evaluation metrics that are commonly used in this area such as root-mean-square error, percentage root-mean-square difference, and measuring with visual inspection. The performance of the proposed method is also compared with the other methods. It is observed that the proposed method achieves high compression ratios with low-level reconstruction error while preserving diagnostic information in the reconstructed EEG signal.
  • Yayın
    Cascaded model adaptation for dialog act segmentation and tagging
    (Elsevier Ltd, 2010-04) Güz, Ümit; Tür, Gökhan; Hakkani Tür, Dilek; Cuendet, Sebastien
    There are many speech and language processing problems which require cascaded classification tasks. While model adaptation has been shown to be useful in isolated speech and language processing tasks, it is not clear what constitutes system adaptation for such complex systems. This paper studies the following questions: In cases where a sequence of classification tasks is employed, how important is to adapt the earlier or latter systems? Is the performance improvement obtained in the earlier stages via adaptation carried on to later stages in cases where the later stages perform adaptation using similar data and/or methods? In this study, as part of a larger scale multiparty meeting understanding system, we analyze various methods for adapting dialog act segmentation and tagging models trained on conversational telephone speech (CTS) to meeting style conversations. We investigate the effect of using adapted and unadapted models for dialog act segmentation with those of tagging, showing the effect of model adaptation for cascaded classification tasks. Our results indicate that we can achieve significantly better dialog act segmentation and tagging by adapting the out-of-domain models, especially when the amount of in-domain data is limited. Experimental results show that it is more effective to adapt the models in the latter classification tasks, in our case dialog act tagging, when dealing with a sequence of cascaded classification tasks
  • Yayın
    Generative and discriminative methods using morphological information for sentence segmentation of Turkish
    (IEEE-INST Electrical Electronics Engineers Inc, 2009-07) Güz, Ümit; Favre, Benoit; Hakkani Tür, Dilek; Tür, Gökhan
    This paper presents novel methods for generative, discriminative, and hybrid sequence classification for segmentation of Turkish word sequences into sentences. In the literature, this task is generally solved using statistical models that take advantage of lexical information among others. However, Turkish has a productive morphology that generates a very large vocabulary, making the task much harder. In this paper, we introduce a new set of morphological features, extracted from words and their morphological analyses. We also extend the established method of hidden event language modeling (HELM) to factored hidden event language modeling (fHELM) to handle morphological information. In order to capture non-lexical information, we extract a set of prosodic features, which are mainly motivated from our previous work for other languages. We then employ discriminative classification techniques, boosting and conditional random fields (CRFs), combined with fHELM, for the task of Turkish sentence segmentation.
  • Yayın
    Effective semi-supervised learning strategies for automatic sentence segmentation
    (Elsevier Science BV, 2018-04-01) Dalva, Doğan; Güz, Ümit; Gürkan, Hakan
    The primary objective of sentence segmentation process is to determine the sentence boundaries of a stream of words output by the automatic speech recognizers. Statistical methods developed for sentence segmentation requires a significant amount of labeled data which is time-consuming, labor intensive and expensive. In this work, we propose new multi-view semi-supervised learning strategies for sentence boundary classification problem using lexical, prosodic, and morphological information. The aim is to find effective semi-supervised machine learning strategies when only small sets of sentence boundary labeled data are available. We primarily investigate two semi-supervised learning approaches, called self-training and co-training. Different example selection strategies were also used for co-training, namely, agreement, disagreement and self-combined. Furthermore, we propose three-view and committee-based algorithms incorporating with agreement, disagreement and self-combined strategies using three disjoint feature sets. We present comparative results of different learning strategies on the sentence segmentation task. The experimental results show that the sentence segmentation performance can be highly improved using multi-view learning strategies that we proposed since data sets can be represented by three redundantly sufficient and disjoint feature sets. We show that the proposed strategies substantially improve the average baseline F-measure of 67.66% to 75.15% and 64.84% to 66.32% when only a small set of manually labeled data is available for Turkish and English spoken languages, respectively.
  • Yayın
    Biometric identification using fingertip electrocardiogram signals
    (Springer London Ltd, 2018-07) Güven, Gökhan; Gürkan, Hakan; Güz, Ümit
    In this research work, we present a newly fingertip electrocardiogram (ECG) data acquisition device capable of recording the lead-1 ECG signal through the right- and left-hand thumb fingers. The proposed device is high-sensitive, dry-contact, portable, user-friendly, inexpensive, and does not require using conventional components which are cumbersome and irritating such as wet adhesive Ag/AgCl electrodes. One of the other advantages of this device is to make it possible to record and use the lead-1 ECG signal easily in any condition and anywhere incorporating with any platform to use for advanced applications such as biometric recognition and clinical diagnostics. Furthermore, we proposed a biometric identification method based on combining autocorrelation and discrete cosine transform-based features, cepstral features, and QRS beat information. The proposed method was evaluated on three fingertip ECG signal databases recorded by utilizing the proposed device. The experimental results demonstrate that the proposed biometric identification method achieves person recognition rate values of 100% (30 out of 30), 100% (45 out of 45), and 98.33% (59 out of 60) for 30, 45, and 60 subjects, respectively.
  • Yayın
    A novel biometric identification system based on fingertip electrocardiogram and speech signals
    (Elsevier Inc., 2022-03) Güven, Gökhan; Güz, Ümit; Gürkan, Hakan
    In this research work, we propose a one-dimensional Convolutional Neural Network (CNN) based biometric identification system that combines speech and ECG modalities. The aim is to find an effective identification strategy while enhancing both the confidence and the performance of the system. In our first approach, we have developed a voting-based ECG and speech fusion system to improve the overall performance compared to the conventional methods. In the second approach, we have developed a robust rejection algorithm to prevent unauthorized access to the fusion system. We also presented a newly developed ECG spike and inconsistent beats removal algorithm to detect and eliminate the problems caused by portable fingertip ECG devices and patient movements. Furthermore, we have achieved a system that can work with only one authorized user by adding a Universal Background Model to our algorithm. In the first approach, the proposed fusion system achieved a 100% accuracy rate for 90 people by taking the average of 3-fold cross-validation. In the second approach, by using 90 people as genuine classes and 26 people as imposter classes, the proposed system achieved 92% accuracy in identifying genuine classes and 96% accuracy in rejecting imposter classes.
  • Yayın
    A novel image compression method based on classified energy and pattern building blocks
    (Springer International Publishing AG, 2011) Güz, Ümit
    In this paper, a novel image compression method based on generation of the so-called classified energy and pattern blocks (CEPB) is introduced and evaluation results are presented. The CEPB is constructed using the training images and then located at both the transmitter and receiver sides of the communication system. Then the energy and pattern blocks of input images to be reconstructed are determined by the same way in the construction of the CEPB. This process is also associated with a matching procedure to determine the index numbers of the classified energy and pattern blocks in the CEPB which best represents (matches) the energy and pattern blocks of the input images. Encoding parameters are block scaling coefficient and index numbers of energy and pattern blocks determined for each block of the input images. These parameters are sent from the transmitter part to the receiver part and the classified energy and pattern blocks associated with the index numbers are pulled from the CEPB. Then the input image is reconstructed block by block in the receiver part using a mathematical model that is proposed. Evaluation results show that the method provides considerable image compression ratios and image quality even at low bit rates.
  • Yayın
    A new method to represent speech signals via predefined signature and envelope sequences
    (Hindawi Publishing Corporation, 2007) Güz, Ümit; Gürkan, Hakan; Yarman, Bekir Sıddık Binboğa
    A novel systematic procedure referred to as "SYMPES" to model speech signals is introduced. The structure of SYMPES is based on the creation of the so-called predefined "signature S = {S(R)(n)} and envelope E = {E(K) (n)}" sets. These sets are speaker and language independent. Once the speech signals are divided into frames with selected lengths, then each frame sequence X(i)( n) is reconstructed by means of the mathematical form X(i)( n) = C(i)E(K) (n) S(R)(n). In this representation, C(i) is called the gain factor, S(R)(n) and E(K) (n) are properly assigned from the predefined signature and envelope sets, respectively. Examples are given to exhibit the implementation of SYMPES. It is shown that for the same compression ratio or better, SYMPES yields considerably better speech quality over the commercially available coders such as G. 726 (ADPCM) at 16 kbps and voice excited LPC-10E (FS1015) at 2.4 kbps.
  • Yayın
    Compression of the biomedical images using quadtree-based partitioned universally classified energy and pattern blocks
    (Springer London, 2019-03-15) Gezer, Murat; Gargari, Sepideh Nahavandi; Güz, Ümit; Gürkan, Hakan
    In this work, an efficient low bit rate image coding/compression method based on the quadtree-based partitioned universally classified energy and pattern building blocks (QB-UCEPB) is introduced. The proposed method combines low bit rate robustness and variable-sized quantization benefits of the well-known classified energy and pattern blocks (CEPB) method and quadtree-based (QB) partitioning technique, respectively. In the new method, first, the QB-UCEPB is constructed in the form of variable length block size thanks to the quadtree-based partitioning rather than fixed block size partitioning which was employed in the conventional CEPB method. The QB-UCEPB is then placed to the transmitter side as well as receiver side of the communication channel as a universal codebook manner. Every quadtree-based partitioned block of the input image is encoded using three quantities: image block scaling coefficient, the index number of the QB-UCEB and the index number of the QB-UCPB. These quantities are sent from the transmitter part to the receiver part through the communication channel. Then, the quadtree-based partitioned input image blocks are reconstructed in the receiver part using a decoding algorithm, which exploits the mathematical model that is proposed. Experimental results show that using the new method, the computational complexity of the classical CEPB is substantially reduced. Furthermore, higher compression ratios, PSNR and SSIM levels are achieved even at low bit rates compared to the classical CEPB and conventional methods such as SPIHT, EZW and JPEG2000
  • Yayın
    Modeling of electrocardiogram signals using predefined signature and envelope vector sets
    (Hindawi Publishing Corporation, 2007) Gürkan, Hakan; Güz, Ümit; Yarman, Bekir Sıddık Binboğa
    A novel method is proposed to model ECG signals by means of "predefined signature and envelope vector sets (PSEVS)." On a frame basis, an ECG signal is reconstructed by multiplying three model parameters, namely, predefined signature vector (PSV)(R)," "predefined envelope vector (PEV)(K)," and frame-scaling coefficient (FSC). All the PSVs and PEVs are labeled and stored in their respective sets to describe the signal in the reconstruction process. In this case, an ECG signal frame is modeled by means of the members of these sets labeled with indices R and K and the frame-scaling coefficient, in the least mean square sense. The proposed method is assessed through the use of percentage root-mean-square difference (PRD) and visual inspection measures. Assessment results reveal that the proposed method provides significant data compression ratio (CR) with low-level PRD values while preserving diagnostic information. This fact significantly reduces the bandwidth of communication in telediagnosis operations. Copyright (c) 2007 Hakan Gurkan et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.