Arama Sonuçları

Listeleniyor 1 - 9 / 9
  • Yayın
    Classification-based macroblock layer rate control for low delay transmission of H.263 video
    (IS & T - Soc Imaging Science Technology, 2003-07) Bayazıt, Uluğ
    Puri and Aravind's method of macroblock bit count estimation for video rate control is based on the classification of the macroblock data into discrete classes and assigning a unique non-linear estimate for each class and quantization parameter pair. This method stands apart from other methods in the literature, since the model of the bit count versus the quantization parameter relation, parameterized by macroblock variance, is a discrete model generated solely from measurements, We extend their technique for low-delay video rate control (tight buffer regulation) in two ways. We propose a strategy of near-uniform quantization parameter assignments to the macroblocks of a frame that can come close to maximizing an objective spatial quality function, such as PSNR, over the entire frame. We also adaptively update the quantization parameter assignments for the yet to be coded macroblocks, after the encoding of each macroblock, to compensate for any errors in the bit count estimation of the encoded macroblock. Our experiments demonstrate that the proposed rate control method can more accurately control the number of bits expended for a frame, as well as yield a higher objective spatial quality than the method adopted by TMN8.
  • Yayın
    Spherical coding algorithm for wavelet image compression
    (IEEE-Inst Electrical Electronics Engineers Inc, 2009-05) Ateş, Hasan Fehmi; Orchard, Michael T.
    In recent literature, there exist many high-performance wavelet coders that use different spatially adaptive coding techniques in order to exploit the spatial energy compaction property of the wavelet transform. Two crucial issues in adaptive methods are the level of flexibility and the coding efficiency achieved while modeling different image regions and allocating bitrate within the wavelet subbands. In this paper, we introduce the "spherical coder," which provides a new adaptive framework for handling these issues in a simple and effective manner. The coder uses local energy as a direct measure to differentiate between parts of the wavelet subband and to decide how to allocate the available bitrate. As local energy becomes available at finer resolutions, i.e., in smaller size windows, the coder automatically updates its decisions about how to spend the bitrate. We use a hierarchical set of variables to specify and code the local energy up to the highest resolution, i.e., the energy of individual wavelet coefficients. The overall scheme is nonredundant, meaning that the subband information is conveyed using this equivalent set of variables without the need for any side parameters. Despite its simplicity, the algorithm produces PSNR results that are competitive with the state-of-art coders in literature.
  • Yayın
    A novel image compression method based on classified energy and pattern building blocks
    (Springer International Publishing AG, 2011) Güz, Ümit
    In this paper, a novel image compression method based on generation of the so-called classified energy and pattern blocks (CEPB) is introduced and evaluation results are presented. The CEPB is constructed using the training images and then located at both the transmitter and receiver sides of the communication system. Then the energy and pattern blocks of input images to be reconstructed are determined by the same way in the construction of the CEPB. This process is also associated with a matching procedure to determine the index numbers of the classified energy and pattern blocks in the CEPB which best represents (matches) the energy and pattern blocks of the input images. Encoding parameters are block scaling coefficient and index numbers of energy and pattern blocks determined for each block of the input images. These parameters are sent from the transmitter part to the receiver part and the classified energy and pattern blocks associated with the index numbers are pulled from the CEPB. Then the input image is reconstructed block by block in the receiver part using a mathematical model that is proposed. Evaluation results show that the method provides considerable image compression ratios and image quality even at low bit rates.
  • Yayın
    Predictive vector quantization of 3-D mesh geometry by representation of vertices in local coordinate systems
    (Elsevier Inc, 2007-08) Bayazıt, Uluğ; Orcay, Özgür; Konur, Umut; Gürgen, Sadık Fikret
    In predictive 3-D mesh geometry coding, the position of each vertex is predicted from the previously coded neighboring vertices and the resultant prediction error vectors are coded. In this work, the prediction error vectors are represented in a local coordinate system in order to cluster them around a subset of a 2-D planar subspace and thereby increase block coding efficiency. Alphabet entropy constrained vector quantization (AECVQ) of Rao and Pearlman is preferred to the previously employed minimum distortion vector quatitization (MDVQ) for block coding the prediction error vectors with high coding efficiency and low implementation complexity. Estimation and compensation of the bias in the parallelogram prediction rule and partial adaptation of the AECVQ codebook to the encoded vector source by normalization using source statistics, are the other salient features of the proposed coding system. Experimental results verify the advantage of the use of the local coordinate system over the global one. The visual error of the proposed coding system is lower than the predictive coding method of Touma and Gotsman especially at low rates, and lower than the spectral coding method of Karni and Gotsman at medium-to-high rates.
  • Yayın
    Compression of the biomedical images using quadtree-based partitioned universally classified energy and pattern blocks
    (Springer London, 2019-03-15) Gezer, Murat; Gargari, Sepideh Nahavandi; Güz, Ümit; Gürkan, Hakan
    In this work, an efficient low bit rate image coding/compression method based on the quadtree-based partitioned universally classified energy and pattern building blocks (QB-UCEPB) is introduced. The proposed method combines low bit rate robustness and variable-sized quantization benefits of the well-known classified energy and pattern blocks (CEPB) method and quadtree-based (QB) partitioning technique, respectively. In the new method, first, the QB-UCEPB is constructed in the form of variable length block size thanks to the quadtree-based partitioning rather than fixed block size partitioning which was employed in the conventional CEPB method. The QB-UCEPB is then placed to the transmitter side as well as receiver side of the communication channel as a universal codebook manner. Every quadtree-based partitioned block of the input image is encoded using three quantities: image block scaling coefficient, the index number of the QB-UCEB and the index number of the QB-UCPB. These quantities are sent from the transmitter part to the receiver part through the communication channel. Then, the quadtree-based partitioned input image blocks are reconstructed in the receiver part using a decoding algorithm, which exploits the mathematical model that is proposed. Experimental results show that using the new method, the computational complexity of the classical CEPB is substantially reduced. Furthermore, higher compression ratios, PSNR and SSIM levels are achieved even at low bit rates compared to the classical CEPB and conventional methods such as SPIHT, EZW and JPEG2000
  • Yayın
    Significance map pruning and other enhancements to SPIHT image coding algorithm
    (Elsevier Science, 2003-10) Bayazıt, Uluğ
    This paper proposes several enhancements to the Set Partitioning in Hierarchical Trees (SPIHT) image coding algorithm without changing the original algorithm's general skeleton. First and foremost, a method for significance map pruning based on a rate-distortion criterion is introduced. Specifically, the (Type A) sets of wavelet coefficients with small ratios of estimated distortion reduction to estimated rate contribution are deemed insignificant and effectively pruned. Even though determining such sets requires the computational complexity of the encoder to increase considerably with respect to the original SPIHT encoder, the original SPIHT decoder may still be used to decode the generated bitstream with a low computational complexity. The paper also proposes three low complexity enhancements by more sophisticated use of the adaptive arithmetic coder. Simulation results demonstrate that all these enhancements yield modest compression gains at moderate to high rates.
  • Yayın
    Hierarchical quantization indexing for wavelet and wavelet packet image coding
    (Elsevier Science BV, 2010-02) Ateş, Hasan Fehmi; Tamer, Engin
    In this paper, we introduce the quantization index hierarchy, which is used for efficient coding of quantized wavelet and wavelet packet coefficients. A hierarchical classification map is defined in each wavelet subband, which describes the quantized data through a series of index classes. Going from bottom to the top of the tree, neighboring coefficients are combined to form classes that represent some statistics of the quantization indices of these coefficients. Higher levels of the tree are constructed iteratively by repeating this class assignment to partition the coefficients into larger Subsets. The class assignments are optimized using a rate-distortion cost analysis. The optimized tree is coded hierarchically from top to bottom by coding the class membership information at each level of the tree. Context-adaptive arithmetic coding is used to improve coding efficiency. The developed algorithm produces PSNR results that are better than the state-of-art wavelet-based and wavelet packet-based coders in literature.
  • Yayın
    Rate-distortion and complexity optimized motion estimation for H.264 video coding
    (IEEE-INST Electrical Electronics Engineers Inc, 2008-02) Ateş, Hasan Fehmi; Altunbaşak, Yücel
    11.264 video coding standard supports several inter-prediction coding modes that use macroblock (MB) partitions with variable block sizes. Rate-distortion (R-D) optimal selection of both the motion vectors (MVs) and the coding mode of each MB is essential for an H.264 encoder to achieve superior coding efficiency. Unfortunately, searching for optimal MVs of each possible subblock incurs a heavy computational cost. In this paper, in order to reduce the computational burden of integer-pel motion estimation (ME) without sacrificing from the coding performance, we propose a R-D and complexity joint optimization framework. Within this framework, we develop a simple method that determines for each MB which partitions are likely to be optimal. MV search is carried out for only the selected partitions, thus reducing the complexity of the ME step. The mode selection criteria is based on a measure of spatiotemporal activity within the MB. The procedure minimizes the coding loss at a given level of computational complexity either for the full video sequence or for each single frame. For the latter case, the algorithm provides a tight upper bound on the worst case complexity/execution time of the ME module. Simulation results show that the algorithm speeds up integer-pel ME by a factor of up to 40 with less than 0.2 dB loss in coding efficiency.
  • Yayın
    Postprocessing of decoded color images by adaptive linear filtering
    (Elsevier Science, 2003-02) Bayazıt, Uluğ
    This paper presents an image adaptive linear filtering method for the reconstruction of the RGB (red, blue, green) color coordinates of a pixel from the lossy compressed luminance/chrominance color coordinates. In the absence of quantization noise, the RGB coordinates of a pixel can be perfectly reconstructed by employing a standard, fixed filter whose support includes only the luminance/chrominance coordinates at the spatial location of the pixel. However, in the presence of quantization noise, a filter with a larger support, that also spatially extends over the luminance/chrominance coordinate planes, is capable of exploiting the statistical dependence among the luminance/chrominance coordinate planes, and thereby yields more accurate reconstruction than the standard, fixed filter. We propose the optimal (in the minimum mean squared error sense) determination of the coefficients of this adaptive linear filter at the image encoder by solving a system of regression equations. When transmitted as side information to the image decoder, the filter coefficients need not incur significant overhead if they are quantized and compressed intelligently. Our simulation results demonstrate that the distortion of the decompressed color coordinate planes can be reduced by several tenths of a dB with negligible overhead rate by the application of our image adaptive linear filtering method.