(Back to Session Schedule)

28th Picture Coding Symposium

Session P4  Poster Session 4
Time: 16:30 - 18:00 Friday, December 10, 2010
Chair: Shinichi Sakaida (Science & Technology Research Laboratories, NHK, Japan)


[3DTV/FTV/multi-view-related topics]

P4-1
TitleOn-Line Statistical Analysis Based Fast Mode Decision for Multi-View Video Coding
AuthorChia-Chi Chan (Dept. Communication Engineering, National Central University, Taiwan), Jheng-Ping Lin (ZyXEL Corp., Taiwan), Chih-Wei Tang (Dept. Communication Engineering, National Central University, Taiwan)
Pagepp. 478 - 481
KeywordMulti-view video coding, fast mode decision, statistical analysis, RD cost, motion and disparity estimation
AbstractThe high computational complexity of multi-view video codecs makes it necessary to speed up for their realization in consumer electronics. Since fast encoding algorithms are expected to adapt to different video sequences, this paper proposes a fast algorithm that consists of fast mode decision and fast disparity estimation for multi-view video coding. The fast mode decision algorithm applies to both temporal and inter-view predictions. The candidates for mode decision are reduced based on a set of thresholds. Differ from the previous fast mode decision algorithms for MVC, this scheme determines the thresholds according to the on-line statistical analysis of motion and disparity costs of the first GOP in each view. Since the inter-view prediction is time consuming, we propose a fast disparity estimation algorithm to save encoding time. Experimental results show that our proposed scheme reduces the computational complexity significantly with negligible degradation of coding efficiency.

P4-2
TitleOptimal Rate Allocation for View Synthesis along a Continuous Viewpoint Location in Multiview Imaging
AuthorVladan Velisavljevic (Deutsche Telekom Laboratories, Germany), Gene Cheung (National Institute of Informatics, Japan), Jacob Chakareski (Ecole Polytechnique Federale de Lausanne, Switzerland)
Pagepp. 482 - 485
KeywordMulti-view imaging, Rate allocation
AbstractWe consider the scenario of view synthesis via depth-image based rendering in multi-view imaging. We formulate a resource allocation problem of jointly assigning an optimal number of bits to compressed texture and depth images such that the maximum distortion of a synthesized view over a continuum of viewpoints between two encoded reference views is minimized, for a given bit budget. We construct simple yet accurate image models that characterize the pixel values at similar depths as first-order Gaussian auto-regressive processes. Based on our models, we derive an optimization procedure that numerically solves the formulated min-max problem using Lagrange relaxation. Through simulations we show that, for two captured views scenario, our optimization provides a significant gain (up to 2dB) in quality of the synthesized views for the same overall bit rate over a heuristic quantization that selects only two quantizers - one for the encoded texture images and the other for the depth images.

P4-3
TitlePanoramic Scene Generation from Multi-view Images with Close Foreground Objects
AuthorSoon-Young Lee (Seoul National University, Republic of Korea), Jae-Young Sim (Ulsan National Institute of Science and Technology, Republic of Korea), Chang-Su Kim (Korea University, Republic of Korea), Sang-Uk Lee (Seoul National University, Republic of Korea)
Pagepp. 486 - 489
Keyword3D scene representation, Image panorama, view expansion, multi-view image processing
AbstractAn algorithm to generate a panorama from multi-view images, which contain foreground objects with varying depths, is proposed in this work. The proposed algorithm constructs a foreground panorama and a background panorama separately, and then merges them into a complete panorama. First, the foreground panorama is obtained by finding the translational displacements of objects between source images. Second, the background panorama is initialized using warped source images and then optimized to preserve spatial consistency and satisfy visual constraints. Then, the background panorama is extended by inserting seams and merged with the foreground panorama. Experimental results demonstrate that the proposed algorithm provides visually satisfying panoramas with all meaningful foreground objects, but without severe artifacts in the backgrounds.

P4-4
TitleA Sub-Pixel Virtual View Synthesis Method for Multiple View Synthesis
AuthorXin Tong, Ping Yang (Tsinghua University, China), Xiaozhen Zheng, Jianhua Zheng (Hisilicon Technologies Co. Ltd., China), Yun He (Tsinghua University, China)
Pagepp. 490 - 493
KeywordSPVVS, virtual view synthesis
AbstractA sub-pixel virtual view synthesis (SPVVS) method is proposed in this paper. In the proposed method, by promoting the sampling rate of the target virtual view, the sub-pixel information of the corresponding pixel position among multiple input views with sub-pixel displacement can be utilized. A directional adaptive image interpolation is manipulated to generate a high resolution intermediate image, which is then down-sampled to obtain the target synthesized virtual view. The realization procedure of SPVVS is also presented. Experiment result shows significant improvements on subjective quality compared to traditional integer pixel synthesis method. Artifacts such as ‘hat’ effects can be significantly reduced.

P4-5
TitleImproving the Visual Quality of AVC/H.264 by Combining It with Content Adaptive Depth Map Compression
AuthorChristian Keimel, Klaus Diepold (TU Muenchen, Germany), Michel Sarkis (Sony Deutschland GmbH, Germany)
Pagepp. 494 - 497
KeywordAVC/H.264, 3D scene analysis, 3DTV, Depth map comression, content adaptive meshing
AbstractThe future of video coding for 3DTV lies in the combination of depth maps and corresponding textures. Most current video coding standards, however, are only optimized for visual quality and are not able to efficiently compress depth maps. We present in this work a content adaptive depth map meshing with tritree and entropy encoding for 3D videos. We show that this approach outperforms the intra frame prediction of AVC/H.264 for the coding of depth maps of still images. We also demonstrate by combining AVC/H.264 with our algorithm that we are able to increase the visual quality of the encoded texture on average by 6~dB. This work is currently limited to still images but an extension to intra coding of 3D video is straightforward.

P4-6
TitleError Concealment for MVC and 3D Video Coding
AuthorOlgierd Stankiewicz, Krzysztof Wegner, Marek Domański (Poznan University of Technology, Poland)
Pagepp. 498 - 501
KeywordError concealment, MVC, 3D video, depth maps, cross-checking
AbstractIn this paper we propose a novel approach to error concealment that can be applied to MVC and other 3D video coding technologies. The image content, that is lost due to errors, is recovered with use of multiple error-concealment techniques. In our work we have used three techniques: well-known temporal- and intra-based techniques and a novel inter-view technique. Proposed inter-view recovery employs Depth Image Based Rendering (DIBR), which requires neighboring views and corresponding depth maps. Those depth maps can be delivered in the bit-stream or estimated in the receiver. In order to obtain the final reconstruction, the best technique is selected locally. For that, an original recovery quality measurement method, based on cross-checking, has been proposed. The idea has been implemented and assessed experimentally, with use of 3D video test sequences. The objective and subjective results show that the proposed approach provide good quality of reconstructed video.

P4-7
TitleDifference Detection Based Early Mode Termination for Depth Map Coding in MVC
AuthorMinghui Wang, Xin Jin, Satoshi Goto (Waseda University, Japan)
Pagepp. 502 - 505
KeywordMVC, mode decision, depth map, difference detection
AbstractDepth map coding is a new topic in multiview video coding (MVC) following the development of depth-image-based rendering (DIBR). Since depth map is monochromatic and has less texture than color map, fast algorithm is necessary and possible to reduce the computation burden of the encoder. This paper proposed a difference detection based early mode termination strategy. The difference detection (DD) algorithms are categorized to reconstructed frame based (RDD) and original frame based.(ODD). And a simplified ODD (sODD) is also proposed. Early mode termination based on these three DD algorithms are implemented and evaluated in the reference software of JMVC8.0 respectively. Simulation results indicate that RDD based one has no performance lost and reduce 25% runtime on average. ODD and sODD based ones can save 54.3% and 43.6% runtime respectively and have a acceptable R-D performance lost.

P4-8
TitleFast Stereo Matching with Predictive Search Range
AuthorYu-Cheng Tseng, Po-Hsiung Hsu, Tian-Sheuan Chang (Dept. of Electronics Engineering & Institute of Electronics, National Chiao Tung University, Taiwan)
Pagepp. 506 - 509
Keyworddisparity estimation, stereo matching
AbstractLocal stereo matching could deliver accurate disparity maps by the associated method, like adaptive support-weight, but suffers from the high computational complexity, O(NL), where N is pixel count in spatial domain, and L is search range in disparity domain. This paper proposes a fast algorithm that groups similar pixels into super-pixels for spatial reduction, and predicts their search range by simple matching for disparity reduction. The proposed algorithm could be directly applied to other local stereo matching, and reduce its computational complexity to only 8.2%-17.4% with slight 1.5%-3.2% of accuracy degradation.

P4-9
TitleInfluences of Frame Delay and Packet Loss between Left and Right Frames in Stereoscopic Video Communications
AuthorShuliang Lin, Yuichiro Sawa, Norishige Fukushima, Yutaka Ishibashi (Nagoya Institute of Technology, Japan)
Pagepp. 510 - 513
Keywordstereo video, network delay, packet loss, subjective assessment, inter-media synchronization
AbstractThis paper analyzes the influences of frame delay and packet loss in stereoscopic vision when stereoscopic video transferred over a IP network. We employ live action videos which are transferred to a head-mount-display (HMD) and do the assessment on stereoscopic perception. As a result, we found that speed and movement direction of the attention object play a great role on the deterioration when frame delay and packet loss occurs.


[Beyond H.264/MPEG-4 AVC and related topics]

P4-10
TitleA Novel Inloop Filter for Video-Compression Based on Temporal Pixel Trajectories
AuthorMarko Esche, Andreas Krutz, Alexander Glantz, Thomas Sikora (Technische Universität Berlin, Germany)
Pagepp. 514 - 517
Keywordvideo-compression, temporal inloop filter, pixel trajectories, deblocking
AbstractThe objective of this work is to investigate the performance of a new inloop filter for video compression, which uses temporal rather than spatial information to improve the quality of reference frames used for prediction. The new filter has been integrated into the H.264/AVC baseline encoder and tested on a wide range of sequences. Experimental results show that the filter achieves a bit rate reduction of up to 12% and more than 4% on average without increasing the complexity of either encoder or decoder significantly.

P4-11
TitleFast Rate-Distortion Optimized Transform for Intra Coding
AuthorXin Zhao (Institute of Computing Technology, Chinese Academy of Sciences; Graduate University of Chinese Academy of Sciences, China), Li Zhang, Siwei Ma, Wen Gao (Institute of Digital Media, Peking University, China)
Pagepp. 518 - 521
Keywordrate-distortion optimization (RDO), Intra coding, mode-dependent directional transform (MDDT), rate-distortion optimized transform (RDOT)
AbstractIn our previous work, rate-distortion optimized transform (RDOT) is introduced. The proposed RDOT achieves remarkable coding gain for KTA Intra coding, but the computational complexity is increased drastically at the encoder. To solve this problem, we propose a fast RDOT scheme using macroblock- and block-level R-D cost thresholding. With the proposed methods, unnecessary mode trials can be efficiently skipped from the encoding process. Experimental results show that, with negligible performance degradation, about 88.9% of the total encoding time is saved.

P4-12
TitleA Hierarchical Variable-Sized Block Transform Coding Scheme for Coding Efficiency Improvement on H.264/AVC
AuthorBumshik Lee, Jaeil Kim, Sangsoo Ahn, Munchurl Kim (Korea Advanced Institute of Science and Technology, Republic of Korea), Hui Yong Kim, Jongho Kim, Jin Soo Choi (Electronic Telecommunications Research Institute (ETRI), Republic of Korea)
Pagepp. 522 - 525
KeywordQuadtree Transform, Variable Block-size Transform, Discrete Cosine Transform, H.264/AVC
AbstractIn this paper, a rate-distortion optimized variable block transform coding scheme is proposed based on a quadtree-structured transform for macroblock (MB) coding with a set of the order-4 and -8 integer cosine transform (ICT) kernels of H.264/AVC as well as a new order-16 ICT kernel. The set of order-4, -8 and -16 ICT kernels are applied for inter-predictive coding in square (4x4, 8x8 or 16x16) or non-square (16x8 or 8x16) transform for each MB in a quadtree structured manner. The proposed quadtree-structured variable block transform scheme using the order-16 ICT kernel achieves significant bitrate reduction up to 15%, compared to the High profile of H.264/AVC. Even if the number of candidates for the transform types increases, the encoding time can be reduced to average 4-6% over the H.264/AVC.

P4-13
TitleEnhanced Region-Based Adaptive Interpolation Filter
AuthorShohei Matsuo, Yukihiro Bandoh, Seishi Takamura, Hirohisa Jozawa (NTT Cyber Space Laboratories, NTT Corporation, Japan)
Pagepp. 526 - 529
Keywordmotion compensation, adaptive interpolation filter, region-division, edge, image locality
AbstractAdaptive interpolation filter (AIF) was proposed to improve motion compensation. The conventional AIF optimizes the filter coefficients on a frame-by-frame basis. However, when the image is divided into multiple regions, each of which has different characteristics, the coding efficiency could improve by performing optimization on a region-by-region basis. In this paper, we proposed a region-based AIF (RBAIF). Simulation results showed that RBAIF offered about 0.43% and 5.05% higher coding gain than the conventional AIF and the H.264/AVC filter, respectively.

P4-14
TitleFractional-Sample Motion Compensation Using Generalized Interpolation
AuthorHaricharan Lakshman, Benjamin Bross, Heiko Schwarz, Thomas Wiegand (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany)
Pagepp. 530 - 533
Keywordvideo coding, motion-compensated prediction, reference picture upsampling, B-splines
AbstractTypical interpolation methods in video coding perform filtering of reference picture samples using FIR filters for motion compensated prediction. This process can be viewed as a signal decomposition using basis functions which are restricted by the interpolating constraint. Using the concept of generalized interpolation provides a greater degree of freedom for selecting basis functions. We implemented generalized interpolation using a combination of IIR and FIR filters. The complexity of the proposed scheme is comparable to that of an 8-tap FIR filter. Bit rate savings up to 20% compared to the H.264/AVC 6-tap filter are shown.


[Image/video coding and related topics]

P4-15
TitleImage Coding Approach Based On Image Decomposition
AuthorYunhui Shi, Yanli Hou, Baocai Yin, Wenpeng Ding (Beijing University of Technology, China)
Pagepp. 534 - 537
Keywordimage decomposition, texture synthesis, region selection, image coding
AbstractTextures in many images or video scenes are difficult to code because of the large amount of visible detail. This paper proposes an image coding approach to solve this problem, in which we incorporate image decomposition and texture synthesis technology into the image coding framework. The key idea of our approach is to first decompose the original image into cartoon component u and texture component v with different basic characteristics, and then to synthesize the selected texture regions in texture component v. The cartoon component u and the non-synthetic regions in texture component v are compressed by JPEG. Experimental results show bit-rate savings of over 30% compared with JPEG at similar visual quality levels.

P4-16
TitleA Real-Time System of Distributed Video Coding
AuthorKazuhito Sakomizu, Takahiro Yamasaki, Satoshi Nakagawa, Takashi Nishi (Oki Electric Industry Co., Ltd., Japan)
Pagepp. 538 - 541
Keyworddistributed video coding, real-time system, Slepian-Wolf theorem, Wyner-Ziv theorem
AbstractThis paper presents a real-time system of distributed video coding (DVC). The decoding process of DVC is normally complex, which causes difficulty in real-time implementation. To address this problem, we propose a new configuration of DVC with three methods. Then we implement the system with parallelization techniques. Experimental results show that the encoder on i.MX31 400 MHz could operates at about CIF 13 fps, and the decoder on Core 2 Quad 2.83 GHz operates at more than CIF 30 fps.

P4-17
TitleBlock-Based Second Order Prediction on AVS-Part 2
AuthorBinbin Yu, Shangwen Li, Lu Yu (Institute of Information and Communication Engineering, Zhejiang University, China)
Pagepp. 542 - 545
KeywordBlock-based Second Order Prediction, Motion-compensated prediction, Mode prediction, Directional operators, AVS
AbstractAVS-Part 2 is a mainstream video coding standard with high compression efficiency similar to H.264/AVC. A technique named Second Order Prediction (SOP) has been presented based on H.264/AVC to decrease the signal correlation after motion-compensated prediction. To achieve better coding performance, this paper presents a method named Block-based Second Order Prediction (BSOP) to ameliorate SOP to adapt to the features of the motion-compensation in AVS-P2 with analysis and demonstration in detail. Experimental results show that the proposed BSOP can outperform AVS-P2 P-picture coding by 3.99% bit-rate saving (0.126dB BD-PSNR gain) on average, and performs better than SOP implemented on AVS by 1.81% bit-rate saving.

P4-18
TitleImproved Local PDF Estimation in the Wavelet Domain for Generalized Lifting
AuthorJulio C. Rolon (National Polytechnic Institute, Mexico), Philippe Salembier (Technical University of Catalonia, Spain)
Pagepp. 546 - 549
KeywordGeneralized lifting, wavelets, pdf estimation, lossy image coding
AbstractGeneralized Lifting has been studied for lossy image compression in [2,3]. It has been demonstrated that the method achieves a significant reduction of the wavelet coefficients energy and entropy. The definition of the GL relies on an estimation of the pdf of the pixel to encode conditioned to a surrounding context. The objective of this paper is to present an improved method for the estimation of the pdf at the local level. We follow the idea of self similarity proposed in [1] for denoising, and propose to estimate the pdf using all the causal contexts within a window. Experimental results show an important increment in the energy and entropy gains when compared to previous strategies [2,3].

P4-19
TitleImage Coding by Using Non-Linear Texture Decomposition and Image Summarization
AuthorChihiro Suzuki, Takamichi Miyata, Yoshinori Sakai (Tokyo Institute of Technology, Japan)
Pagepp. 550 - 553
KeywordTexture Synthesis, TV-regularization, Bidirectional Similarity, Image Coding
AbstractTV-regularization can be used to decompose any natural image into a structure image(S) and a texture image(T). We proposed a novel image coding method that is coding separately these images. We make a compaction image that has all texture patterns. Then the compaction image is divided into S and T. Then, the encoder sends the compaction’s S, T and input’s S to the decoder. At the decoder, original-size T is synthesized from compaction’s S by matching the compaction’s S and the original-size S.

P4-20
TitleCoding Efficiency Improvement Using Inter-Picture Processing of Quantization Error
AuthorKenji Sugiyama, Naoya Sagara, Masao Arizumi (Seikei University, Japan)
Pagepp. 554 - 557
KeywordCompatibility, I-picture coding, B-picture coding, Motion compensation, Quantization error
AbstractAs standard video encoder techniques have matured, their rate of improvement has slowed. As an alternative, a new coding concept with semi-compatibility has been proposed, and an enhancement to I-pictures efficiency has been discussed. This applied method reduces the quantization error using motion compensated inter-picture processing. In this report, we apply this method to P-pictures to improve the efficiency of B-pictures. The quantization error component of the prediction signal is canceled by averaging the bi-directional prediction. Experiments using MPEG-4 show significant improvement in the coding efficiency using the proposed method. The maximum PSNR gain reaches 2.3 dB in a static sequence. At least 0.5 dB can be achieved in a high motion sequence.

P4-21
TitleMultiple Description Video Transcoding with Temporal Drift Control
AuthorPedro Correia, Pedro Assunção, Vitor Silva (Instituto de Telecomunicações, Portugal)
Pagepp. 558 - 561
KeywordMultiple Description Transcoding, Multiple Description Scalar Quantisation, Drift Distortion
AbstractThis paper proposes a multiple description (MD) transcoding scheme capable of preventing drift by distortion accumulation in temporally predicted motion compensated slices. Drift compensation is achieved by generating a controlled amount of side information to be used for decoding whenever a description fails to reach the end user terminal. The side information is generated by re-encoding the transcoding residue with an independent quantisation parameter which also controls redundancy. A simplified architecture is devised to reduce transcoding complexity in regard to the number of processing functions and buffer requirements. The experimental results show that temporally predicted frames do not suffer from drift and their quality is significantly improved at reduced redundancy cost in comparison with a classic MD transcoding scheme.

P4-22
TitleH.264/AVC to Wavelet-Based Scalable Video Transcoding Supporting Multiple Coding Configurations
AuthorEduardo Peixoto, Toni Zgaljic, Ebroul Izquierdo (Queen Mary, University of London, U.K.)
Pagepp. 562 - 565
KeywordTranscoding, Scalable Video Coding
AbstractScalable Video Coding (SVC) enables low complexity adaptation of the compressed video, providing an efficient solution for video content delivery through heterogeneous networks and to different displays. However, legacy video and most commercially available content capturing devices use conventional non-scalable coding, e.g., H.264/AVC. This paper proposes an efficient transcoder from H.264/AVC to a wavelet-based SVC to exploit the advantages offerend by the SVC technology. The proposed transcoder is able to cope with different coding configurations in H.264/AVC, such as IPP or IBBP with multiple reference frames. To reduce the transcoder's complexity, motion information and presence of the residual data extracted from the decoded H.264/AVC video are exploited. Experimental results show a good performance of the proposed transcoder in terms of decoded video quality and system complexity.

P4-23
TitleEdge-Adaptive Transforms for Efficient Depth Map Coding
AuthorGodwin Shen, Woo-shik Kim, Sunil Kumar Narang, Antonio Ortega (University of Southern California, U.S.A.), Jaejoon Lee, HoCheon Wey (Samsung Advanced Institute of Technology, Republic of Korea)
Pagepp. 566 - 569
KeywordMultiview plus depth (MVD), Depth coding, Rate-distortion optimization
AbstractIn this work a new set of edge-adaptive transforms (EATs) is presented as an alternative to the standard DCTs used in image and video coding applications. These transforms avoid filtering across edges in each image block, thus, they avoid creating large high frequency coefficients. These transforms are then combined with the DCT in H.264/AVC and a transform mode selection algorithm is used to choose between DCT and EAT in an RD-optimized manner. These transforms are applied to coding depth maps used for view synthesis in a multi-view video coding system, and provides up to 29% bit rate reduction for a fixed quality in the synthesized views.

P4-24
TitleDirection-Adaptive Hierarchical Decomposition for Image Coding
AuthorTomokazu Murakami, Keita Takahashi, Takeshi Naemura (The University of Tokyo, Japan)
Pagepp. 570 - 573
KeywordImage decomposition, directional prediction, L1 norm, directional transform, wavelet
AbstractA new model of decomposing an image hierarchically into direction-adaptive subbands using pixel-wise direction estimation is presented. For each decomposing operation, an input image is divided into two parts: a base image subsampled from the input image and subband components. The subband components consist of residuals of estimating the pixels skipped through the subsampling, which ensures the invertibility of the decomposition. The estimation is performed in a direction-adaptive way, whose optimal direction is determined by a L1 norm criterion for each pixel, aiming to achieve good energy compaction that is suitable for image coding. Furthermore, since the L1 norms are obtained from the base image alone, we do not need to retain the directional information explicitly, which is another advantage of our model. Experimental results show that the proposed model can achieve lower entropy than conventional Haar or D5/3 discrete wavelet transform in case of lossless coding.


[Image/video processing and related topics]

P4-25
TitleA Robust Video Super-Resolution Algorithm
AuthorXinfeng Zhang (Institute of Computing Technology, Chinese Academy of Sciences, China), Ruiqin Xiong, Siwei Ma, Wen Gao (School of Electronic Engineering and Computer Science, Peking University, China)
Pagepp. 574 - 577
Keywordsuper-resolution, kernel regression, irregular interpolation
AbstractIn this paper, we proposed a robust video super-resolution reconstruction method based on spatial-temporal orienta-tion-adaptive kernel regression. First, we proposed a robust registration efficiency model to reflect the temporal information reliability. Second, we proposed a spatial-temporal steering kernel considering motions between frames and structures in each low resolution frame. Simula-tion results demonstrate that our new super-resolution method substantially improves both the subjective quality and objective quality than other resolution enhancement methods.

P4-26
TitleAn Efficient Method for the Detection of Ringing Artifacts and De-Ringing in JPEG Image
AuthorShen-Chuan Tai, Bo-Jhih Chen, Mankit Choi (Institute of Computer and Communication Engineering, Department of Electrical Engineering, National Cheng Kung University, Taiwan)
Pagepp. 578 - 581
KeywordJPEG, ringing artifacts, image compression
AbstractJPEG standard is commonly used for the still image compression. However, the DCT-based coding of JPEG is one of the lossy compression tools and introduces most artifacts in the decompressed image such as blocking and ringing artifacts. In this paper, our proposed method focuses on how to efficiently detect the blocks which cause ringing artifacts. Moreover, these ringing blocks will be further filtered and the texture region will be able to preserve as well as the smooth region. Simulation results show that our proposed method can outperform the related algorithms in subjectively and objectively. By comparing JPEG compressed image, the decompressed image using our algorithm can achieve the better PSNR as well as the visual performance, especially at the lower quality coding (the higher compression rate).

P4-27
TitleLow Delay Distributed Video Coding Using Data Hiding
AuthorKrishna Rao Vijayanagar, Bowen Dan, Joohee Kim (Illinois Institute of Technology, U.S.A.)
Pagepp. 582 - 585
KeywordDistributed Video Coding, Data hiding, Low Delay DVC
AbstractDistributed Video Coding (DVC) is a popular topic in the research community and the past years have seen several different implementations. DVC has been proposed as a solution for applications that have limited battery resources and low hardware complexity, thus necessitating a low complexity encoder. An ideal application would be in remote surveillance/monitoring or live video conferencing. However, current solutions use iteratively decodable channel codes like LDPCA or Turbo codes that have large latencies. In order to make real-time communication possible. The proposed architecture makes efficient use of Skip blocks to reduce the bitrate, eliminates the iterative decoding nature of the Wyner-Ziv (WZ) channel and uses a simple data-hiding based compression algorithm. This drastically cuts down on the time complexity of the decoding procedure while still maintaining an rate-distortion performance better than that of H.264/AVC Intra coding and other current DVC solutions.

P4-28
TitleFFT-Based Full-Search Block Matching Using Overlap-Add Method
AuthorHidetake Sasaki, Zhen Li, Hitoshi Kiya (Tokyo Metropolitan University, Japan)
Pagepp. 586 - 589
Keywordblock matching, FFT, overlap-add method, pattern recognition, motion estimation
AbstractOne category of fast full-search block matching algorithms (BMAs) is based on the fast Fourier transformation (FFT). In conventional methods in this category, the macroblock size must be adjusted to the search window size by zero-padding. In these methods, the memory consumption and computational complexity heavily depend on the size difference between the macroblock and the search window. Thus, we propose a novel FFT-based BMA to solve this problem. The proposed method divides the search window into multiple sub search windows to versatilely control the difference between the macroblock and the search window sizes. Simulation results show the effectiveness of the proposed method.


[Quality, system, applications, and other topics]

P4-29
TitleTemporal Inconsistency Measure for Video Quality Assessment
AuthorSongnan Li, Lin Ma, Fan Zhang, King Ngi Ngan (The Chinese University of Hong Kong, Hong Kong)
Pagepp. 590 - 593
Keywordvideo quality assessment, spatial visual quality measure, temporal inconsistency measure
AbstractVisual quality assessment plays a crucial role in many vision related signal processing applications. In the literature, more efforts have been spent on spatial visual quality measure. Although a large number of video quality metrics have been proposed, the methods to use temporal information for quality assessment are less diversified. In this paper, we propose a novel method to measure the temporal impairments. The proposed method can be incorporated into any image quality metric to extend it into a video quality metric. Moreover, it is easy to apply the proposed method in video coding system to incorporate with MSE for rate-distortion optimization.

P4-30
TitleThe Dependence of Visual Noise Perception on Background Color and Luminance
AuthorMakoto Shohara, Kazunori Kotani (Japan Advanced Institute of Science and Technology, Japan)
Pagepp. 594 - 597
KeywordVisual system, Noise measurement, Color measurement, Noise generators, Shot noise
AbstractThis paper describes the dependency of noise perception on background color and luminance of noise quantitatively. We use the luminance and chromatic noise models derived from shot noise model. And we conduct subjective and quantitative experiments using a modified gray scale method. The subjective experiment results show the perceived color noise depends on the background color, but the perceived luminance noise does not. In addition, the perceived chromatic noise level is about 8 times smaller than calculated color noise.

P4-31
TitleAn Adaptive Low-Complexity Global Motion Estimation Algorithm
AuthorMd Nazmul Haque, Moyuresh Biswas, Mark R. Pickering, Michael R. Frater (The University of New South Wales, Australia)
Pagepp. 598 - 601
Keywordglobal motion estimation, video coding, image registration, gradient-descent optimization
AbstractA limitation of current global motion estimation approaches is the additional complexity of the gradient-descent optimization that is typically required to calculate the optimal set of global motion parameters. In this paper we propose a new low-complexity algorithm for global motion estimation. The complexity of the proposed algorithm is reduced by performing the majority of the operations in the gradient-descent optimization using logic operations rather than full-precision arithmetic operations. This use of logic operations means that the algorithm can be implemented much more easily in hardware platforms such as field programmable gate arrays (FPGAs). Experimental results show that the execution time for software implementations of the new algorithm is reduced by a factor of almost four when compared to existing fast implementations without any significant loss in registration accuracy.

P4-32
TitleScalable Multiple Description Video Coding Using Successive Refinement of Side Quantizers
AuthorMuhammad Majid, Charith Abhayaratne (The University of Sheffield, U.K.)
Pagepp. 602 - 605
KeywordMultiple description coding, quality scalability, resilience
AbstractIn this paper, we present a new method for scalable multiple description video coding based on motion compensated temporal filtering and multiple description scalar quantizer with successive refinement. In our method quality scalability is achieved by successively refining the side quantizers of a multiple description scalar quantizer. The rate of each description is allocated by considering different refinement levels for each spatio-temporal subband. The performance of the proposed scheme under lossless and lossy channel conditions are presented and compared with single scalable description video coding.

P4-33
TitleBit-Plane Compressive Sensing with Bayesian Decoding for Lossy Compression
AuthorSz-Hsien Wu (Electronic Engineering, National Chiao Tung University, Taiwan), Wen-Hsiao Peng (Computer Science, National Chiao Tung University, Taiwan), Tihao Chiang (Electronic Engineering, National Chiao Tung University, Taiwan)
Pagepp. 606 - 609
KeywordCompressive Sensing, Bayesian estimation, Bit-plane
AbstractThis paper addresses the problem of reconstructing a compressively sampled sparse signal from its lossy and possibly insufficient measurements. The process involves estimations of sparsity pattern and sparse representation, for which we derived a vector estimator based on the Maximum a Posteriori Probability (MAP) rule. By making full use of signal prior knowledge, our scheme can use a measurement number close to sparsity to achieve perfect reconstruction. It also shows a much lower error probability of sparse pattern than prior work, given insufficient measurements. To better recover the most significant part of the sparse representation, we further introduce the notion of bit-plane separation. When applied to image compression, the technique in combination with our MAP estimator shows promising results as compared to JPEG: the difference in compression ratio is seen to be within a factor of two, given the same decoded quality.

P4-34
TitleA Reduced-Reference Metric Based on the Interest Points in Color Images
AuthorMichael Nauge, Mohamed-Chaker Larabi, Christine Fernandez (University of Poitiers, France)
Pagepp. 610 - 613
KeywordMetric, Interest point, Reduced reference, quality, saliency
AbstractIn the last decade, an important research effort has been dedicated to quality assessment from the subjective and the objective points of view. The focus was mainly on Full Reference (FR) metrics because of the ability to compare to an original. Only few works were oriented to Reduced Reference (RR) or No Reference (NR) metrics, very useful for applications where the original image is not available such as transmission or monitoring. In this work, we propose an RR metric based on two concepts, the interest points of the image and the objects saliency on color images. This metric needs a very low amount of data (lower than 8 bytes) to be able to compute the quality scores. The results show a high correlation between the metric scores and the human judgement and a better quality range than well-known metrics like PSNR or SSIM. Finally, interest points have shown that they can predict the quality of color images.