(Go to Top Page)

28th Picture Coding Symposium
Technical Program


Technical Program:   SIMPLE version   DETAILED version with abstract
Author Index:   HERE

Session Schedule


Wednesday, December 8, 2010

Opening Session
8:30 - 8:45
K1  Keynote Speech 1
8:45 - 9:30
Break
9:30 - 9:45
O1  Oral Session 1: FTV
9:45 - 11:45
Lunch
11:45 - 13:00
S1  Special Session 1: 3DTV/FTV
13:00 - 15:30
D1  Panel Discussion 1: 3DTV/FTV
15:30 - 16:30
Break
16:30 - 16:45
P1  Poster Session 1
16:45 - 18:15
Welcome Reception

Thursday, December 9, 2010

K2  Keynote Speech 2
8:45 - 9:30
P2  Poster Session 2
9:30 - 11:00
Break
11:00 - 11:15
T1  Tutorial Session 1
11:15 - 12:00
Lunch
12:00 - 13:15
O2  Oral Session 2: Depth Map Coding
13:15 - 15:15
Break
15:15 - 15:30
O3  Oral Session 3: New Techniques for Video Coding
15:30 - 17:30
Banquet

Friday, December 10, 2010

K3  Keynote Speech 3
8:45 - 9:30
P3  Poster Session 3
9:30 - 11:00
Break
11:00 - 11:15
T2  Tutorial Session 2
11:15 - 12:00
Lunch
12:00 - 13:15
S2  Special Session 2: Beyond H.264/MPEG-4 AVC
13:15 - 15:15
D2  Panel Discussion 2: Beyond H.264/MPEG-4 AVC
15:15 - 16:15
Break
16:15 - 16:30
P4  Poster Session 4
16:30 - 18:00



List of Papers



Wednesday, December 8, 2010

Session K1  Keynote Speech 1
Time: 8:45 - 9:30 Wednesday, December 8, 2010
Chair: Takahiro Saito (Kanagawa University, Japan)

K1-1 (Time: 8:45 - 9:30)
Title(Keynote Speech) Research and Activities on Ultra-realistic Communications
AuthorKazumasa Enami (National Institute of Information and Communications Technology, Japan)
AbstractUltra-realistic communications are future means of communicating that provide users with a highly realistic presence. The technologies to achieve these communications are varied and include ultra-high definition/3-dimensional images, reproduction of highly realistic surround sound, and multisensory communication that includes touch and smell. Japan Broadcasting Corporation (NHK) is developing Ultra-HDTV system, called Super Hi-vision, which has four times the pixels of HDTV both horizontally and vertically. The images can be viewed as close as 0.75 times the height of the screen. At the proximity, the angle of viewing is 100 degrees, which confers an enhanced sense of reality. National Institute of Information and Communications Technology (NICT) is researching Ultra-realistic communication technologies that provide natural and real information to everybody. Main research subjects are holographic 3D video, super multi-view 3D video without special glasses on and super surround audio systems. We also researching necessary requirements for ultra-realistic systems based on underlying principles of human information processing. The Ultra-Realistic Communications Forum (URCF) was established on 2007 to ensure efficient industry-academia-government collaboration in research, development, verification experiments, and standardization on ultra-realistic technologies. Especially standardization of FTV (Free-viewpoint TV) is being actively discussed in a working group of URCF. Ultra-realistic communication systems have quite huge amount of data. For example, the data rate of Super Hi-vision, super multi-view 3D video and electronic holography video is 72Gbps, more than 200Gbps and more than several hundred Tbps respectively. Then efficient video coding algorithms are required for the ultra-realistic communications. In the Symposium, I will introduce research on ultra-realistic communications, activities of URCF and describe the expectation of new picture coding technology.


Session O1  Oral Session 1: FTV
Time: 9:45 - 11:45 Wednesday, December 8, 2010
Chair: Marek Domanski (Poznan University of Technology, Poland)

O1-1 (Time: 9:45 - 10:15)
TitleFree Viewpoint Image Generation with Super Resolution
AuthorNorishige Fukushima, Yutaka Ishibashi (Graduate School of Engineering, Nagoya Institute of Technology, Japan)
Pagepp. 1 - 4
KeywordFree Viewpoint TV, Image Based Rendering, Ray Space, Light Field, Super Resolution
AbstractIn this paper, we propose a method of free viewpoint image generation with super resolution. In the conventional approach, such as nearest neighbor and linear interpolation, the synthetic image on zoomed virtual view tends to have low resolution. To overcome this problem, we combine the super resolution process with free viewpoint image generation. The experimental results show that synthesized image in the effective range has higher PSNR than the conventional method.

O1-2 (Time: 10:15 - 10:45)
TitleReducing Bitrates of Compressed Video with Enhanced View Synthesis for FTV
AuthorLu Yang, Meindert Onno Wildeboer, Tomohiro Yendo, Mehrdad Panahpour Tehrani (Nagoya University, Japan), Toshiaki Fujii (Tokyo Institute of Technology, Japan), Masayuki Tanimoto (Nagoya University, Japan)
Pagepp. 5 - 8
KeywordFree-viewpoint TV (FTV), video compression, bitrates, view synthesis, MVC
AbstractWe deal with the bitrates of view synthesis at the decoder side of FTV that would use compressed depth maps and views. The focus is to reduce bitrates required for generating the high-quality virtual view. We employ a reliable view synthesis method which is compared with standard MPEG view synthesis software. The experimental results show that the bitrates required for synthesizing high-quality virtual view could be reduced by utilizing our enhanced view synthesis technique to improve the PSNR at medium bitrates.

O1-3 (Time: 10:45 - 11:15)
TitleDepth Map Processing with Iterative Joint Multilateral Filtering
AuthorPoLin Lai (Samsung Telecommunications America, U.S.A.), Dong Tian (Mitsubishi Electric Research Laboratories, U.S.A.), Patrick Lopez (Technicolor, Research and Innovation, France)
Pagepp. 9 - 12
Keyword3D video, depth maps, joint filtering, iterative filtering, view synthesis
AbstractDepth maps estimated using stereo matching between frames from different video views typically exhibit false contours and noisy artifacts around object boundaries. In this paper, iterative joint multilateral filtering is proposed to deal with these artifacts. The proposed filter consists of multiple filter kernels. Knowing that the estimated depth maps are erroneous, besides the kernels which measure the proximity of depth samples and the similarity between depth sample values, we further develop kernels which measure similarity between the corresponding video pixel values. To increase reliability, these novel kernels operate on the color (RGB) domain instead of only on the luminance domain. Furthermore, the filter shapes are designed to adapt brightness variations. Finally, to tackle large misalignment between boundaries in depth maps and in the corresponding video frames, iterative approach is utilized. Our results demonstrate that the proposed method can significantly improve the boundaries in depth maps and can reduce false contours. With the processed depth maps, it is observed that the quality of object boundaries in synthesized views can be improved.

O1-4 (Time: 11:15 - 11:45)
TitleStereoscopic Depth Estimation Using Fuzzy Segment Matching
AuthorKrzysztof Wegner, Olgierd Stankiewicz, Marek Domański (Poznan University of Technology, Poland)
Pagepp. 13 - 16
KeywordStereo matching, depth estimation, soft segmentation, disparity calculation, fuzzy set
AbstractStereo matching techniques usually match segments or blocks of pixels. This paper proposes to match segments defined as fuzzy sets of pixels. The proposed matching method is applicable to various techniques of stereo matching as well as to different measures of differences between pixels. In the paper, embedment of this approach into the state-of-the-art depth estimation software is described. Obtained experimental results show that the proposed way of stereo matching increases reliability of various depth estimation techniques.


Session S1  Special Session 1: 3DTV/FTV
Time: 13:00 - 15:30 Wednesday, December 8, 2010
Chair: Toshiaki Fujii (Tokyo Institute of Technology, Japan)

S1-1 (Time: 13:00 - 13:30)
Title(Invited Talk) Challenges in Multiview Video - The 3 D's
AuthorAntonio Ortega (University of Southern California, U.S.A.)
Pagep. 17
Keywordmultiview video, multiview video coding, depth estimation
AbstractWe highlight some of the challenges facing wide adoption of multiview video technologies, by focusing on design issues involving displays, depth data and tools for content delivery. Our goal is to show that uncertainty about what choices will be made in each of these areas have a significant impact on the design of coding tools.

S1-2 (Time: 13:30 - 14:00)
Title(Invited Talk) Vision Field Capturing and Its Applications in 3DTV
AuthorQionghai Dai (Tsinghua University, China)
Pagep. 18
Keywordmultiview capturing, vision field capturing, 3DTV
Abstract3D video capturing acquires the visual information in 3D manner, which possesses the first step of the entire 3DTV system chain before 3D coding, transmission and visualization. As the cost of sensors reduces recently, many systems utilize multiple cameras to acquire visual information, which is called multiview capturing. We start with reviewing these multiview systems built in the past decades. Then, a new concept of vision field capturing is introduced. Finally, some new techniques using TOF(time-of-flight) camera and 3D scanner will also be presented.

S1-3 (Time: 14:00 - 14:30)
Title(Invited Talk) 3D Information Coding
AuthorJoern Ostermann (Hannover University, Germany)
Pagep. 19
KeywordMPEG, 3D reconstruction, 3DTV, stereo video, multiview display
Abstract3D visual information can be presented to viewers using stereo displays. For these displays, the two video sequences of a stereo pair are jointly coded using MVC. The extension of coding standards for use with the existing MPEG-2 or AVC infrastructure is currently developed. Auto stereoscopic displays enabling motion parallax will be supported by the future 3DV standard. Several video streams and a depth map will enable the display to render the correct views of the scene.

S1-4 (Time: 14:30 - 15:00)
Title(Invited Talk) 3D Television System Based on Integral Photography
AuthorTomoyuki Mishina (NHK, Japan)
Pagep. 20
Keyword3D image, integral photography, full-parallax, Super Hi-Vision
AbstractIntegral photography is a photographic technique which is able to reconstruct three-dimensional (3D) images having full-parallax, and the reconstructed 3D images can be seen without wearing special glasses. We have been researching into a 3D television system based on integral photography. This paper describes our system in which an ultra-high-definition imaging system called Super Hi-Vision is applied to integral photography.

S1-5 (Time: 15:00 - 15:30)
Title(Invited Talk) Progress from Stereoscopic to Three-Dimensional Displays Based on Visual Perception
AuthorSumio Yano (NHK, formerly NICT, Japan)
Pagep. 21
Keywordstereoscopic displays, multi-view stereoscopic displays, three-dimensional display, light field reproduction, human visual field
AbstractThe proper characteristics of stereoscopic and multi-view displays are described in the context of human visual perception at first. Next, the development of three-dimensional displays, matched to the function of human visual field, is described accounting for these points. These three-dimensional displays were developed based on the principle of light field reproduction. From these issues, the processing technology for large capacity image information is important for the development of the three-dimensional image systems in the future.


Session D1  Panel Discussion 1: 3DTV/FTV
Time: 15:30 - 16:30 Wednesday, December 8, 2010
Chair: Toshiaki Fujii (Tokyo Institute of Technology, Japan)

D1-1 (Time: 15:30 - 16:30)
Title(Panel Discussion) 3DTV/FTV
AuthorChair: Toshiaki Fujii (Tokyo Institute of Technology, Japan), Panelists: Kazumasa Enami (National Institute of Information and Communications Technology, Japan), Antonio Ortega (University of Southern California, U.S.A.), Qionghai Dai (Tsinghua University, China), Joern Ostermann (Hannover University, Germany), Tomoyuki Mishina (NHK, Japan), Sumio Yano (NHK, formerly NICT, Japan)


Session P1  Poster Session 1
Time: 16:45 - 18:15 Wednesday, December 8, 2010
Chair: Jiro Katto (Waseda University, Japan)


[3DTV/FTV/multi-view-related topics]

P1-1
Title3D Space Representation Using Epipolar Plane Depth Image
AuthorTakashi Ishibashi, Tomohiro Yendo, Mehrdad Panahpour Tehrani (Graduate School of Engineering, Nagoya University, Japan), Toshiaki Fujii (Graduate School of Science and Engineering, Tokyo Institute of Technology, Japan), Masayuki Tanimoto (Graduate School of Engineering, Nagoya University, Japan)
Pagepp. 22 - 25
KeywordFTV, EPI, EPDI, Ray-Space, GDM
AbstractWe propose a novel 3D space representation for multi-view video, using epipolar plane depth images (EPDI). Multi-view video plus depth (MVD) is used as common data format for FTV(Free-viewpoint TV), which enables synthesizing virtual view images. Due to large amount of data and complexity of the multi-view video coding (MVC), compression of MVD is a challenging issue. We address this problem and propose a new representation that is constructed from MVD using rayspace. MVD is converted into image and depth ray-spaces. The proposed representation is obtained by converting each of ray-spaces into a global depth map and a texture map using EPDI. Experiments demonstrates the analysis of this representation, and its efficiency.

P1-2
TitleView Synthesis Error Analysis for Selecting the Optimal QP of Depth Map Coding in 3D Video Application
AuthorYanwei Liu, Song Ci, Hui Tang (Institute of Acoustics, Chinese Academy of Sciences, China)
Pagepp. 26 - 29
Keywordview synthesis error analysis, depth map coding, selecing QP
AbstractIn 3D video communication, how to select the appropriate quantization parameter (QP) for depth map coding is very important for obtaining the optimal view synthesis quality. This paper first analyzes the depth uncertainty induced two kinds of view synthesis errors, namely the original depth error induced view synthesis error and the depth compression induced view synthesis error, and then proposes a quadratic model to characterize the relationship between the view synthesis quality and the depth quantization step size. The proposed model can find the inflexion point in the curve of the view synthesis quality with the increasing depth quantization step size. Experimental results show that, given the rate constraint for depth map, the proposed model can accurately find the optimal QP for depth map coding.

P1-3
TitleSuppressing Texture-Depth Misalignment for Boundary Noise Removal in View Synthesis
AuthorYin Zhao (Zhejiang University, China), Zhenzhong Chen (Nanyang Technological University, Singapore), Dong Tian (Mitsubishi Electric Research Labs, U.S.A.), Ce Zhu (Nanyang Technological University, Singapore), Lu Yu (Zhejiang University, China)
Pagepp. 30 - 33
Keyword3D video, boundary noise, DIBR, texture-depth misalignment, view synthesis
AbstractDuring view synthesis based on depth maps, also known as Depth-Image-Based Rendering (DIBR), annoying artifacts are often generated around foreground objects, yielding the visual effects that slim silhouettes of foreground objects are scattered into the background. The artifacts are referred as the boundary noises. We investigate the cause of boundary noises, and find out that they result from the misalignment between texture and depth information along object boundaries. Accordingly, we propose a novel solution to remove such boundary noises by applying restrictions during forward warping on the pixels within the texture-depth misalignment regions. Experiments show this algorithm can effectively eliminate most boundary noises and it is also robust for view synthesis with compressed depth and texture information.

P1-4
TitleA High Efficiency Coding Framework for Multiple Image Compression of Circular Camera Array
AuthorDongming Xue (Interdisciplinary Graduate School of Science and Engineering, Tokyo Institute of Technology, Japan), Akira Kubota (Faculty of Science and Engineering, Chuo University, Japan), Yoshinori Hatori (Interdisciplinary Graduate School of Science and Engineering, Tokyo Institute of Technology, Japan)
Pagepp. 34 - 37
Keywordmulti-view video coding, virtual plane, polar axis, poxel framework
AbstractMany existing multi-view video coding techniques remove inter-viewpoint redundancy by applying disparity compensation in conventional video coding frameworks, e.g., H264/MPEG4. However, conventional methodology works ineffectively as it ignores the special features of the inter-view-point disparity. This paper proposed a framework using virtual plane for multi-view image compression, such as we can largely reduce the disparity compensation cost. Based on this VP predictor, we design a poxel (probabilistic voxelized volume) framework, which can integrate the information of the cameras in different view-points in the polar axis to obtain a more effective compression performance. In addition, considering the replay convenience of the multi-view video at the receiving side, we reformed overhead information in polar axis at the sending side in advance.

P1-5
TitleExploiting Depth Information for Fast Multi-View Video Coding
AuthorBrian W. Micallef, Carl J. Debono, Reuben A. Farrugia (University of Malta, Malta)
Pagepp. 38 - 41
Keyword3DTV, disparity vector estimation, geometric disparity vector predictor, multi-view video coding
AbstractMulti-view video coding exploits inter-view redundancies to compress the video streams and their associated depth information. These techniques utilize disparity estimation techniques to obtain disparity vectors (DVs) across different views. However, these methods contribute to the majority of the computational power needed for multi-view video encoding. This paper proposes a solution for fast disparity estimation based on multi-view geometry and depth information. A DV predictor is first calculated followed by an iterative or a fast search estimation process which finds the optimal DV in the search area dictated by the predictor. Simulation results demonstrate that this predictor is reliable enough to determine the area of the optimal DVs to allow a smaller search range. Furthermore, results show that the proposed approach achieves a speedup of 2.5 while still preserving the original rate-distortion performance.

P1-6
TitleAn Adaptive Early Skip Mode Decision Scheme for Multiview Video Coding
AuthorBruno Zatt (Federal University of Rio Grande do Sul, Brazil), Muhammad Shafique (Karlsruhe Institute of Technology, Germany), Sergio Bampi (Federal University of Rio Grande do Sul, Brazil), Jörg Henkel (Karlsruhe Institute of Technology, Germany)
Pagepp. 42 - 45
KeywordMVC, Early Skip, Video Coding, Mode Decision
AbstractIn this work a novel scheme is proposed for adaptive early SKIP mode decision in the multiview video coding based on mode correlation in the 3D-neighborhood, variance, and rate-distortion properties. Our scheme employs an adaptive thresholding mechanism in order to react to the changing values of Quantization Parameter (QP). Experimental results demonstrate that our scheme provides a consistent time saving over a wide range of QP values. Compared to the exhaustive mode decision, our scheme provides a significant reduction in the encoding complexity (up to 77%) at the cost of a small PSNR loss (0.172 dB in average). Compared to state-of-the-art, our scheme provides an average 2x higher complexity reduction with a relatively higher PSNR value (avg. 0.2 dB).

P1-7
TitleElectoronic Hologram Generation Using High Quality Color and Depth Information of Natural Scene
AuthorKousuke Nomura (Tokyo University of Science, Japan), Ryutaro Oi, Taiichiro Kurita (National Institute of Information and Communications Technology, Japan), Takayuki Hamamoto (Tokyo University of Science, Japan)
Pagepp. 46 - 49
KeywordCGH, color hologram, real Scene, natural light, optically reconstruction
AbstractRecently, the computer generated hologram (CGH) methods are heavily researched. In those conventional CGH, however, it is not so common to provide a hologram of real scene. This time we report that we could get the properly reconstruction of image by using our proposal method against the color hologram anew. We confirmed that our proposed color hologram properly recorded both color and the 3-D information of the space by an assessment experiment which optically reconstructs the hologram.


[Beyond H.264/MPEG-4 AVC and related topics]

P1-8
TitleComplementary Coding Mode Design Based on R-D Cost Minimization for Extending H.264 Coding Technology
AuthorTomonobu Yoshino, Sei Naito, Shigeyuki Sakazawa, Shuichi Matsumoto (KDDI R&D Laboratories, Japan)
Pagepp. 50 - 53
Keywordlow bit-rate video coding, high resolution video coding, H.264, SKIP mode
AbstractTo improve high resolution video coding efficiency under low bit-rate condition, an appropriate coding mode is required from an R-D optimization (RDO) perspective, although a coding mode defined within the H.264 standard is not always optimal for RDO criteria. With this in mind, we previously proposed extended SKIP modes with close-to-optimal R-D characteristics. However, the additional modes did not always satisfy the optimal R-D characteristics, especially for low bit-rate coding. In response, in this paper, we propose an enhanced coding mode capable of providing a candidate corresponding to the minimum R-D cost by controlling the residual signal associated with the extended SKIP mode. The experimental result showed that the PSNR improvement against H.264 and our previous approach reached 0.42 dB and 0.24 dB in the maximum case, respectively.

P1-9
TitleEnhanced Video Compression with Region-Based Texture Models
AuthorFan Zhang, David Bull (University of Bristol, U.K.)
Pagepp. 54 - 57
Keywordvideo, compression, texture, warping, synthesis
AbstractThis paper presents a region-based video compression algorithm based on texture warping and synthesis. Instead of encoding whole images or prediction residuals after translational motion estimation, this algorithm employs a perspective motion model to warp static textures and uses a texture synthesis approach to synthesise dynamic textures. Spatial and temporal artefacts are prevented by an in-loop video quality assessment module. The proposed method has been integrated into an H.264 video coding framework. The results show significant bitrate savings, up to 55%, compared with H.264, for similar visual quality.

P1-10
TitleCoding Efficient Improvement by Adaptive Search Center Definition
AuthorKyohei Oba, Takahiro Bandou, Tian Song, Takashi Shimamoto (Tokushima University, Japan)
Pagepp. 58 - 61
KeywordH.264/AVC, Motion Estimation, Inter Prediction, RDO
AbstractIn this paper, an efficient search center definition algorithm is proposed for H.264/AVC. H.264/AVC achieved high coding efficiency by introducing some new coding tools including a new definition of the search center. However, the definition of the search center is not efficient in the case of significant motions. This work proposes some new search center candidates using spatial and temporal correlations of motion vectors to improve the coding efficiency. Simulation results show that the proposed search centers can achieve very much bit saving but induced high computation complexity. Additional complexity reduction algorithm is also introduced to improve the trade off between bit saving and implementation performance. This work realized a maximum bit saving of 19%.

P1-11
TitleA Novel Coding Scheme for Intra Pictures of H.264/AVC
AuthorJin Young Lee, Jaejoon Lee, Hochen Wey, Du-Sik Park (Samsung Advanced Institute of Technology, Samsung Electronics Co., Ltd., Republic of Korea)
Pagepp. 62 - 65
KeywordH.264/AVC, intra coding, intra prediction
AbstractA novel intra coding scheme is proposed to improve coding performance in intra pictures of H.264/AVC. The proposed method generates two sub-images, which are defined as a sampled image and a prediction error image in this paper, from an original image, and then encodes them separately. Especially, in an intra prediction process of encoding, the sampled image employs the original intra prediction modes, while the prediction error image uses newly defined four intra prediction modes. Experimental results demonstrate that the proposed method achieves significantly higher intra coding performance and reduces encoding complexity with the smaller number of a rate-distortion (RD) cost calculation process, as compared with the original intra coding method of H.264/AVC.

P1-12
TitleEntropy Coding in Video Compression Using Probability Interval Partitioning
AuthorDetlev Marpe, Heiko Schwarz (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany), Thomas Wiegand (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute/Technical University of Berlin, Germany)
Pagepp. 66 - 69
Keywordentropy coding, variable length coding
AbstractWe present a novel approach to entropy coding, which provides the coding efficiency and simple probability modeling capability of arithmetic coding at the complexity level of Huffman coding. The key element of the proposed approach is a partitioning of the unit interval into a small set of probability intervals. An input sequence of discrete source symbols is mapped to a sequence of binary symbols and each of the binary symbols is assigned to one of the probability intervals. The binary symbols that are assigned to a particular probability interval are coded at a fixed probability using a simple code that maps a variable number of binary symbols to variable length codewords. The probability modeling is decoupled from the actual binary entropy coding. The coding efficiency of the probability interval partitioning entropy (PIPE) coding is comparable to that of arithmetic coding.

P1-13
TitleSeparable Wiener Filter Based Adaptive In-Loop Filter for Video Coding
AuthorMischa Siekmann, Sebastian Bosse, Heiko Schwarz, Thomas Wiegand (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany)
Pagepp. 70 - 73
KeywordVideo Compression, In-loop Filtering, Wiener Filter, Separable Image Filtering, Regional Filtering
AbstractRecent investigations have shown that a non-separable Wiener filter, that is applied inside the motion-compensation loop, can improve the coding efficiency of hybrid video coding designs. In this paper, we study the application of separable Wiener filters. Our design includes the possibility to adaptively choose between the application of the vertical, horizontal, or combined filter. The simulation results verify that a separable in-loop Wiener filter is capable of providing virtually the same increase in coding efficiency as a non- separable Wiener filter, but at a significantly reduced decoder complexity.


[Image/video coding and related topics]

P1-14
TitleHyperspectral Image Compression Suitable for Spectral Analysis Application
AuthorKazuma Shinoda, Yukio Kosugi, Yuri Murakami, Masahiro Yamaguchi, Nagaaki Ohyama (Tokyo Institute of Technology, Japan)
Pagepp. 74 - 77
KeywordHyperspectral image, Image compression, Vegetation index, JPEG2000
AbstractThis paper presents a HSI compression considering the error of both vegetation index and spectral data. The proposed method separates a hyperspectral data into spectral data for vegetation index and residual data. Both of the data are encoded by using a seamless coding individually. By holding the spectral channels required for vegetation index in the head of the code-stream, a precise vegetation analysis can be done in a low bit rate. Additionally, by decoding the residual data, the spectral data can be reconstructed in low distortion.

P1-15
TitleA Background Model Based Method for Transcoding Surveillance Videos Captured by Stationary Camera
AuthorXianguo Zhang (Peking University, China), Luhong Liang, Qian Huang (Institute of Computing Technology, Chinese Academy of Sciences, China), Tiejun Huang, Wen Gao (Peking University, China)
Pagepp. 78 - 81
Keywordsurveillance, archive, transcode, background model, difference frame
AbstractReal-world video surveillance applications require storing videos without neglecting any part of scenarios for weeks or months. To reduce the storage cost, the high bit-rate videos from cameras should be transcoded into a more efficient compressed format with little quality loss. We propose a background model based method to improve the transcoding efficiency for surveillance videos captured by stationary cameras, and objectively measure it. Experimental Results show this method saves about half the used bits compared with the full-decoding-full-encoding method.

P1-16
TitleAnalysis of In-Loop Denoising in Lossy Transform Coding
AuthorEugen Wige, Gilbert Yammine (Multimedia Communications and Signal Processing / University of Erlangen-Nuremberg, Germany), Peter Amon, Andreas Hutter (Siemens Corporate Technology / Information and Automation Technologies, Germany), Andre Kaup (Multimedia Communications and Signal Processing / University of Erlangen-Nuremberg, Germany)
Pagepp. 82 - 85
Keywordhigh quality compression, predictor denoising, quantization effects
AbstractWhen compressing noisy image sequences, the compression efficiency is limited by the noise amount within these image sequences as the noise part cannot be predicted. In this paper, we investigate the influence of noise within the reference frame on lossy video coding of noisy image sequences. We estimate how much noise is left within a lossy coded reference frame. Therefore we analyze the transform and quantization step inside a hybrid video coder, specifically H.264/AVC. The noise power after transform, quantization, and inverse transform is calculated analytically. We use knowledge of the noise power within the reference frame in order to improve the inter frame prediction. For noise filtering of the reference frame, we implemented a simple denoising algorithm inside the H.264/AVC reference software JM15.1. We show that the bitrate can be decreased by up to 8.1% compared to the H.264/AVC standard for high resolution noisy image sequences.

P1-17
TitleAn Efficient Side Information Generation Using Seed Blocks for Distributed Video Coding
AuthorDongYoon Kim, DongSan Jun, HyunWook Park (KAIST, Republic of Korea)
Pagepp. 86 - 89
KeywordDistrivuted vidoe coding, Wyner-Ziv video coding, Side information
AbstractRecently, a new video coding technique, distributed video coding (DVC), is an emerging research area for low power video coding applications. In the DVC, the encoder is much simpler than the conventional video codec, whereas the decoder is very heavy. The DVC decoder exploits side information which is generated by motion compensated frame interpolation to reconstruct the Wyner-Ziv frame. This paper proposes an efficient side information generation algorithm using seed blocks for DVC. Seed blocks are firstly selected to be used for motion estimation of the other blocks. As the side information is close to the target image, the final reconstructed image in the DVC decoder has better quality and the compression ratio becomes high. The proposed method contributes to improve the DVC compression performance with reduced computing time. Experimental results show that accurate motion vectors are estimated by the proposed method and its computational complexity of motion estimation is significantly reduced in comparison with the previous methods.

P1-18
TitleCompressive Video Sensing Based on User Attention Model
AuthorJie Xu (Advanced Computing Research Laboratory, Institute of Computing Technology, Chinese Academy of Sciences, China), Jianwei Ma (School of computational Science, Florida State University, U.S.A.), Dongming Zhang, Yongdong Zhang, Shouxun Lin (Advanced Computing Research Laboratory, Institute of Computing Technology, Chinese Academy of Sciences, China)
Pagepp. 90 - 93
Keywordcompressive sensing, video, user attention model, ROI
AbstractWe propose a compressive video sensing scheme based on user attention model (UAM) for real video sequences acquisition. In this work, for every group of consecutive video frames, we set the first frame as reference frame and build a UAM with visual rhythm analysis (VRA) to automatically determine region-of-interest (ROI) for non-reference frames. The determined ROI usually has significant movement and attracts more attention. Each frame of the video sequence is divided into non-overlapping blocks of 16x16 pixel size. Compressive video sampling is conducted in a block-by-block manner on each frame through a single operator and in a whole region manner on the ROIs through a different operator. Our video reconstruction algorithm involves alternating direction l1-norm minimization algorithm (ADM) for the frame difference of non-ROI blocks and minimum total-variance (TV) method for the ROIs. Experimental results showed that our method could significantly enhance the quality of reconstructed video and reduce the errors accumulated during the reconstruction.

P1-19
TitleAdvanced Inpainting-Based Macroblock Prediction with Regularized Structure Propagation in Video Compression
AuthorYang Xu, Hongkai Xiong (Shanghai Jiao Tong University, China)
Pagepp. 94 - 97
Keywordbelief propagation, tensor voting, inpainting, model selection, H.264
AbstractIn this paper, we propose an optimized inpainting-based macroblock (MB) prediction mode (IP- mode) in the state-of-the-art H.264/AVC video compression engine, and investigate a natural extension of structured sparsity over the ordered Belief Propagation (BP) inference in inpainting-based prediction. The IP- mode is regularized by a global spatio-temporal consistency between the predicted content and the co-located known texture, and could be adopted in both Intra and Inter frames without redundant assistant information. It is solved by an optimization problem under Markov Random Field (MRF), and the structured sparsity of the predicted macroblock region is inferred by tensor voting projected from the decoded regions to tune the priority of message scheduling in BP with a more convergent manner. Rate-distortion optimization is maintained to select the optimal mode among the inpainting-based prediction (IP-), the intra-, and inter- modes. Compared to the existing prediction modes in H.264/AVC, the proposed inpainting-based prediction scheme is validated to achieve a better R-D performance for homogeneous visual patterns and behave a more robust error resilience capability with an intrinsic probabilistic inference.

P1-20
TitleAn Improved Iterative Algorithm for Calculating the Rate-Distortion Performance of Causal Video Coding for Continuous Sources and its Application to Real Video Data
AuthorEn-hui Yang, Chang Sun, Lin Zheng (University of Waterloo, Canada)
Pagepp. 98 - 101
Keywordvideo coding, causal video coding, rate-distortion performance, soft decision quantization, interactive algorithm
AbstractAn improved iterative algorithm is first proposed to calculate the rate-distortion performance of causal video coding for any continuous sources. Instead of using continuous reproduction alphabets, it utilizes finite reproduction alphabets and iteratively updates them along with transitional probabilities from the continuous source to reproduction letters, thus overcoming the computation complexity problem encountered when applying the algorithm recently proposed by Yang et al for discrete sources to continuous sources. The proposed algorithm converges in the sense that the rate-distortion cost is monotonically decreasing until a stationary point is reached. It is then applied to practical video data to establish some theoretic coding performance benchmark. In comparison with H.264, experiments show that under the same motion compensation setting, causal video coding offers a roughly 1 dB coding gain on average over H.264 for the IPPIPP…GOP structure. This suggests that an area one could explore to further improve the rate-distortion performance of H.264 be how quantization and coding should be performed conditionally given previous frames and coded frames and given motion compensation.

P1-21
TitleTechnical Design & IPR Analysis for Royalty-Free Video Codecs
AuthorCliff Reader (none, U.S.A.)
Pagepp. 102 - 105
KeywordRoyalty-free, video codec, performance, complexity, IPR
AbstractRoyalty-free standards for image and video coding have been actively discussed for over 20 years. This paper breaks down the issues of designing royalty-free codecs into the major topics of requirements, video coding tools, classes of patents and performance. By dissecting the codec using a hierarchy of major to minor coding tools, it is possible to pinpoint where a patent impacts the video coding, and what the consequence will be of avoiding the patented tool.

P1-22
TitleAccelerating Pixel Predictor Evolution Using Edge-Based Class Separation
AuthorSeishi Takamura, Masaaki Matsumura, Hirohisa Jozawa (NTT Cyber Space Laboratories, NTT Corporation, Japan)
Pagepp. 106 - 109
KeywordGenetic programming, lossless image coding, pixel prediction, divide and conquer
AbstractEvolutionary methods based on genetic programming (GP) enable dynamic algorithm generation, and have been successfully applied to many areas such as plant control, robot control, and stock market prediction. However, one of the challenges of this approach is its high computational complexity. Conventional image/video coding methods such as JPEG and H.264 all use fixed (non-dynamic) algorithms without exception. However, one of the challenges of this approach is its high computational complexity. In this article, we introduce a GP-based image predictor that is specifically evolved for each input image, as well as local image properties such as edge direction. Via the simulation, proposed method demonstrated ~180 times faster evolution speed and 0.02-0.1 bit/pel lower bit rate than previous method.

P1-23
TitleAvoidance of Singular Point in Reversible KLT
AuthorMasahiro Iwahashi (Nagaoka University of Technology, Japan), Hitoshi Kiya (Tokyo Metropolitan University, Japan)
Pagepp. 110 - 113
KeywordKLT, error, reversible, coding
AbstractPermutation of order and sign of signals are introduced to avoid the singular point problem of a reversible transform. Even though a transform in the lifting structure can be "reversible" in spite of rounding operations, its multiplier coefficients have singular points (SP). Around the SP, rounding errors are magnified to huge amount and the coding efficiency is decreased. We investigate effect of the permutations on rotation angles and confirmed that PSNR was improved by 14 [dB] for RGB color components.


[Image/video processing and related topics]

P1-24
TitleSuper-Resolution Decoding of JPEG-Compressed Image Data with the Shrinkage in the Redundant DCT Domain
AuthorTakashi Komatsu, Yasutaka Ueda, Takahiro Saito (Kanagawa University, Japan)
Pagepp. 114 - 117
KeywordJPEG, decoding, redundant DCT, shrinkage, super-resolution
AbstractAlter, Durand and Froment introduced the total-variation (TV) minimization approach to the artifact-free JPEG decoding. They formulated the decoding problem as the constrained TV restoration problem, in which the TV semi-norm of its restored color image is minimized under the constraint that each DCT coefficient of the restored color image should be in the quantization interval of its corresponding DCT coefficient of the JPEG-compressed data. This paper proposes a new restoration approach to the JPEG decoding. Instead of the TV regularization, our new JPEG-decoding method employs a shrinkage operation in the redundant DCT domain, to mitigate degradations caused by the JPEG coding.

P1-25
TitleImage Interpolation via Regularized Local Linear Regression
AuthorXianming Liu, Debin Zhao (Harbin Institute of Technology, China), Ruiqin Xiong, Siwei Ma, Wen Gao (Peking University, China)
Pagepp. 118 - 121
KeywordImage interpolation, regularized local linear regression, edge preservation
AbstractIn this paper, we present an efficient image interpolation scheme by using regularized local linear regression (RLLR). On one hand, we introduce a robust estimator of local image structure based on moving least squares, which can efficiently handle the statistical outliers compared with ordinary least squares based methods. On the other hand, motivated by recent progress on manifold based semi-supervise learning, the intrinsic manifold structure is explicitly considered by making use of both measured and unmeasured data points. In particular, the geometric structure of the marginal probability distribution induced by unmeasured samples is incorporated as an additional locality preserving constraint. The optimal model parameters can be obtained with a closed-form solution by solving a convex optimization problem. Experimental results demonstrate that our method outperform the existing methods in both objective and subjective visual quality over a wide range of test images.

P1-26
TitleFast and Efficient Gaussian Noise Image Restoration Algorithm by Spatially Adaptive Filtering
AuthorTuan Anh Nguyen, Myoung Jin Kim, Min Cheol Hong (Soongsil University, Republic of Korea)
Pagepp. 122 - 125
Keywordnoise, detection, removal, constraint, smoothness
AbstractIn this paper, we propose a spatially adaptive noise removal algorithm using local statistics that consists of two stages: noise detection and removal. To corporate desirable properties into denoising process, the local weighted mean, local weighted activity, and local maximum are defined. With these local statistics, the noise detection function is defined and a modified Gaussian filter is used to suppress the detected noise components. The experimental results demonstrate the effectiveness of the proposed algorithm.

P1-27
TitleImage Simplification by Frequency-Selective Means Filtering
AuthorJohannes Ballé (RWTH Aachen University, Germany)
Pagepp. 126 - 129
Keywordnarrowband filters, non-local means filtering, phase congruency
AbstractIn this paper, we present an algorithm to remove high-frequent texture and detail from images. The algorithm effectively removes texture regardless of its contrast without destroying high-level image structure or introducing artificial edges. While based on the same general framework, this “image simplification” filter differs from noise filtering methods such as bilateral filtering by its frequency-selectivity and edge awareness. Applications include artistic filtering of images (edge-preserving smoothing), image reconstruction and preprocessing for low-bitrate image compression.

P1-28
TitleTheoretical Analysis of Trend Vanishing Moments for Directional Orthogonal Transforms
AuthorShogo Muramatsu, Dandan Han, Tomoya Kobayashi, Hisakazu Kikuchi (Niigata University, Japan)
Pagepp. 130 - 133
KeywordM-D filter banks, M-D Wavelets, Directional transforms, Nonseparable filter design, Image coding
AbstractThis work contributes to investigate theoretical properties of the trend vanishing moments (TVMs) which the authors have defined in a previous work and applied to the directional design of 2-D nonseparable GenLOT. Some significant properties of TVMs are shown theoretically and experimentally.

P1-29
TitleA Study on Memorability and Shoulder-Surfing Robustness of Graphical Password Using DWT-Based Image Blending
AuthorTakao Miyachi, Keita Takahashi, Madoka Hasegawa, Yuichi Tanaka, Shigeo Kato (Utsunomiya University, Japan)
Pagepp. 134 - 137
KeywordGraphical password, discrete wavelet transform, authentication, usable security
AbstractWe propose a graphical password method which is difficult to steal original pass-image by using characteristics of human vision system. In our method, we combine low frequency components of a decoy picture with high frequency components of a pass-image. It is easy for legitimate users to recognize the pass-image in the blended image. On the other hand, this task is difficult for attackers. We used discrete wavelet transform (DWT) to blend a decoy image and a pass-image. User studies are conducted to evaluate memorability and shoulder-surfing robustness of this method.

P1-30
TitleScalable Image Scrambling Method Using Unified Constructive Permutation Function on Diagonal Blocks
AuthorKokSheik Wong (University of Malaya, Malaysia), Kiyoshi Tanaka (Shinshu University, Japan)
Pagepp. 138 - 141
Keywordscalable scrambling, UCPF, diagonal block, image encryption
AbstractIn this paper, an extension of ScaScra [1] is proposed to scalably scramble an image in the diagonal direction for achieving distorted scanline-like effect. The non-overlapping diagonal blocks are first defined and the unified constructive permutation function is applied to scramble pixels in each diagonal block. Scalability in scrambling is achieved by varying the block size. Experiments were carried out to objectively and subjectively verify the basic performance of the proposed extension and compare them to the results of ScaScra by using standard test images. Evaluations on pixel correlation and entropy are also carried out to verify the performance of both ScaScra and the proposed extension.


[Quality, system, applications, and other topics]

P1-31
TitleMemory-Efficient Parallelization of JPEG-LS with Relaxed Context Update
AuthorSimeon Wahl, Zhe Wang, Chensheng Qiu, Marek Wroblewski, Lars Rockstroh, Sven Simon (University of Stuttgart, Germany)
Pagepp. 142 - 145
KeywordJPEG-LS, Lossless Image Coding, Parallelization, Context Update
AbstractA relaxation to the context update of JPEG-LS by delaying the update procedure is proposed, in order to achieve a guaranteed degree of parallelism with a negligible effect on the compression ratio. The lossless mode of JPEG-LS including the run-mode is considered. A descewing scheme is provided generating a bit-stream that preserves the order needed for the decoder to mimic the prediction in a consistent way.

P1-32
TitleH.264 Hierarchical P Coding in the Context of Ultra-Low Delay, Low Complexity Applications
AuthorDanny Hong, Michael Horowitz, Alexandros Eleftheriadis, Thomas Wiegand (Vidyo, Inc., U.S.A.)
Pagepp. 146 - 149
KeywordH.264, hierarchical P coding
AbstractDespite the attention that hierarchical B picture coding has received, little attention has been given to a related technique called hierarchical P picture coding. P picture only coding without reverse prediction is necessary in constrained bit rate applications that require ultra-low delay and/or low complexity, such as videoconferencing. Such systems, however, have been using the traditional IPPP picture coding structure almost exclusively. In this paper, we investigate the use of hierarchical P coding vs. traditional IPPP coding and demonstrate that it has significant advantages which have not yet been well documented or understood. From a pure coding efficiency point of view, we show that for encoders configured to use ultra-low delay and low complexity coding tools, hierarchical P coding achieves an average advantage of 7.86% BD-rate and 0.34 dB BD-SNR.

P1-33
TitleCorrelation Modeling with Decoder-Side Quantization Distortion Estimation for Distributed Video Coding
AuthorJozef Skorupa, Jan De Cock, Jürgen Slowack, Stefaan Mys (Ghent University -- IBBT, Belgium), Nikos Deligiannis (Vrije Universiteit Brussel -- IBBT, Belgium), Peter Lambert (Ghent University -- IBBT, Belgium), Adrian Munteanu (Vrije Universiteit Brussel -- IBBT, Belgium), Rik Van de Walle (Ghent University -- IBBT, Belgium)
Pagepp. 150 - 153
Keyworddistributed video coding, correlation modeling, quantization distortion
AbstractAiming for low-complexity encoding, distributed video coders still fail to achieve the performance of current industrial standards for video coding. One of most important problems in this area is the accurate modeling of the correlation between the predicted signal and the original video. In our previous work we showed that exploiting the quantization distortion can significantly improve the accuracy of a correlation estimator. In this paper we describe how the quantization distortion can be exploited purely at the decoder side without any performance penalty when compared to an encoder-aided system. As a result, the proposed correlation estimator delivers state-of-the-art modeling accuracy while neatly fitting the low-encoder-complexity characteristic of distributed video coding.

P1-34
TitleContent-Based Retrieval by Multiple Image Examples for Sign Board Retrieval
AuthorAtsuo Yoshitaka (JAIST, Japan), Terumasa Hyoudou (Hiroshima University, Japan)
Pagepp. 154 - 157
KeywordMultiple image examples, QBE, image retrieval
AbstractIn the area of retrieving image databases, one of the promising approaches is to retrieve it by specifying image example. However, specifying a single image example is not always sufficient to get satisfactory result, since one image example does not give comprehensive ranges of values that reflect the various aspects of the object to be retrieved. In this paper, we propose a method of retrieving images by specifying multiple image examples that is designed for retrieving sign boards. Features of color, shape, and spatial relation of color regions are extracted from example images, and they are clustered so as to obtain proper range of values. Compared with QBE systems that accept only a single image as the query condition, MIERS (Multi-Image Example-based Retrieval System) returns better retrieval result, where the experimental result showed that specifying more examples helps to improve recall with little deterioration of precision.



Thursday, December 9, 2010

Session K2  Keynote Speech 2
Time: 8:45 - 9:30 Thursday, December 9, 2010
Chair: Kiyoharu Aizawa (University of Tokyo, Japan)

K2-1 (Time: 8:45 - 9:30)
Title(Keynote Speech) Decoding Visual Perception from Human Brain Activity
AuthorYukiyasu Kamitani (ATR Computational Neuroscience Laboratories, Japan)
AbstractIn modern neuroscience, brain activity is considered as "codes" that encode mental and behavioral contents. Recent advances in human neuroimaging, in particular functional magnetic resonance imaging (fMRI), have revealed brain regions that appear to encode specific behavior and cognition. Despite the wide-spread use of human neuroimaging in such "functional brain mapping", its potential to read out, or "decode", mental contents from brain activity has not been fully explored. Mounting evidence from animal neurophysiology has revealed the roles of the early visual cortex in representing visual features such as orientation and motion direction. However, non-invasive neuroimaging methods have been thought to lack the resolution to probe into these putative feature representations in the human brain. In this talk, I present methods for decoding early visual representations from fMRI voxel patterns based on machine learning techniques. I first show how early visual features represented in "subvoxel" neural structures could be decoded from ensemble fMRI responses. Decoding of stimulus features is extended to the method for neural mind-reading, which attempts to predict a person's subjective state using a decoder trained with unambiguous stimulus presentation. We discuss how a multivoxel pattern can represent more information than the sum of individual voxels, and how an effective set of voxels for decoding can be selected from all available voxels. Based on these methods, we have recently proposed a modular decoding approach, in which a wide variety of percepts can be predicted by combining the outputs of multiple modular decoders. I show an example of visual image reconstruction where arbitrary visual images can be accurately reconstructed by the decoding model trained on fMRI responses to several hundred random images. Finally, I discuss potential applications of neural decoding for brain-based communications.


Session P2  Poster Session 2
Time: 9:30 - 11:00 Thursday, December 9, 2010
Chair: Kazunori Kotani (Japan Advanced Institute of Science and Technology, Japan)


[3DTV/FTV/multi-view-related topics]

P2-1
TitleFocus on Visual Rendering Quality through Content-Based Depth Map Coding
AuthorEmilie Bosc, Luce Morin, Muriel Pressigout (INSA of Rennes, France)
Pagepp. 158 - 161
Keyword3D video coding, adaptive coding, depth coding
AbstractMulti-view video plus depth (MVD) data is a set of multiple sequences capturing the same scene at different viewpoints, with their associated per-pixel depth value. Overcoming this large amount of data requires an effective coding framework. Yet, a simple but essential question refers to the means assessing the proposed coding methods. While the challenge in compression is the optimization of the rate-distortion ratio, a widely used objective metric to evaluate the distortion is the Peak-Signal-to-Noise-Ratio (PSNR), because of its simplicity and mathematically easiness to deal with such purposes. This paper points out the problem of reliability, concerning this metric, when estimating 3D video codec performances. We investigated the visual performances of two methods, namely H.264/MVC and Locally Adaptive Resolution (LAR) method, by encoding depth maps and reconstructing existing views from those degraded depth images. The experiments revealed that lower coding efficiency, in terms of PSNR, does not imply a lower rendering visual quality and that LAR method preserves the depth map properties correctly.

P2-2
TitleBit Allocation of Vertices and Colors for Patch-Based Coding in Time-Varying Meshes
AuthorToshihiko Yamasaki, Kiyoharu Aizawa (The University of Tokyo, Japan)
Pagepp. 162 - 165
Keyword3DTV, Time-varying mesh (TVM), inter-frame compression, vector quantization (VQ)
AbstractThis paper discusses the optimal bit rates assignment for vertices and color and for reference frames (I frames) and target frames (P frames) in the patch-based compression method for Time-Varying Meshes (TVMs). TMVs are non-isomorphic 3D mesh sequences generated from multi-view images. Experimental results demonstrated that, the bit rate for vertices affects the visual quality of the rendered 3D model very much whereas that for color does not contribute to the quality improvement. Therefore, as many bits as possible should be assigned to vertices and 8-10 bits per vertex (bpv) is enough for color. In inter-frame coding, the bit rates for the target frames improves the visual quality proportionally but at the same time it is demonstrated that less bits (5~6 bpv) are enough to achieve the same visual quality as the intra-frames.

P2-3
TitleMotion Activity-Based Block Size Decision for Multi-view Video Coding
AuthorHuanqiang Zeng, Kai-Kuang Ma (School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore), Canhui Cai (Institute of Information Science and Technology, Huaqiao University, China)
Pagepp. 166 - 169
Keywordmulti-view video coding, motion estimation, disparity estimation, block size decision, motion activity
AbstractMotion estimation and disparity estimation using variable block sizes have been exploited in multi-view video coding to effectively improve the coding efficiency, but at the expense of yielding higher computational complexity. In this paper, a fast block size decision algorithm, called motion activity-based block size decision (MABSD), is proposed. In our approach, the various motion estimation and disparity estimation block sizes are classified into four classes, and only one of them will be chosen to further identify the optimal block size within that class according to the measured motion activity of the current macroblock. The above-mentioned motion activity can be measured by the maximum city-block distance of a set of motion vectors taken from the adjacent macroblocks in the current view and its neighboring view. Experimental results have shown that compared with exhaustive block size decision, which is a default approach set in the JMVM reference software, the proposed MABSD algorithm achieves a reduction of computational complexity by 42% on average, while incurring only 0.01 dB loss in peak signal-to-noise ratio (PSNR) and 1% increment on the total bit rate.

P2-4
TitleColor Based Depth Up-Sampling for Depth Compression
AuthorMeindert Onno Wildeboer, Tomohiro Yendo, Mehrdad Panahpour Tehrani (Nagoya University, Japan), Toshiaki Fujii (Tokyo Institute of Technology, Japan), Masayuki Tanimoto (Nagoya University, Japan)
Pagepp. 170 - 173
Keyworddepth up-sampling, depth coding, depth-map, FTV, 3DTV
Abstract3D scene information can be represented in several ways. In applications based on a (N-)view plus (N-)depth representation, both view and depth data is compressed. In this paper we present a depth compression method containing an depth up-sample filter which uses the color view as prior. Our method of depth down-/up-sampling is able to maintain clear object boundaries in the reconstructed depth maps. Our experimental results show that the proposed depth re-sampling filter, used in combination with a standard state-of-the art video encoder, can increase both the coding efficiency and rendering quality.

P2-5
TitleEfficient Free Viewpoint Video-On-Demand Scheme Realizing Walk-Through Experience
AuthorAkio Ishikawa, Hiroshi Sankoh, Sei Naito, Shigeyuki Sakazawa (KDDI R&D Laboratories Inc., Japan)
Pagepp. 174 - 177
KeywordFreeviewpoint Television, Multi-view Video, Freeviewpoint Video Transmission, Walk-through, Multi-texturing
AbstractThis paper presents an efficient VOD scheme for FTV, and proposes a data format and its data generation method to provide a walk-through experience. We employ a hybrid rendering approach to describe a 3D scene using 3D model data for objects and textures. In this paper we propose an efficient texture data format, which removes the redundancy due to occlusion of objects by employing an orthogonal projection image for each object. The advantage of the data format is great simplification at the server to choose the transmitted images that correspond to the requested viewpoint.

P2-6
Title3-D Video Coding Using Depth Transition Data
AuthorWoo-Shik Kim, Antonio Ortega (University of Southern California, U.S.A.), Jaejoon Lee, HoCheon Wey (Samsung Electronics Co., Ltd., Republic of Korea)
Pagepp. 178 - 181
Keyword3-D video coding, multiview plus depth, view synthesis
AbstractThe objective is to develop new 3-D video coding system to provide better coding efficiency with improved subjective quality. We analyzed rendered view distortions in DIBR system, and found that the depth map coding distortion leads to “erosion artifacts”, which lead to significant perceptual quality degradation. To solve this, we propose a solution using depth transition data which indicates the camera position where depth changes. Simulation results show that significant subjective quality improvement with maximum PSNR gains of 0.5 dB.

P2-7
TitleSubjective Assessment of Frame Loss Concealment Methods in 3D Video
AuthorJoao Carreira, Luis Pinto, Nuno Rodrigues, Sergio Faria, Pedro Assuncao (Institute of Telecommunications / Polytechnic Institute of Leiria - ESTG, Portugal)
Pagepp. 182 - 185
Keyword3D video, Frame loss concealment, Subjective Assessment
AbstractThis paper investigates the subjective impact resulting from different concealment methods for coping with lost frames in 3D video communication systems. It is assumed that a high priority channel is assigned to the main view and only the auxiliary view is subject to either transmission errors or packet loss, leading to missing frames at the decoder output. Three methods are used for frame concealment under different loss ratios. The results show that depth is well perceived by users and the subjective impact of frame loss not only depends on the concealment method but also exhibits high correlation with the disparity of the original sequence. It is also shown that under heavy loss conditions it is better to switch from 3D to 2D rather than presenting concealed 3D video to users.

P2-8
TitleA Novel Upsampling Scheme for Depth Map Compression in 3DTV System
AuthorYanjie Li, Lifeng Sun (Tsinghua University, China)
Pagepp. 186 - 189
KeywordDepth map compression, Resolution reduction, Upsampling, Rate-distortion
AbstractIn this paper, we propose a novel two-step depth map upsampling scheme to address the problem on 3D videos. The first step utilizes the full resolution 2D color map to help reconstruct a more accurate full resolution depth map. And the second step further flats the reconstructed depth map to ensure its local uniformity. Test results show that the proposed upsampling scheme achieves up to 2dB coding gains for the rendering of free-viewpoint video, and improves its perceptual quality significantly.


[Beyond H.264/MPEG-4 AVC and related topics]

P2-9
TitleAdaptive Direct Vector Derivation for Video Coding
AuthorYusuke Itani, Shunichi Sekiguchi, Yoshihisa Yamada (Information Technology R&D Center, Mitsubishi Electric Corporation, Japan)
Pagepp. 190 - 193
Keyworddirect mode, motion vector predictor, HEVC, H.264, extended macroblock
AbstractThis paper proposes a new method for improving direct prediction scheme that has been employed in conventional video coding standards such as AVC/H.264. We extend direct prediction concept to achieve better adaptation to local statistics of video source with the assumption of the use of larger motion blocks than conventional macroblock size. Experimental results show the proposed method provides up to 3.3% bitrate saving in low-bitrate coding.

P2-10
TitleInter Prediction Based on Spatio-Temporal Adaptive Localized Learning Model
AuthorHao Chen, Ruimin Hu, Zhongyuan Wang, Rui Zhong (Wuhan University, China)
Pagepp. 194 - 197
KeywordInter prediction, STALL, LSP
AbstractInter prediction based on block matching motion estimation is important for video coding. But this method suffers from the additional overhead in data rate representing the motion information that needs to be transmitted to the decoder. To solve this problem, we present an improved implicit motion information inter prediction algorithm for P slice in H.264/AVC based on the spatio-temporal adaptive localized learning (STALL) model. According to 4*4 block transform structure in H.264/AVC, we first adaptively choose nine spatial neighbors and nine temporal neighbors, and a localized 3D casual cube is designed as training window. By using these information, the model parameters could be adaptively computed based on the Least Square Prediction (LSP) method. Finally, we add a new inter prediction mode into H.264/AVC standard for P slice. The experimental results show that our algorithm improves encoding efficiency compared with H.264/AVC standard, with relatively increases in complexity.

P2-11
TitleIntra Picture Coding with Planar Representations
AuthorJani Lainema, Kemal Ugur (Nokia Research Center, Finland)
Pagepp. 198 - 201
Keywordvideo coding, intra coding, planar coding, H.264/AVC, HEVC
AbstractIn this paper we introduce a novel concept for Intra coding of pictures especially suitable for representing smooth image segments. Traditional block based transform coding methods cause visually annoying blocking artifacts for image segments with gradually changing smooth content. The proposed solution overcomes this drawback by defining a fully continuous surface of sample values approximating the original image. The gradient of the surface is indicated by transmitting values for selected control points within the image segment and the surface itself is obtained by interpolating sample values in-between the control points. This approach is found to provide up to 30 percent bitrate reductions in the case of natural imagery and it has also been adopted to the initial HEVC codec design by JCT-VC.

P2-12
TitleAdaptive Global Motion Temporal Prediction for Video Coding
AuthorAlexander Glantz, Andreas Krutz, Thomas Sikora (Technische Universität Berlin, Germany)
Pagepp. 202 - 205
KeywordH.264/AVC, video coding, global motion, temporal filtering, prediction
AbstractDepending on the content of a video sequence, the amount of bits spent for the transmission of motion vectors can be enormous. A global motion model can be a better representation of movement in these regions than a motion vector. This paper presents a novel prediction technique that is based on global motion compensation and temporal filtering. The new approach is incorporated into H.264/AVC and outperforms the reference by up to 14%.

P2-13
TitleHighly Efficient Video Compression Using Quadtree Structures and Improved Techniques for Motion Representation and Entropy Coding
AuthorDetlev Marpe, Heiko Schwarz, Thomas Wiegand, Sebastian Bosse, Benjamin Bross, Philipp Helle, Tobias Hinz, Heiner Kirchhoffer, Haricharan Lakshman, Tung Nguyen, Simon Oudin, Mischa Siekmann, Karsten Suehring, Martin Winken (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany)
Pagepp. 206 - 209
KeywordVideo coding, H.265, HEVC
AbstractThis paper describes a novel video coding scheme that can be considered as a generalization of the block-based hybrid video coding approach of H.264/AVC. While the individual building blocks of our approach are kept simple similarly as in H.264/AVC, the flexibility of the block partitioning for prediction and transform coding has been substantially in-creased. This is achieved by the use of nested and pre-configurable quadtree structures, such that the block parti-tioning for temporal and spatial prediction as well as the space-frequency resolution of the corresponding prediction residual can be adapted to the given video signal in a highly flexible way. In addition, techniques for an improved motion representation as well as a novel entropy coding concept are included. The presented video codec was submitted to a Call for Proposals of ITU-T VCEG and ISO/IEC MPEG and was ranked among the five best performing proposals, both in terms of subjective and objective quality.


[Image/video coding and related topics]

P2-14
TitleDictionary Learning-Based Distributed Compressive Video Sensing
AuthorHung-Wei Chen, Li-Wei Kang, Chun-Shien Lu (Academia Sinica, Taiwan)
Pagepp. 210 - 213
Keywordcompressive sensing, sparse representation, dictionary learning, single-pixel camera, l1-minimization
AbstractWe address an important issue of fully low-cost and low-complex video compression for use in resource-extremely limited sensors/devices. Conventional motion estimation-based video compression or distributed video coding (DVC) techniques all rely on the high-cost mechanism, namely, sensing/sampling and compression are disjointedly performed, resulting in unnecessary consumption of resources. That is, most acquired raw video data will be discarded in the (possibly) complex compression stage. In this paper, we propose a dictionary learning-based distributed compressive video sensing (DCVS) framework to “directly” acquire compressed video data. Embedded in the compressive sensing (CS)-based single-pixel camera architecture, DCVS can compressively sense each video frame in a distributed manner. At DCVS decoder, video reconstruction can be formulated as an l1-minimization problem via solving the sparse coefficients with respect to some basis functions. We investigate adaptive dictionary/basis learning for each frame based on the training samples extracted from previous reconstructed neighboring frames and argue that much better basis can be obtained to represent the frame, compared to fixed basis-based representation and recent popular “CS-based DVC” approaches without relying on dictionary learning.

P2-15
TitleMedium-Granularity Computational Complexity Control for H.264/AVC
AuthorXiang Li, Mathias Wien, Jens-Rainer Ohm (Institute of Communications Engineering, RWTH Aachen University, Germany)
Pagepp. 214 - 217
KeywordComputational complexity control, computational scalability
AbstractToday, video applications on handheld devices become more and more popular. Due to limited computational capability of handheld devices, complexity constrained video coding draws much attention. In this paper, a medium-granularity computational complexity control (MGCC) is proposed for H.264/AVC. First, a large dynamic range in complexity is achieved by taking 16x16 motion estimation in a single reference frame as the basic computational unit. Then a high coding efficiency is obtained by an adaptive computation allocation at MB level. Simulations show that coarse-granularity methods cannot work when the normalized complexity is below 15\%. In contrast, the proposed MGCC performs well even when the complexity is reduced to 8.8\%. Moreover, an average gain of 0.3 dB over coarse-granularity methods in BD-PSNR is obtained for 11 sequences when the complexity is around 20\%.

P2-16
TitleAn Improved Wyner-Ziv Video Coding With Feedback Channel
AuthorFeng Ye, Aidong Men, Bo Yang, Manman Fan, Kan Chang (School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, China)
Pagepp. 218 - 221
KeywordWyner-Ziv video coding, motion activity, 3DRS motion estimation, side information
AbstractThis paper presents a improved feedback-assisted low complexity WZVC scheme. The performance of this scheme is improved by two enhancements: an improved mode-based key frame encoding and a 3DRS-assisted (three-dimensional recursive search assisted) motion estimation algorithm for WZ encoding. Experimental results show that our coding scheme can achieve significant gain compared to state-of-the-art TDWZ codec while still low encoding complexity.

P2-17
TitleBackground Aided Surveillance-Oriented Distributed Video Coding
AuthorHongbin Liu (Harbin Institute of Technology, China), Siwei Ma (Peking University, China), Debin Zhao (Harbin Institute of Technology, China), Wen Gao (Peking University, China), Xiaopeng Fan (Harbin Institute of Technology, China)
Pagepp. 222 - 225
Keywordsurveillance, background, distributed video coding
AbstractThis paper presents a background aided surveillance-oriented distributed video coding system. A high quality background frame is encoded for each group of pictures (GOP), which can provides high quality SI for the background parts of the Wyner-Ziv (WZ) frames. Consequently, bit rate for the WZ frames can be reduced. Experimental results demonstrate that the proposed system can decrease the bit rate by up to 67.4% when compared with traditional DVC codec.

P2-18
TitleContent-Adaptive Spatial Scalability for Scalable Video Coding
AuthorYongzhe Wang (Shanghai Jiao Tong University, China), Nikolce Stefanoski (Disney Research Zurich, Switzerland), Xiangzhong Fang (Shanghai Jiao Tong University, China), Aljoscha Smolic (Disney Research Zurich, Switzerland)
Pagepp. 226 - 229
KeywordH.264/AVC, scalable video coding, spatial scalability, content-adaptation, non-linear image warping
AbstractThis paper presents an enhancement of the SVC extension of the H.264/AVC standard by content-adaptive spatial scalability (CASS). The video streams (spatial layers), which are used as input to the encoder, are created by content-adaptive and art-directable retargeting of existing high resolution video. The non-linear dependencies between such video streams are efficiently exploited by CASS for scalable coding. This is achieved by integrating warping-based non-linear texture prediction and warp coding into the SVC framework.

P2-19
TitleColorization-Based Coding by Focusing on Characteristics of Colorization Bases
AuthorShunsuke Ono, Takamichi Miyata, Yoshinori Sakai (Tokyo Institute of Technology, Japan)
Pagepp. 230 - 233
KeywordColorization, Colorization-based coding, Representative pixels, Redundancy, Correct color
AbstractA novel approach to image compression called colorization-based coding has recently been proposed. It automatically extracts representative pixels from an original color image at an encoder and restores a full color image by using colorization at a decoder. However, previous studies on colorization-based coding extract redundant representative pixels. We propose a new colorization-based coding method focuses on the colorization basis. Experimental results revealed that our method can drastically suppress the information amount compared conventional method while objective quality is maintained.

P2-20
TitleWyner-Ziv Coding of Multispectral Images for Space and Airborne Platforms
AuthorShantanu Rane, Yige Wang, Petros Boufounos, Anthony Vetro (Mitsubishi Electric Research Laboratories, U.S.A.)
Pagepp. 234 - 237
KeywordMultispectral, Wyner-Ziv coding, LDPC code
AbstractThis paper investigates the application of lossy distributed source coding to high resolution multispectral images. The choice of distributed source coding is motivated by the need for very low encoding complexity on space and airborne platforms. The data consists of red, blue, green and infra-red channels and is compressed in an asymmetric Wyner-Ziv setting. One image channel is compressed using traditional JPEG and transmitted to the ground station where it is available as side information for Wyner-Ziv coding of the other channels. Encoding is accomplished by quantizing the image data, applying a Low-Density Parity Check code to the remaining three image channels, and transmitting the resulting syndromes. At the ground station, the image data is recovered from the syndromes by exploiting the correlation in the frequency spectrum of the band being decoded and the JPEG-decoded side information band. In experiments with real uncompressed images obtained by a satellite, the rate-distortion performance is found to be vastly superior to JPEG compression of individual image channels and rivals that of JPEG2000 at much lower encoding complexity.

P2-21
TitleReversible Component Transforms by the LU Factorization
AuthorHisakazu Kikuchi, Junghyeun Hwang, Shogo Muramatsu (Niigata University, Japan), Jaeho Shin (Dongguk University, Republic of Korea)
Pagepp. 238 - 241
Keywordcomponent transform, image compression, LU factorization, lifting, round-off error
AbstractA scaled transform is defined for a given irreversible linear transformation based on the LU factorization of a nonsingular matrix so that the transformation may be computed in a lifting form and hence may be reversible. Round-off errors in the lifting computation and the computational complexity are analyzed. Some reversible component transforms are presented and experimented to give some remarks to image compression applications. Discussions are developed with coding gain and actual bit rates.

P2-22
TitleMOS-Based Bit Allocation in SNR-Temporal Scalable Coding
AuthorYuya Yamasaki, Toshiyuki Yoshida (University of Fukui, Japan)
Pagepp. 242 - 245
Keywordscalable coding, video coding, MOS, bit allocation
AbstractIn a scalable video coding, improvement in the motion smoothness (frame rate) and in the spatial quality (SNR) conflicts within a given bit rate. Since the motion and the spatial activities in a target video varies from scene to scene, these activities should be taken into account in order to optimally allocate the temporal and quality scalability. This paper proposes a fundamental idea for allocating bit rates in the temporal and quality scalability based on the maximization of the estimated mean opinion score (MOS) for each scene. A reduction technique of MOS fluctuation in the enhancement layer is discussed as an application of the proposed technique.


[Image/video processing and related topics]

P2-23
TitleAutomatic Moving Object Extraction Using X-means Clustering
AuthorKousuke Imamura, Naoki Kubo, Hideo Hashimoto (Kanazawa University, Japan)
Pagepp. 246 - 249
Keywordmoving object extraction, x-means clustering, watershed algorithm, voting method
AbstractThe present paper proposes an automatic extraction technique of moving objects using x-means clustering. The proposed technique is an extended k-means clustering and can determine the optimal number of clusters based on the Bayesian Information Criterion(BIC). In the proposed method, the feature points are extracted from a current frame, and x-means clustering classifies the feature points based on their estimated affine motion parameters. A label is assigned to the segmented region, which is obtained by morphological watershed, by voting for the feature point cluster in each region. The labeling result represents the moving object extraction. Experimental results reveal that the proposed method provides extraction results with the suitable object number.

P2-24
TitleAccurate Motion Estimation for Image of Spatial Periodic Pattern
AuthorJun-ichi Kimura, Naohisa Komatsu (School of Fundamental Science and Engineering, Waseda University, Japan)
Pagepp. 250 - 253
Keywordmotion estimation, motion vector
AbstractWe investigate the mechanism of the error of motion estimation for image including spatial periodic patterns. With a motion estimation model, we conclude that block matching distortion caused by motion vector sampling error (DVSE) degrades the motion estimation accuracy. We propose a new motion estimation using a maximum value of DVSE for each block. Simulations show that precision of proposed method is over 98% which are superior to full search method.

P2-25
TitleDirection-Adaptive Image Upsampling Using Double Interpolation
AuthorYi-Chun Lin, Yi-Nung Liu, Shao-Yi Chien (National Taiwan University, Taiwan)
Pagepp. 254 - 257
Keyworddouble interpolation, upsampling, direction-adaptive, zigzagging, bicubic
AbstractDouble interpolation quality evaluation can be used as a measurement of an interpolation operation. By using this double interpolation framework, a high efficient direction-adaptive upsampling algorithm is proposed without any threshold setting and post-processing. With the proposed upsampling algorithm, the problem of zigzagging artifacts on the edge no longer exists. Moreover, the proposed algorithm has low computation complexity. The experimental results show that the proposed algorithm has high quality image output.

P2-26
TitleBlind GOP Structure Analysis of MPEG-2 and H.264/AVC Decoded Video
AuthorGilbert Yammine, Eugen Wige, Andre Kaup (Multimedia Communications and Signal Processing - University of Erlangen-Nuremberg, Germany)
Pagepp. 258 - 261
KeywordGOP Structure, Blind Analysis, Noise Estimation
AbstractIn this paper, we provide a simple method for analyzing the GOP structure of an MPEG-2 or H.264/AVC decoded video without having access to the bitstream. Noise estimation is applied on the decoded frames and the variance of the noise in the different I-, P-, and B-frames is measured. After the encoding process, the noise variance in the video sequence shows a periodic pattern, which helps in the extraction of the GOP period, as well as the type of frames. This algorithm can be used along with other algorithms to blindly analyze the encoding history of a video sequence. The method has been tested on several MPEG-2 DVB and DVD streams, as well as on H.264/AVC encoded sequences, and shows successful results in both cases.

P2-27
TitleDistributed Video Coding Based on Adaptive Slice Size Using Received Motion Vectors
AuthorKyungyeon Min, Seanae Park, Donggyu Sim (Kwangwoon University, Republic of Korea)
Pagepp. 262 - 265
KeywordDVC, crossover, adaptive slice control
AbstractIn this paper, we propose a new distributed video coding (DVC) method based on adaptive slice size using received motion vectors (MVs). In the proposed algorithm, the MVs estimated at a DVC decoder are transmitted to a corresponding encoder. In the proposed encoder, a predicted side information (PSI) is reconstructed with the transmitted MVs and key frames. Therefore, the PSI can be generated same to side information (SI) at the decoder. We can, also, calculate an exact crossover rate between the SI and original input frame using PSI and the original frame. As a result, the proposed method can transmit minimum parity bits to maximize error correction ability of a channel decoder with minimal computational complexity. Experimental results show that the proposed algorithm is better than several conventional DVC methods.

P2-28
TitleImproved Watermark Sharing Scheme Using Minimum Error Selection and Shuffling
AuthorAroba Khan, Yohei Yokoyama, Kiyoshi Tanaka (Shinshu University, Japan)
Pagepp. 266 - 269
KeywordWatermark Sharing, Halftoning, error diffussion
AbstractIn this work, we focus on a watermark sharing scheme using error diffusion called DHCED, and try to overcome some drawbacks of this method. The proposed method simultaneously generates carrier halftone images that share the watermark information by selecting the minimum error caused in the noise function for watermark embedding. Also, the proposed method shuffles watermark image before embedding not only to increase the secrecy of the embedded watermark information but also improve the watermark detection ratio as well as the watermark appearance in the detection process.


[Quality, system, applications, and other topics]

P2-29
TitleOn the Duality of Rate Allocation and Quality Indices
AuthorThomas Richter (University of Stuttgart, Germany)
Pagepp. 270 - 273
KeywordJPEG 2000, SSIM
AbstractIn a recent work, the author proposed to study the performance of still image quality indices such as the SSIM by using them as bjective function of rate allocation algorithms. The outcome of that work was not only a multi-scale SSIM optimal JPEG 2000 implementation, but also a first-order approximation of the MS-SSIM that is surprisingly similar to more traditional contrast-sensitivity and visual masking based approaches. It will be seen in this work that the only difference between the latter works and the MS-SSIM index is the choice of the exponent of the masking term, and furthermore, that a slight modification of the SSIM definition reproducing the traditional exponent is able to improve the performance of the index at or below the visual threshold. It is hence demonstrated that the duality of quality indices and rate allocation helps to improve both the visual performance of the compression codec and the performance of the index.

P2-30
TitleImage Quality Assessment Based on Local Orientation Distributions
AuthorYue Wang (Graduate University of Chinese Academy of Sciences, China), Tingting Jiang, Siwei Ma, Wen Gao (Institute of Digital Media, Peking University, China)
Pagepp. 274 - 277
Keywordmage quality assessment (IQA), human visual system (HVS), Histograms of Oriented Gradients (HOG)
AbstractImage quality assessment (IQA) is very important for many image and video processing applications, e.g. compression, archiving, restoration and enhancement. An ideal image quality metric should achieve consistency between image distortion prediction and psychological perception of human visual system (HVS). Inspired by that HVS is quite sensitive to image local orientation features, in this paper, we propose a new structural information based image quality metric, which evaluates image distortion by computing the distance of Histograms of Oriented Gradients (HOG) descriptors. Experimental results on LIVE database show that the proposed IQA metric is competitive with state-of-the-art IQA metrics, while keeping relatively low computing complexity.

P2-31
TitleDistance and Relative Speed Estimation of Binocular Camera Images Based on Defocus and Disparity Information
AuthorMitsuyasu Ito, Yoshiaki Takada, Takayuki Hamamoto (Tokyo University of Science, Japan)
Pagepp. 278 - 281
KeywordITS, focus blur, disparity information, distance estimation, relative speed
AbstractIn this paper, we discuss a method of distance and relative speed estimation for ITS. In this method, we use different focus positions of two cameras for obtaining the amount of focus blur. Next, we propose the method of distance estimation by the amount and disparity information. According to the result of simulation, proposed method was reasonably. In addition, we compose a prototype system for the real-time estimation. As a result of the implementation of processing, system was properly validated.

P2-32
TitleComparing Two Eye-Tracking Databases: The Effect of Experimental Setup and Image Presentation Time on the Creation of Saliency Maps
AuthorUlrich Engelke (Blekinge Institute of Technology, Sweden), Hantao Liu (Delft University of Technology, Netherlands), Hans-Jürgen Zepernick (Blekinge Institute of Technology, Sweden), Ingrid Heynderickx (Philips Research Laboratories, Netherlands), Anthony Maeder (University of Western Sydney, Australia)
Pagepp. 282 - 285
KeywordEye tracking experiments, Visual saliency, Correlation analysis, Natural images
AbstractVisual attention models are typically designed based on human gaze patterns recorded through eye tracking. In this paper, two similar eye tracking experiments from independent laboratories are presented, in which humans observed natural images under task-free condition. The resulting saliency maps are analysed with respect to two criteria; the consistency between the experiments and the impact of the image presentation time. It is shown, that the saliency maps between the experiments are strongly correlated independent of presentation time. It is further revealed that the presentation time can be reduced without substantially sacrificing the accuracy of the convergent saliency map. The results provide valuable insight into the similarity of saliency maps from independent laboratories and are highly beneficial for the creation of converging saliency maps at reduced experimental time and cost.

P2-33
TitleSuccessive Refinement of Overlapped Cell Side Quantizers for Scalable Multiple Description Coding
AuthorMuhammad Majid, Charith Abhayaratne (The University of Sheffield, U.K.)
Pagepp. 286 - 289
KeywordMultiple description coding, scalability, robustness
AbstractScalable multiple description coding (SMDC) provides reliability and facility to truncate the descriptions according to the user rate-distortion requirements. In this paper we generalize the conditions of successive refinement of the side quantizer of a multiple description scalar quantizer that has overlapped quantizer cells generated by a modified linear index assignment matrix. We propose that the split or refinement factor for each of the refinement side quantizers should be greater than the maximum side quantizer bin spread and should not be integer multiples of each other for satisfying the SMDC distortion conditions and verify through simulation results on scalable multiple description image coding.

P2-34
TitleVQ Based Data Hiding Method for Still Images by Tree-Structured Links
AuthorHisashi Igarashi, Yuichi Tanaka, Madoka Hasegawa, Shigeo Kato (Utsunomiya University, Japan)
Pagepp. 290 - 293
KeywordData Hiding, Vector Quantization, Tree-Structured Links
AbstractIn this paper, we propose a data embedding method into still images based on Vector Quantization (VQ). Several VQ-based data embedding methods, called 'Mean Gray-Level Embedding method (MGLE)' or 'Pair wise Nearest-Neighbor Embedding method (PNNE)' have been proposed. Those methods are, however, not sufficiently effective. Meanwhile an efficient adaptive data hiding method called 'Adaptive Clustering Embedding method (ACE)' was pro-posed, but is somewhat complicated because the VQ indices have to be adaptively clustered in the embedding process. In our proposed method, output vectors are linked with tree structure and information is embedded by using some of linked vectors. The simulation results show that our proposed method indicates higher SNR than the conventional methods under the same amounts of embedded data.


Session T1  Tutorial Session 1
Time: 11:15 - 12:00 Thursday, December 9, 2010
Chair: Kenji Sugiyama (Seikei University, Japan)

T1-1 (Time: 11:15 - 12:00)
Title(Tutorial) Evolutive Video Coding - From Generic Algorithm towards Content-Specific Algorithm -
AuthorSeishi Takamura (NTT Cyber Space Laboratories, NTT Corporation, Japan)
AbstractEvolutive methods based on genetic programming (GP) enable dynamic algorithm generation, and have been successfully applied to many areas such as plant control, robot control, and stock market prediction. However, conventional image/video coding methods such as JPEG and H.264 all use fixed (non-dynamic) algorithms without exception. We have proposed a GP-based image predictor that is specifically evolved for each input image. It is a radical departure from conventional "fixed algorithm", "man-made algorithm" and "hand-made program" towards new paradigm such as "content-specific algorithm", "machine-generated algorithm/program". In this course, the predictor generation algorithm, its basic performance, application examples and some speeding up techniques of the evolution process will be presented.


Session O2  Oral Session 2: Depth Map Coding
Time: 13:15 - 15:15 Thursday, December 9, 2010
Chair: Kiyoharu Aizawa (University of Tokyo, Japan)

O2-1 (Time: 13:15 - 13:45)
TitleMultiscale Recurrent Pattern Matching Approach for Depth Map Coding
AuthorDanillo B. Graziosi (UFRJ, Brazil), Nuno M. M. Rodrigues (IT, Portugal), Carla L. Pagliari (IME, Brazil), Eduardo A. B. da Silva (UFRJ, Brazil), Sérgio M. M. de Faria (IT, Portugal), Marcelo M. Perez (IME, Brazil), Murilo B. de Carvalho (UFF, Brazil)
Pagepp. 294 - 297
KeywordDepth Maps, 3D Image Coding, Recurrent Pattern Matching
AbstractIn this article we propose to compress depth maps using a coding scheme based on multiscale recurrent pattern matching and evaluate its impact on depth image based rendering (DIBR). Depth maps are usually converted into gray scale images and compressed like a conventional luminance signal. However, using traditional transform-based encoders to compress depth maps may result in undesired artifacts at sharp edges due to the quantization of high frequency coefficients. The Multiscale Multidimensional Parser (MMP) is a pattern matching-based encoder, that is able to preserve and efficiently encode high frequency patterns, such as edge information. This ability is critical for encoding depth map images. Experimental results for encoding depth maps show that MMP is much more efficient in a rate-distortion sense than standard image compression techniques such as JPEG2000 or H.264/AVC. In addition, the depth maps compressed with MMP generate reconstructed views with a higher quality than all other tested compression algorithms.

O2-2 (Time: 13:45 - 14:15)
TitleSparse Representation of Depth Maps for Efficient Transform Coding
AuthorGene Cheung (National Institute of Informatics, Japan), Akira Kubota (Chuo University, Japan), Antonio Ortega (University of Southern California, U.S.A.)
Pagepp. 298 - 301
KeywordDepth-image-based rendering, transform coding, sparse representation
AbstractCompression of depth maps is important for "image plus depth" representation of multiview images, which enables synthesis of novel intermediate views via depth-image-based rendering (DIBR) at decoder. Previous depth map coding schemes exploit unique depth characteristics to compactly and faithfully reproduce the original signal. In contrast, given that depth maps are not directly viewed but are only used for view synthesis, in this paper we manipulate depth values themselves, without causing severe synthesized view distortion, in order to maximize sparsity in the transform domain for compression gain. We formulate the sparsity maximization problem as an l0-norm optimization. Given l0-norm optimization is hard in general, we first find a sparse representation by iteratively solving a weighted l1 minimization via linear programming (LP). We then design a heuristic to push resulting LP solution away from constraint boundaries to avoid quantization errors. Using JPEG as an example transform codec, we show that our approach gained up to 2.5dB in rate-distortion performance for the interpolated view.

O2-3 (Time: 14:15 - 14:45)
TitleA Novel Approach for Efficient Multi-View Depth Map Coding
AuthorJin Young Lee, Hochen Wey, Du-Sik Park (Samsung Advanced Institute of Technology, Samsung Electronics Co., Ltd., Republic of Korea)
Pagepp. 302 - 305
KeywordMulti-view video plus depth format, Depth map coding, Video coding
AbstractMulti-view video plus depth (MVD) format, which consists of texture and depth images, has been recently presented as video representation to support depth perception of scenes and efficient view generation at the arbitrary positions. In particular, a depth image has been one of the significantly important issues for successful services of highly advanced multi-media video applications, such as three-dimensional television (3DTV) and free-viewpoint television (FTV). In this paper, we present a novel approach for efficient multi-view depth map coding. We assume that a texture image in the MVD format is first encoded and then the corresponding depth image is encoded. According to an analysis of inter-view correlation between the previously encoded texture images, the proposed method skips some blocks of the depth image without encoding. The skipped blocks in the depth map are predicted from the neighboring depth images at the same time instant. Experimental results demonstrate that the proposed method reduces the coding bitrate of up to 74.8% and improves PSNR of up to 3.51dB in P and B views.

O2-4 (Time: 14:45 - 15:15)
TitleDiffusion Filtering of Depth Maps in Stereo Video Coding
AuthorGerhard Tech, Karsten Müller, Thomas Wiegand (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany)
Pagepp. 306 - 309
Keyworddiffusion filtering, noise reduction, stereo video, video plus depth coding
AbstractA method for removing irrelevant information from depth maps in Video plus Depth coding is presented. The depth map is filtered in several iterations using a diffusional approach. In each iteration smoothing is carried out in local sample neighborhoods considering the distortion introduced to a rendered view. Smoothing is only applied when the rendered view is not affected. Therefore irrelevant edges and features in the depth map can be damped while the quality of the rendered view is retained. The processed depth maps can be coded at a reduced rate compared to unaltered data. Coding experiments show gains up to 0.5dB for the rendered view at the same bit rate.


Session O3  Oral Session 3: New Techniques for Video Coding
Time: 15:30 - 17:30 Thursday, December 9, 2010
Chair: Yoshiyuki Yashima (Chiba Institute of Technology, Japan)

O3-1 (Time: 15:30 - 16:00)
TitleParallel Intra Prediction for Video Coding
AuthorAndrew Segall, Jie Zhao (Sharp Labs of America, U.S.A.), Tomoyuki Yamamoto (Sharp Corporation, Japan)
Pagepp. 310 - 313
Keywordintra prediction, video coding, parallel
AbstractIn this paper, we propose an intra-prediction system that is both parallel friendly and with high coding efficiency. This is achieved by combining a novel prediction strategy that reduces serial dependencies and a novel, multi-directional and adaptive prediction system. The resulting technique is compared with state-of-the-art ITU-T H.264/MPEG-4 AVC. We observe a 2x and 8x increase in parallelism for 8x8 and 4x4 partitions, respectively, and an average rate increase of less than 0.08% for predictive coding scenarios.

O3-2 (Time: 16:00 - 16:30)
TitleQuantization Noise Reduction in Hybrid Video Coding by a System of Three Adaptive Filters
AuthorMatthias Narroschke (Panasonic R&D Center Germany GmbH, Germany)
Pagepp. 314 - 317
KeywordVideo coding, Noise reduction, Quantization noise, Wiener filter
AbstractHybrid video coding algorithms, e.g. H.264/MPEG-4 AVC, apply prediction and subsequent prediction error coding introducing quantization noise. The quantized prediction error signal and the prediction signal are added for reconstruction. Deblocking filters reduce quantization noise of the reconstructed signal at block boundaries. To further reduce quantization noise, adaptive Wiener filters are applied to the deblocked signal. In this paper, the adaptive Wiener filter is extended to a system of three adaptive filters in order to improve the quantization noise reduction. A first filter is applied to the deblocked signal, a second filter to the quantized prediction error signal, and a third filter to the prediction signal. The three filtered signals are added for reconstruction. For a set of thirteen test sequences, the system of three adaptive filters achieves an average bit rate reduction at the same quality of 1.9% compared to the adaptive Wiener filter and of 4.9% compared to no Wiener filter. For particular sequences, a bit rate reduction of 6.1% is achieved compared to the adaptive Wiener filter and of 17.1% compared to no Wiener filter.

O3-3 (Time: 16:30 - 17:00)
TitleSpatio-Temporal Prediction in Video Coding by Non-Local Means Refined Motion Compensation
AuthorJürgen Seiler, Thomas Richter, André Kaup (University of Erlangen-Nuremberg, Germany)
Pagepp. 318 - 321
KeywordVideo Coding, Prediction, Signal Extrapolation
AbstractThe prediction step is a very important part of hybrid video codecs. In this contribution, a novel spatio-temporal prediction algorithm is introduced. For this, the prediction is carried out in two steps. Firstly, a preliminary temporal prediction is conducted by motion compensation. Afterwards, spatial refinement is carried out for incorporating spatial redundancies from already decoded neighboring blocks. Thereby, the spatial refinement is achieved by applying Non-Local Means denoising to the union of the motion compensated block and the already decoded blocks. Including the spatial refinement into H.264/AVC, a rate reduction of up to 14 % or respectively a gain of up to 0.7 dB PSNR compared to unrefined motion compensated prediction can be achieved.

O3-4 (Time: 17:00 - 17:30)
TitleA Novel Video Coding Scheme for Super Hi-Vision
AuthorShun-ichi Sekiguchi, Akira Minezawa, Kazuo Sugimoto (Information Technology R&D Center, Mitsubishi Electric Corporation, Japan), Atsuro Ichigaya, Kazuhisa Iguchi, Yoshiaki Shishikui (Science & Technology Research Laboratories, NHK, Japan)
Pagepp. 322 - 325
Keywordvideo coding, SHV, motion compensation prediction, intra prediction, transform
AbstractWe propose a novel video coding scheme targeting Super Hi-Vision (SHV) video sources. While it takes a conventional block-based MC + Transform hybrid coding approach that is suitable for hardware implementation of a SHV video codec, the proposed scheme achieved significant coding efficiency improvement by introducing several coding tools such as intra prediction and adaptive transform. According to our experimental analysis, the proposed scheme achieves significant bit-rate saving compared to the state-of-the-art AVC/H.264 high profile.



Friday, December 10, 2010

Session K3  Keynote Speech 3
Time: 8:45 - 9:30 Friday, December 10, 2010
Chair: Kohtaro Asai (Mitsubishi Electric Corporation, Japan)

K3-1 (Time: 8:45 - 9:30)
Title(Keynote Speech) Advances in Video Compression
AuthorThomas Wiegand (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany)
AbstractRecent work in the research and standards community has shown that significant advances are possible in the field of video compression. These gains have been demonstrated in particular for high-resolution video content. The talk will present the new techniques and analyze the reasons for the improvements in coding efficiency. Comparisons against the state-of-the-art will be provided and potential new developments will be discussed.


Session P3  Poster Session 3
Time: 9:30 - 11:00 Friday, December 10, 2010
Chair: Akira Kubota (Chuo University, Japan)


[3DTV/FTV/multi-view-related topics]

P3-1
TitleA Fast Graph Cut Algorithm for Disparity Estimation
AuthorCheng-Wei Chou, Jang-Jer Tsai, Hsueh-Ming Hang, Hung-Chih Lin (National Chiao Tung University, Taiwan)
Pagepp. 326 - 329
KeywordFTV, stereo correspondence, disparity estimation, graph cut
AbstractIn this paper, we propose a fast graph cut (GC) algorithm for disparity estimation. Two accelerating techniques are suggested: one is the early termination rule, and the other is prioritizing the alpha-beta swap pair search order. Our simulations show that the proposed fast GC algorithm outperforms the original GC scheme by 210% in the average computation time while its disparity estimation quality is almost similar to that of the original GC.

P3-2
TitleParallel Processing Method for Realtime FTV
AuthorKazuma Suzuki (Graduate School of Engineering, Nagoya University, Japan), Norishige Fukushima (Graduate School of Engineering, Nagoya Institute of Technology, Japan), Tomohiro Yendo, Mehrdad Panahpour Tehrani (Graduate School of Engineering, Nagoya University, Japan), Toshiaki Fujii (Graduate School of Engineering, Tokyo Institute of Technology, Japan), Masayuki Tanimoto (Graduate School of Engineering, Nagoya University, Japan)
Pagepp. 330 - 333
KeywordFTV, Free Viewpoint Image Generation, Image Based Rendering, Realtime Processing, Parallel Processing
AbstractIn this paper, we propose a parallel processing method to generate free viewpoint image in realtime. It is impossible to arrange the cameras in a high density realistically though it is necessary to capture images of the scene from innumerable cameras to express the free viewpoint image. Therefore, it is necessary to interpolate the image of arbitrary viewpoint from limited captured images. However, this process has the relation of the trade-off between the image quality and the computing time. In proposed method, it aimed to generate the high-quality free viewpoint image in realtime by applying the parallel processing to time-consuming interpolation part.

P3-3
TitleInfluence of Wavelet-Based Depth Coding in Multiview Video Systems
AuthorIsmael Daribo, Hideo Saito (Keio University, Japan)
Pagepp. 334 - 337
Keywordwavelet, depth, coding, 3dtv, mvv
AbstractMultiview video representation based on depth data, such as multiview video-plus-depth (MVD), is emerging 3D video communication services raising in the meantime the problem of coding and transmitting depth video in addition to classical texture video. Depth video is considered as a key side information in novel view synthesis within multiview video systems, such as three-dimensional television (3DTV) or free viewpoint television (FTV), wherein the influence of depth compression on the novel synthesized view is still a contentious issue. In this paper, we propose to discuss and investigate the impact of wavelet-based compression of the depth video on the quality of the view synthesis. Experimental results show that significant gains can be obtained by improving depth edge preservation through shorter wavelet-based filtering on depth edges.

P3-4
TitleAn Epipolar Resticted Inter-Mode Selection for Stereoscopic Video Encoding
AuthorGuolei Yang (Peking University, China), Luhong Liang (Institute of Computing Technology, Chinese Academy of Sciences, China), Wen Gao (Peking University, China)
Pagepp. 338 - 341
KeywordStereoscopic video, Epipolar restriction, Inter-mode selection
AbstractIn this paper, we propose a fast inter-prediction mode selection algorithm for stereoscopic video encoding. Different from methods using disparity estimation, candidate modes are generated by sliding a window along the macro-block line restricted by the epipolar. Then the motion information is utilized to rectify the candidate modes. A selection failure handling algorithm is also proposed to preserve coding quality. The proposed algorithm is evaluated using independent H.264/AVC encoders for left and right views and can be extended to MVC.

P3-5
TitleTemporal Consistency Enhancement on Depth Sequences
AuthorDeliang Fu, Yin Zhao, Lu Yu (Zhejiang University, China)
Pagepp. 342 - 345
Keyword3D video, temporal consistency, temporal depth filtering, view synthesis
AbstractCurrently, depth sequences generated by automatic depth estimation suffer from the temporal inconsistency problem. Estimated depth values of some objects vary in adjacent frames, whereas the objects actually remain on the same depth planes. These temporal depth errors significantly impair the visual quality of the synthesized virtual view as well as the coding efficiency of the depth sequences. Since depth sequences correspond to texture sequences, some erroneous temporal depth variations can be detected by analyzing temporal variations of the texture sequences. Utilizing this property, we propose a novel solution to enhance the temporal consistency of depth sequences by applying adaptive temporal filtering on them. Experiments demonstrate that the proposed depth filtering algorithm can effectively suppress transient depth errors and generate more stable depth sequences, resulting in notable temporal quality improvement of the synthesized views and higher coding efficiency on the depth sequences.

P3-6
TitleReal-Time Free Viewpoint Television for Embedded Systems
AuthorDavide Aliprandi, Emiliano Piccinelli (STMicroelectronics, Italy)
Pagepp. 346 - 349
KeywordViewpoint, Depth, 3DTV
AbstractIn this paper we describe an image-based rendering pipeline for interactive real-time Free Viewpoint Television (FTV) on embedded systems. Description of the processing steps and optimizations implemented targeting the hardware acceleration of a commercial programmable Graphics Processing Unit (GPU) is given. As a result, real-time view synthesis at 70 fps in XGA resolution has bee achieved. Restrictions and modifications introduced to support the application on OpenGL ES 2.0 based GPUs for embedded systems have also been discussed.

P3-7
TitlePower-Aware Complexity-Scalable Multiview Video Coding for Mobile Devices
AuthorMuhammad Shafique (Karlsruhe Institute of Technology, Germany), Bruno Zatt, Sergio Bampi (Federal University of Rio Grande do Sul, Brazil), Jörg Henkel (Karlsruhe Institute of Technology, Germany)
Pagepp. 350 - 353
KeywordMVC, Mobile Devices, Complexity reduction, Power-Aware, Adaptivity
AbstractWe propose a novel power-aware scheme for complexity-scalable multiview video coding on mobile devices. Our scheme exploits the asymmetric view quality which is based on the binocular suppression theory. Our scheme employs different quality-complexity classes (QCCs) and adapts at run time depending upon the current battery state. It thereby enables a run-time tradeoff between complexity and video quality. The experimental results show that our scheme is superior to state-of-the-art and it provides an up to 87% complexity reduction while keeping the PSNR close to the exhaustive mode decision. We have demonstrated the power-aware adaptivity between different QCCs using a laptop with battery charging and discharging scenarios.

P3-8
Title3D Pose Estimation in High Dimensional Search Spaces with Local Memorization
AuthorWeilan Luo, Toshihiko Yamasaki, Kiyoharu Aizawa (The University of Tokyo, Japan)
Pagepp. 354 - 357
Keywordtracking, annealing, twist, particle filter
AbstractIn this paper, a stochastic approach for extracting the articulated 3D human postures by synchronized multiple cameras is presented in the high-dimensional configuration spaces. Annealed Particle Filtering (APF) seeks for the globally optimal solution of the likelihood. We improve and extend the APF with local memorization to estimate the suited kinematic postures for a volume sequence directly instead of projecting a rough simplified body model to 2D images. Our method guides the particles to the global optimization on the basis of local constraints. A segmentation algorithm is performed on the volumetric models and the process is repeated. We assign the articulated models 42 degrees of freedom. The matching error is about 6% on average while tracking the posture between two neighboring frames.

P3-9
TitleFree-Viewpoint Image Generation Using Different Focal Length Camera Array
AuthorKengo Ando (Graduate School of Engineering, Nagoya University, Japan), Norishige Fukushima (Graduate School of Engineering, Nagoya Institute of Technology, Japan), Tomohiro Yendo, Mehrdad Panahpour Tehrani (Graduate School of Engineering, Nagoya University, Japan), Toshiaki Fujii (Graduate School of Engineering, Tokyo Institute of Technology, Japan), Masayuki Tanimoto (Graduate School of Engineering, Nagoya University, Japan)
Pagepp. 358 - 361
KeywordFree-viewpoint image generation, Image Based Rendering
AbstractThe availability of multi-view images including Free-Viewpoint TV. Virtual viewpoint images are synthesized by Image-Based Rendering. In this paper, we introduce a depth estimation method for forward virtual viewpoints and view generation method using a zoom camera in our camera setup to improve virtual viewpoints’ image quality. Simulation results confirm reduced error during depth estimation using our proposed method in comparison with conventional stereo matching scheme. We have demonstrated the improvement in image resolution of virtually moved forward camera.


[Beyond H.264/MPEG-4 AVC and related topics]

P3-10
TitleDecoder-Side Hierarchical Motion Estimation for Dense Vector Fields
AuthorSven Klomp, Marco Munderloh, Jörn Ostermann (Leibniz Universität Hannover, Germany)
Pagepp. 362 - 365
Keywordvideo coding, motion compensation, dense vector field, block matching
AbstractRecent research revealed that the data rate can be reduced by performing an additional motion estimation at the decoder. This paper addresses an improved hierarchical motion estimation algorithm to be used in a decoder-side motion estimation system. A special motion vector latching is used to be more robust for very small block sizes and to better adapt to object borders. A dense motion vector field is estimated which reduces the rate by 6.9% in average compared to H.264/AVC.

P3-11
TitleEdge-Based Adaptive Directional Intra Prediction
AuthorFeng Zou, Oscar C. Au, Wen Yang, Chao Pang, Jingjing Dai, Xing Wen (The Hong Kong University of Science and Technology, Hong Kong), Yu Liu (Hong Kong Applied Science and Technology Research Institute, Hong Kong)
Pagepp. 366 - 369
KeywordH.264/AVC, intra prediction, edge
AbstractH.264/AVC employs intra prediction to reduce spatial redundancy between neighboring blocks. Different directional prediction modes are used to cater diversified video content. Although it achieves quite high coding efficiency, it is desirable to analyze its drawbacks in the existing video coding standard, since it allows us to design better ones. Basically, even after intra prediction, the residue still contains a lot of edge or texture information. Unfortunately, these high frequency components consume a large quantity of bits and the distortion is usually quite high. Based on this drawback, an Edge-based Adaptive Directional Intra Prediction is proposed (EADIP) to reduce the residue energy especially for the edge region. In particular, we establish an edge model in EADIP, which is quite flexible for natural images. Within the model, the edge splits the macroblock into two regions, each being predicted separately. In implementation, we consider the current trend of mode selection and complexity issues. A mode extension is made on INTRA 16x16 in H.264/AVC. Experimental results show that the proposed algorithm outperforms H.264/AVC. And the proposed mode is more likely to be chosen in low bitrate situations.

P3-12
TitleAn Improved Low Delay Inter Frame Coding Using Template Matching Averaging
AuthorYoshinori Suzuki, Choong Seng Boon (NTT DOCOMO, INC., Japan)
Pagepp. 370 - 373
KeywordVideo coding, Prediction methods, Motion compensation
AbstractThis paper presents an efficient inter prediction method for video coding. The method applies the idea of template matching averaging to the conventional motion compensated prediction. While one of the candidate is specified by a motion vector, the remaining candidates are obtained by a template matching without using explicit motion vector. The averaging of multiple predictors reduces coding noise residing in each of the predictors. Simulation results show that the proposed scheme improves coding efficiency up to 4.5%.

P3-13
TitleGenerating Subject Oriented Codec by Evolutionary Approach
AuthorMasaaki Matsumura, Seishi Takamura, Hirohisa Jozawa (NTT Cyber Space Laboratories, NTT Corporation, Japan)
Pagepp. 374 - 377
KeywordEvolutive image coding, coding tools combination, subject oriented codec, lossless image coding
AbstractIn this paper, we propose an automatic optimization method for deriving the combination that suits for categorized pictures. We prepare some categorised pictures, and optimize the combination for each category. In the case of optimization for lossless image coding, our method achieves a bit-rate reduction of over 2.8% (maximum) compared to the combination that offers the best bit-rate averagely prepared beforehand.

P3-14
TitleImproved Context Modeling for Coding Quantized Transform Coefficients in Video Compression
AuthorTung Nguyen, Heiko Schwarz, Heiner Kirchhoffer, Detlev Marpe (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany), Thomas Wiegand (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute/Technical University of Berlin, Germany)
Pagepp. 378 - 381
Keywordcontext modeling, transform coding
AbstractRecent investigations have shown that the support of ex-tended block sizes for motion-compensated prediction and transform coding can significantly increase the coding effi-ciency for high-resolution video relative to H.264/AVC. In this paper, we present a new context-modeling scheme for the coding of transform coefficient levels that is particularly suitable for transform blocks greater than 8x8. While the basic concept for transform coefficient coding is similar to CABAC, the probability model selection has been optimized for larger block transforms. The proposed context modeling is compared to a straightforward extension of the CABAC context modeling; both schemes have been implemented in a hybrid video codec design that supports block sizes of up to 128x128 samples. In our simulations, we obtained overall bit rate reductions of up to 4%, with an average of 1.7% with the proposed context modeling scheme.


[Image/video coding and related topics]

P3-15
TitleBitwise Prediction Error Correction for Distributed Video Coding
AuthorAxel Becker-Lakus, Ka-Ming Leung, Zhonghua Ma (Canon Information Systems Research Australia (CiSRA), Australia)
Pagepp. 382 - 385
KeywordDistributed Video Coding, Wyner-Ziv Coding, Side Information Generation
AbstractSide information plays a key role in the performance of a Distributed Video Coding (DVC) system. However, the generation of side information often relies on complex motion estimation/interpolation operation. The correlation between the source data and the side information, sometimes referred as virtual channel, is also very difficult to model accurately. In this paper, we propose a bitwise prediction error correction method to improve the quality of the side information during Wyner-Ziv decoding. Whenever a bit error is detected in a bit plane, the less significant bits of the corresponding pixel are adjusted to match the initial prediction. The proposed method has been evaluated using a pixel-domain DVC system and delivers a better coding performance with improved decoding quality and reduced bitrate.

P3-16
TitleImproved Texture Compression for S3TC
AuthorYifei Jiang, Dandan Huan (Institute of Computing Technology, Chinese Academy of Sciences, China)
Pagepp. 386 - 389
Keywordcomputer graphics, texture compression, clustering algorithms
AbstractTexture compression is a specialized form of still image compression employed in computer graphics systems to reduce memory bandwidth consumption. Modern texture compression schemes cannot generate satisfactory qualities for both alpha channel and color channel of texture images. We propose a novel texture compression scheme, named ImTC, based on the insight into the essential difference between transparency and color. ImTC defines new data formats and compresses the two channels flexibly. While keeping the same compression ratio as the de facto standard texture compression scheme, ImTC improves compression qualities of both channels. The average PSNR score of alpha channel is improved by about 0.2 dB, and that of color channel can be increased by 6.50 dB over a set of test images, which makes ImTC a better substitute for the standard scheme.

P3-17
TitleCompression of Pre-Computed Per-Pixel Texture Features Using MDS
AuthorWai-Man Pang (Spatial Media Group, Computer Arts Lab., University of Aizu, Japan), Hon-Cheng Wong (Faculty of Information Technology, Macau University of Science and Technology, Macau)
Pagepp. 390 - 393
KeywordCompressed texture features, Gabor wavelet transform, Multidimensional scaling, Compression
AbstractThere are many successful experiences on employing texture analysis to improve the accuracy and robustness on image segmentation. Usually, a per-pixel based texture analysis is required, this involves intensive computation especially for large images. While, precomputation and storing of the texture features involves large file space which is not cost effective. To adopt to this novel needs, we propose in this paper the use of multidimensional scaling (MDS) technqiue to reduce the size of per-pixel texture features of an image, while preserving the textural discrminiability for segmentation. As per-pixel texture features will create very large dissimilarity matrix, and make the solving of MDS intractable. A sampling-based MDS is therefore introduced to tackle the problem with a divide-and-conquer approach. A compression ratio of 1:24 can be achieved with an average error lower than 7%. Preliminary experiments on segmentation using the compressed data show satisfactory results as good as using the uncompressed features. We foresee that such a method will enable texture features to be stored and transferred more effectively on low processing power devices or embedded system like mobile phones.

P3-18
TitleTemporal Signal Energy Correction and Low-Complexity Encoder Feedback for Lossy Scalable Video Coding
AuthorMarijn J.H. Loomans, Cornelis J. Koeleman (VDG Security BV, Netherlands), Peter H.N. de With (Eindhoven University of Technology, Netherlands)
Pagepp. 394 - 397
KeywordScalable Video Coding, Wavelets, Embedded systems
AbstractWe address two problems found in embedded Scalable Video Codec implementations: the temporal signal energy distribution and frame-to-frame quality fluctuations. To solve these problems, we move the temporal energy correction to the leaves of the temporal tree, and feed back the decoded first frame of the GOP into the temporal coding chain. The first modification saves on required memory size, bandwidth and computations, while reducing floating/fixed-point conversion errors and the second is achieved without entropy decoding and an unmodified decoder.

P3-19
TitleImproving Colorization-Based Coding by Using Local Correlation between Luminance and Chrominance in Texture Component
AuthorYoshitaka Inoue, Takamichi Miyata, Yoshinori Sakai (Tokyo Institute of Technology, Japan)
Pagepp. 398 - 401
Keywordimage coding, colorization, total variation, correlation between luminance and chrominance
AbstractRecently, a novel approach to color image compression based on colorization has been presented. Although the conventional method of colorization-based coding outperforms JPEG in terms of subjective quality, the decoded chrominance components lose the local oscillation that the original images had. We focus on the local correlation that exists between luminance and chrominance in separated texture components, and we present a new colorization-based coding method. Experimental results showed that our coding method can improve the coding efficiency.

P3-20
TitleVideo Encoding with the Original Picture as the Reference Picture
AuthorTaiga Muromoto, Naoya Sagara, Kenji Sugiyama (Seikei University, Japan)
Pagepp. 402 - 405
KeywordInter-picture prediction, Reference picture, Quantization error, Group of picture
AbstractInter-picture prediction uses a local decoded picture for the reference, in order to avoid a mismatch between encoding and decoding. However, this scheme does not necessarily result in optimal coding efficiency since it requires encoding the processing altogether. Therefore, we study the use of the original picture as the reference. In this case, although the mismatch causes degradation of the picture quality, the bit amount is reduced. Therefore, we propose an adaptive method based on rate distortion optimization. The original picture is used only in the macroblock, if it is lower cost than the local decoded picture is used. Experimental results show a 0.1 to 1.0 dB gain in PSNR in each sequence.

P3-21
TitleA New Hybrid Parallel Intra Coding Method Based on Interpolative Prediction
AuthorCui Wang (Tokyo Institute of Technology, Japan), Akira Kubota (Chuo University, Japan), Yoshinori Hatori (Tokyo Institute of Technology, Japan)
Pagepp. 406 - 409
Keywordhybrid parallel coding, interpolative prediction, new shape of block
AbstractThe hybrid coding method to combine the predictive coding with the orthogonal transformation and the quantization is mainly used recently. This paper proposes a new hybrid parallel intra coding method based on interpolative prediction which uses correlations between neighboring pixels. For high performance and parallel, the optimal quantizing scheme and the new shape of the block are used. The result of experiments shows that the proposed technique achieves 1~4 dB improvement in Luminance PSNR, especially for image with more details.

P3-22
TitleRBF-Based VBR Controller for Real-Time H.264/SVC Video Coding
AuthorSergio Sanz-Rodríguez, Fernando Díaz-de-María (Carlos III University of Madrid, Spain)
Pagepp. 410 - 413
KeywordRate Control, Variable Bit Rate, Scalable Video Coding, H.264/SVC, streaming
AbstractIn this paper we propose a novel VBR controller for real-time H.264/SVC video coding. Since consecutive pictures within the same scene often exhibit similar degrees of complexity, the proposed VBR controller allows for just an incremental variation of QP with respect to that of the previous picture, so preventing unnecessary QP fluctuations. For this purpose, a RBF network has been carefully designed to estimate the QP increment at each dependency (spatial or CGS) layer. A mobile live streaming application scenario was simulated to assess the performance of the proposed VBR controller, which was compared to a recently proposed CBR controller for H.264/SVC. The experimental results show a remarkably consistent quality, notably outperforming the reference CBR controller.

P3-23
TitleScalable Video Compression Framework with Adaptive Multiresolution Directional Filter Bank Design
AuthorLingchen Zhu, Hongkai Xiong (Shanghai Jiao Tong University, China)
Pagepp. 414 - 417
KeywordScalable video coding, directional filter banks, multiscale geometric, sparse coding
AbstractRegarding orientation resolution as an isolated variable from scale, this paper proposes a dual (scale and orientation) multiresolution transform into scalable video coding (SVC) framework. By projecting 2D signals (textures and edges) onto nonuniformly divided orientation subspaces, the dual multiresolution SVC (DMSVC) can capture 2-D curve smoothness with less coefficients, and provide more flexible spatial decomposition structures than traditional wavelet-based SVC (WSVC). In the spatial decomposition module of DMSVC, the nonuniform directional distribution along scale of each frame is detected by phase congruency in the overcomplete wavelet domain. The corresponding orientational multi-resolution is achieved by nonuniform directional filter banks (NUDFB) which is fulfilled via a non-symmetric binary tree (NSBT) structured frequency division. The wavelet basis function in each scale is converted to an adaptive set of nonuniform directional basis by employing nonuniform directional filter banks. Experimental results validate a superior coding performance and visual quality over WSVC especially on those sequences full of directional edges and textures.

P3-24
TitleA Four-Description MDC for High Loss-Rate Channels
AuthorMeilin Yang, Mary Comer, Edward J. Delp (School of Electrical and Computer Engineering, Purdue University, U.S.A.)
Pagepp. 418 - 421
KeywordMDC: four-description MDC, high packet loss rate, Gilbert model
AbstractOne of the most difficult problems in video transmission is communication over error-prone channels, especially when retransmission is unacceptable. To address this problem, Multiple Description Coding (MDC) has been proposed as an effective solution due to its robust error resilience. Considering applications in scalable, multicast and P2P environments, it is advantageous to use more than two descriptions (which is designated multi-description MDC in this paper). In this paper, we present a new four-description MDC for high loss-rate channel using a hybrid structure of temporal and spatial correlations. A Gilbert model is used as the channel model for burst packet loss simulation. Experimental results demonstrate the efficacy of the proposed method.

P3-25
TitleBi-Directional Optical Flow for Improving Motion Compensation
AuthorAlexander Alshin, Elena Alshina, Tammy Lee (Samsung Electronics Co., Ltd., Republic of Korea)
Pagepp. 422 - 425
Keywordbi-directional prediction, optical flow
AbstractNew method improving B-slice prediction is proposed. By combining the optical flow concept and high accuracy gradients evaluation we construct the algorithm which allows pixel-wise refinement of motion. This approach does not require any signaling for decoder. According to tests with WQVGA sequences bit-saving of 2%-6% can be achieved using this tool.


[Image/video processing and related topics]

P3-26
TitleTwo-Dimensional Chebyshev Polynomials for Image Fusion
AuthorZaid Omar, Nikolaos Mitianoudis, Tania Stathaki (Imperial College London, U.K.)
Pagepp. 426 - 429
KeywordImage and data fusion, Chebyshev polynomials, orthogonal moments
AbstractThis report documents in detail the research carried out by the author throughout his first year. The paper presents a novel method for fusing images in a domain concerning multiple sensors and modalities. Using Chebyshev polynomials as basis functions, the image is decomposed to perform fusion at feature level. Results show favourable performance compared to previous efforts on image fusion, namely ICA and DT-CWT, in noise affected images. The work presented here aims at providing a novel framework for future studies in image analysis and may introduce innovations in the fields of surveillance, medical imaging and remote sensing.

P3-27
TitleImage Denoising with Hard Color-Shrinkage and Grouplet Transform
AuthorTakahiro Saito, Ken-ichi Ishikawa, Yasutaka Ueda, Takashi Komatsu (Kanagawa University, Japan)
Pagepp. 430 - 433
KeywordColor-image processing, denoising, wavelet transform, grouplet transform, shrinkage
AbstractTo remove signal-dependent noise of a digital color camera, we propose a denoising method with our hard color-shrinkage in the tight-frame grouplet transform domain. The classic hard-shrinkage works well for monochrome-image denoising. To utilize inter-channel color cross-correlations, a noisy image undergoes the color transformation from the RGB to the luminance-and-chrominance color space, and the luminance and the chrominance components are separately denoised; but this approach cannot cope with actual signal-dependent noise. To utilize the noise’s signal-dependencies, we construct the hard color-shrinkage where the inter-channel color cross-correlations are directly utilized in the RGB color space. The hard color-shrinkage alleviates denoising artifacts, and improves picture quality of denoised images.

P3-28
TitleImproved FMO Based H.264 Frame Layer Rate Control for Low Bit Rate Video Transmission
AuthorRhandley Domingo Cajote (University of the Philippines, Diliman, Philippines), Supavadee Aramvith (Chulalongkorn University, Thailand)
Pagepp. 434 - 437
KeywordFMO, Rate Control, H.264/AVC, video coding
AbstractThe use of Flexible Macroblock Ordering (FMO) in H.264/AVC as an error-resilient tool incurs extra overhead bits that reduces coding efficiency at low bit rate. To improve coding efficiency, we present an improved frame-layer H.264/AVC rate control that takes into consideration the effects of using FMO for video transmission. In this paper, we propose a new header bits model, an enhanced frame complexity measure and a quantization parameter (QP) adjustment scheme. Simulation results show that the proposed method performed better than the existing frame layer rate control with FMO enabled using different number of slice groups.

P3-29
TitleImprovement of Spatial Resolution by Integration of High-Speed Sub-Frame Images
AuthorDaisuke Kashiwagura, Kanae Matsuzaki, Takayuki Hamamoto (Tokyo University of Science, Japan)
Pagepp. 438 - 441
Keywordsuper resolution, sub-frame image, high frame rate, motion estimation
AbstractThe super resolution technique based on the integration of successive frames depends on the accuracy of the motion estimation. However it requires much amount of calculation and tends to make estimation errors for some images. In this paper, we propose a super resolution method by the motion estimation based on the block matching using high-speed sub-frame images.

P3-30
TitleImproved Autoregressive Image Model Estimation for Directional Image Interpolation
AuthorRuiqin Xiong (Peking University, China), Wenpeng Ding (Beijing University of Technology, China), Siwei Ma, Wen Gao (Peking University, China)
Pagepp. 442 - 445
Keywordimage interpolation, model estimation, autoregressive model, regularization
AbstractFor image interpolation algorithms employing autoregressive models, a mechanism is required to estimate the model parameters piecewisely and accurately so that local structures of image can be exploited efficiently. This paper proposes a new strategy for better estimating the model. Different from conventional schemes which build the model solely upon the covariance matrix of low-resolution image, the proposed strategy utilizes the covariance matrix of high-resolution image itself, with missing pixels properly initialized. To make the estimation robust, we adopt a general solution which exploits the covariance matrices of both scales. Experimental results demonstrate that the proposed strategy improves model estimation and the interpolation performance remarkably.


[Quality, system, applications, and other topics]

P3-31
TitleSubjective Evaluation of Hierarchical B-Frames Using Video-MUSHRA
AuthorHussain Mohammed, Nikolaus Färber, Jens Garbas (Fraunhofer IIS, Germany)
Pagepp. 446 - 449
KeywordH.264/AVC, Hierarchical B-Frames, Subjective Quality, MUSHRA
AbstractHierarchical B-Frames (HBF) has emerged as an efficient video coding tool in recent years. As shown in the literature, this approach results in excellent PSNR gains of >1 dB. However these PSNR gains are not sufficiently assessed in a scientific manner by subjective tests. Hence in this paper, we evaluate HBF coding pattern subjectively by using the MUSHRA test methodology. While MUSHRA is well established in audio coding research, its application to video is a novelty of this paper. We compare HBF with simple IPP coding pattern at either same PSNR or same bit rate. Our results indicate that, HBF gains are clearly subjectively perceptible. Hence, it can be shown that PSNR gains also correlate with a subjective gain. Interestingly, even at same PSNR, HBF is found to be subjectively superior to simple IPP coding.

P3-32
TitleIntra Prediction Architecture for H.264/AVC QFHD Encoder
AuthorGang He, Dajiang Zhou, Jinjia Zhou, Satoshi Goto (Waseda University, Japan)
Pagepp. 450 - 453
KeywordH.264, intra prediction, hardware architecture
AbstractThis paper proposes a high-performance intra prediction architecture that can support H.264/AVC high profile. The proposed MB/block co-reordering can avoid data dependency and improve pipeline utilization. Therefore, the timing constraint of real-time 4kx2k encoding can be achieved with negligible quality loss. 16x16 prediction engine and 8x8 prediction engine work parallel for prediction and coefficients generating. A reordering interlaced reconstruction is also designed for fully pipelined architecture. It takes only 160 cycles to process one macroblock (MB). Hardware utilization of prediction and reconstruction modules is almost 100%. Furthermore, PE-reusable 8x8 intra predictor and hybrid SAD & SATD mode decision are proposed to save hardware cost. The design is implemented by 90nm CMOS technology with 113.2k gates and can encode 4kx2k video sequences at 60 fps with operation frequency of 310MHz.

P3-33
TitleCompressed Signature for Video Identification
AuthorNikola Sprljan, Paul Brasnett, Stavros Paschalakis (Mitsubishi Electric R&D Centre Europe, U.K.)
Pagepp. 454 - 457
Keywordvideo descriptor, lossless compression
AbstractThis paper presents a new application-specific lossless compression scheme developed for video identification descriptors, also known as video fingerprints or signatures. In designing such a descriptor, one usually has to balance the descriptor size against discriminating power and temporal localisation performance. The proposed compression scheme alleviates this problem by efficiently exploiting the temporal redundancies present in the video fingerprint, allowing highly accurate fingerprints which also entail low transmission and storage costs. In this paper we provide a detailed description of our compression scheme and a comparative evaluation against well known state-of-the-art generic compression tools.

P3-34
TitleA Subjective Image Quality Metric for Bit-Inversion-Based Watermarking
AuthorTadahiko Kimoto, Fumihiko Kosaka (Toyo University, Japan)
Pagepp. 458 - 461
Keywordimage watermark, subjective quality, perceptually adaptive system
AbstractAn image watermarking scheme using the previously proposed bit embedding method is developed. Based on the properties of the bit embedding method, the perceptual model of two kinds of objective quality measures is assumed. Then, the measurements of human subjective image quality are analyzed from the viewpoint of the correlation with these two measures. Thereby, the estimating function that can yield an estimate of the subjective quality from two objective measurements is determined. By using the estimating function, the perceptually adaptive watermarking can be achieved.


Session T2  Tutorial Session 2
Time: 11:15 - 12:00 Friday, December 10, 2010
Chair: Takayuki Hamamoto (Tokyo Science University, Japan)

T2-1 (Time: 11:15 - 12:00)
Title(Tutorial) Quality Assessment for Image Compression Purpose
AuthorChaker Larabi (University of Poitiers, France)
AbstractIn the last years, image quality assessment became a very hot research topic especially for compression of image and video, first because of the large availability of multimedia applications and contents, and then because many scientists/engineers need to make a selection of algorithms and tools. This tutorial is designed to cover several aspects of the field of image quality assessment. After a brief introduction about the needs of quality assessment for multimedia applications, a review of the main approaches will be made by giving a description of the metrics categories and the subjective paradigms. At this point, it is important to distinguish between Full reference, reduced reference and no reference metrics but also understand the difference between image and video quality assessment. The focus of this course will be put on how to answer the question: which quality procedure for which application and which content? Another important topic for this course is on how to measure the performance of a given metric for a given application. Several practical examples will allow to better handle the quality assessment problem.


Session S2  Special Session 2: Beyond H.264/MPEG-4 AVC
Time: 13:15 - 15:15 Friday, December 10, 2010
Chair: Seishi Takamura (NTT Corporation, Japan)

S2-1 (Time: 13:15 - 13:45)
TitleRecent Advances in Video Coding Using Static Background Models
AuthorAndreas Krutz, Alexander Glantz, Thomas Sikora (Technische Universität Berlin, Germany)
Pagepp. 462 - 465
KeywordVideo Coding, Model-based Video Coding, H.264/AVC
AbstractSprite coding, as standardized in MPEG-4 Visual, can result in superior performance compared to common hybrid video codecs both objectively and subjectively. However, state-ofthe- art video coding standard H.264/AVC clearly outperforms MPEG-4 Visual sprite coding in broad bit rate ranges. Based on the sprite coding idea, this paper proposes a video coding technique that merges the advantages of H.264/AVC and sprite coding. For that, sophisticated algorithms for global motion estimation, sprite generation and object segmentation – all needed for thorough sprite coding – are incorporated into an H.264/AVC coding environment. The proposed approach outperforms H.264/AVC especially in lower bit rate ranges. Savings up to 21% can be achieved.

S2-2 (Time: 13:45 - 14:15)
TitleNovel Video Coding Paradigm with Reduction/Restoration Processes
AuthorToshie Misu, Yasutaka Matsuo, Shinichi Sakaida, Yoshiaki Shishikui, Eisuke Nakasu (Science & Technology Research Laboratories, NHK, Japan)
Pagepp. 466 - 469
Keywordvideo coding, super-resolution, image reduction, image restoration, nonuniform sampling
AbstractTo optimally design distortions in lossy video coding, we propose the use of a novel coding paradigm with adaptive nonlinear transforms as pre/post-processors of a conventional video codec. The preprocessor decimates less important pixels based on an image analysis. A conventional video encoder such as MPEG-4 AVC/H.264 further eliminates the redundancy of the decimated images. On the decoder side, the postprocessor restores small decoded images of the conventional decoder to the original resolution using an inverse mapping including a super-resolution technique that uses a priori knowledge on the decimation in the preprocessing. Results of the experiments showed the proposed coding scheme poses distortion that has a more straightforward appearance than that of a directly encoded/decoded image by a sole conventional H.264 codec.

S2-3 (Time: 14:15 - 14:45)
TitleTowards Efficient Intra Prediction Based on Image Inpainting Methods
AuthorDimitar Doshkov, Patrick Ndjiki-Nya, Haricharan Lakshman, Martin Koeppel, Thomas Wiegand (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany)
Pagepp. 470 - 473
KeywordIntra prediction, Texture synthesis, PDEs, Template matching, Inpainting
AbstractIn this paper, novel intra prediction methods based on image inpainting approaches are proposed. The H.264/AVC intra prediction modes are not well suited for processing complex textures at low bit rates. Our algorithm utilizes an efficient combination of partial differential equations (PDEs) and patch based texture synthesis in addition to the standard directional predictors. Bit rate savings up to 3.5% compared to that of the H.264/AVC standard are shown.

S2-4 (Time: 14:45 - 15:15)
TitleLow Complexity Video Coding and the Emerging HEVC Standard
AuthorKemal Ugur (Nokia Research Center, Finland), Kenneth Andersson (Ericsson Research, Sweden), Arild Fuldseth, Gisle Bjontegaard, Lars Peter Enderssen (Tandberg Telecom (Cisco), Norway), Jani Lainema, Antti Hallapuro, Dmytro Rusanovskyy, Cixun Zhang (Nokia Research Center, Finland), Andrey Norkin, Clinton Priddle, Thomas Rusert, Jonatan Samuelsson, Rickard Sjoberg, Zhuangfei Wu (Ericsson Research, Sweden), Justin Ridge (Nokia, U.S.A.)
Pagepp. 474 - 477
KeywordHEVC, standardization, video coding, H.264/AVC
AbstractThis paper describes a low complexity video codec with high coding efficiency. It was proposed to the High Efficiency Video Coding (HEVC) standardization effort of MPEG and VCEG, and has been partially adopted into the initial HEVC Test Model under Consideration design. The proposal utilizes a quad-tree structure with a support of large macroblocks of size 64x64 and 32x32, in addition to macroblocks of size 16x16. The entropy coding is done using a low complexity variable length coding based scheme with improved context adaptation over the H.264/AVC design. In addition, the proposal includes improved interpolation and deblocking filters, giving better coding efficiency while having low complexity. Finally, an improved intra coding method is presented. The subjective quality of the proposal is evaluated extensively and the results show that the proposed method achieves similar visual quality as H.264/AVC High Profile anchors with around 50% and 35% bit rate reduction for low delay and random-access experiments respectively at high definition sequences. This is achieved with less complexity than H.264/AVC Baseline Profile, making the proposal especially suitable for resource constrained environments.


Session D2  Panel Discussion 2: Beyond H.264/MPEG-4 AVC
Time: 15:15 - 16:15 Friday, December 10, 2010
Chair: Kohtaro Asai (Mitsubishi Electric Corporation, Japan)

D2-1 (Time: 15:15 - 16:15)
Title(Panel Discussion) Beyond H.264/MPEG-4 AVC
AuthorChair: Kohtaro Asai (Mitsubishi Electric Corporation, Japan), Panelists: Thomas Wiegand (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany), Kemal Ugur (Nokia Research Center, Finland), Woo-Jin Han (Samsung Electronics, Republic of Korea), Andrew Segall (Sharp Labs of America, U.S.A.), Teruhiko Suzuki (Sony, Japan)


Session P4  Poster Session 4
Time: 16:30 - 18:00 Friday, December 10, 2010
Chair: Shinichi Sakaida (Science & Technology Research Laboratories, NHK, Japan)


[3DTV/FTV/multi-view-related topics]

P4-1
TitleOn-Line Statistical Analysis Based Fast Mode Decision for Multi-View Video Coding
AuthorChia-Chi Chan (Dept. Communication Engineering, National Central University, Taiwan), Jheng-Ping Lin (ZyXEL Corp., Taiwan), Chih-Wei Tang (Dept. Communication Engineering, National Central University, Taiwan)
Pagepp. 478 - 481
KeywordMulti-view video coding, fast mode decision, statistical analysis, RD cost, motion and disparity estimation
AbstractThe high computational complexity of multi-view video codecs makes it necessary to speed up for their realization in consumer electronics. Since fast encoding algorithms are expected to adapt to different video sequences, this paper proposes a fast algorithm that consists of fast mode decision and fast disparity estimation for multi-view video coding. The fast mode decision algorithm applies to both temporal and inter-view predictions. The candidates for mode decision are reduced based on a set of thresholds. Differ from the previous fast mode decision algorithms for MVC, this scheme determines the thresholds according to the on-line statistical analysis of motion and disparity costs of the first GOP in each view. Since the inter-view prediction is time consuming, we propose a fast disparity estimation algorithm to save encoding time. Experimental results show that our proposed scheme reduces the computational complexity significantly with negligible degradation of coding efficiency.

P4-2
TitleOptimal Rate Allocation for View Synthesis along a Continuous Viewpoint Location in Multiview Imaging
AuthorVladan Velisavljevic (Deutsche Telekom Laboratories, Germany), Gene Cheung (National Institute of Informatics, Japan), Jacob Chakareski (Ecole Polytechnique Federale de Lausanne, Switzerland)
Pagepp. 482 - 485
KeywordMulti-view imaging, Rate allocation
AbstractWe consider the scenario of view synthesis via depth-image based rendering in multi-view imaging. We formulate a resource allocation problem of jointly assigning an optimal number of bits to compressed texture and depth images such that the maximum distortion of a synthesized view over a continuum of viewpoints between two encoded reference views is minimized, for a given bit budget. We construct simple yet accurate image models that characterize the pixel values at similar depths as first-order Gaussian auto-regressive processes. Based on our models, we derive an optimization procedure that numerically solves the formulated min-max problem using Lagrange relaxation. Through simulations we show that, for two captured views scenario, our optimization provides a significant gain (up to 2dB) in quality of the synthesized views for the same overall bit rate over a heuristic quantization that selects only two quantizers - one for the encoded texture images and the other for the depth images.

P4-3
TitlePanoramic Scene Generation from Multi-view Images with Close Foreground Objects
AuthorSoon-Young Lee (Seoul National University, Republic of Korea), Jae-Young Sim (Ulsan National Institute of Science and Technology, Republic of Korea), Chang-Su Kim (Korea University, Republic of Korea), Sang-Uk Lee (Seoul National University, Republic of Korea)
Pagepp. 486 - 489
Keyword3D scene representation, Image panorama, view expansion, multi-view image processing
AbstractAn algorithm to generate a panorama from multi-view images, which contain foreground objects with varying depths, is proposed in this work. The proposed algorithm constructs a foreground panorama and a background panorama separately, and then merges them into a complete panorama. First, the foreground panorama is obtained by finding the translational displacements of objects between source images. Second, the background panorama is initialized using warped source images and then optimized to preserve spatial consistency and satisfy visual constraints. Then, the background panorama is extended by inserting seams and merged with the foreground panorama. Experimental results demonstrate that the proposed algorithm provides visually satisfying panoramas with all meaningful foreground objects, but without severe artifacts in the backgrounds.

P4-4
TitleA Sub-Pixel Virtual View Synthesis Method for Multiple View Synthesis
AuthorXin Tong, Ping Yang (Tsinghua University, China), Xiaozhen Zheng, Jianhua Zheng (Hisilicon Technologies Co. Ltd., China), Yun He (Tsinghua University, China)
Pagepp. 490 - 493
KeywordSPVVS, virtual view synthesis
AbstractA sub-pixel virtual view synthesis (SPVVS) method is proposed in this paper. In the proposed method, by promoting the sampling rate of the target virtual view, the sub-pixel information of the corresponding pixel position among multiple input views with sub-pixel displacement can be utilized. A directional adaptive image interpolation is manipulated to generate a high resolution intermediate image, which is then down-sampled to obtain the target synthesized virtual view. The realization procedure of SPVVS is also presented. Experiment result shows significant improvements on subjective quality compared to traditional integer pixel synthesis method. Artifacts such as ‘hat’ effects can be significantly reduced.

P4-5
TitleImproving the Visual Quality of AVC/H.264 by Combining It with Content Adaptive Depth Map Compression
AuthorChristian Keimel, Klaus Diepold (TU Muenchen, Germany), Michel Sarkis (Sony Deutschland GmbH, Germany)
Pagepp. 494 - 497
KeywordAVC/H.264, 3D scene analysis, 3DTV, Depth map comression, content adaptive meshing
AbstractThe future of video coding for 3DTV lies in the combination of depth maps and corresponding textures. Most current video coding standards, however, are only optimized for visual quality and are not able to efficiently compress depth maps. We present in this work a content adaptive depth map meshing with tritree and entropy encoding for 3D videos. We show that this approach outperforms the intra frame prediction of AVC/H.264 for the coding of depth maps of still images. We also demonstrate by combining AVC/H.264 with our algorithm that we are able to increase the visual quality of the encoded texture on average by 6~dB. This work is currently limited to still images but an extension to intra coding of 3D video is straightforward.

P4-6
TitleError Concealment for MVC and 3D Video Coding
AuthorOlgierd Stankiewicz, Krzysztof Wegner, Marek Domański (Poznan University of Technology, Poland)
Pagepp. 498 - 501
KeywordError concealment, MVC, 3D video, depth maps, cross-checking
AbstractIn this paper we propose a novel approach to error concealment that can be applied to MVC and other 3D video coding technologies. The image content, that is lost due to errors, is recovered with use of multiple error-concealment techniques. In our work we have used three techniques: well-known temporal- and intra-based techniques and a novel inter-view technique. Proposed inter-view recovery employs Depth Image Based Rendering (DIBR), which requires neighboring views and corresponding depth maps. Those depth maps can be delivered in the bit-stream or estimated in the receiver. In order to obtain the final reconstruction, the best technique is selected locally. For that, an original recovery quality measurement method, based on cross-checking, has been proposed. The idea has been implemented and assessed experimentally, with use of 3D video test sequences. The objective and subjective results show that the proposed approach provide good quality of reconstructed video.

P4-7
TitleDifference Detection Based Early Mode Termination for Depth Map Coding in MVC
AuthorMinghui Wang, Xin Jin, Satoshi Goto (Waseda University, Japan)
Pagepp. 502 - 505
KeywordMVC, mode decision, depth map, difference detection
AbstractDepth map coding is a new topic in multiview video coding (MVC) following the development of depth-image-based rendering (DIBR). Since depth map is monochromatic and has less texture than color map, fast algorithm is necessary and possible to reduce the computation burden of the encoder. This paper proposed a difference detection based early mode termination strategy. The difference detection (DD) algorithms are categorized to reconstructed frame based (RDD) and original frame based.(ODD). And a simplified ODD (sODD) is also proposed. Early mode termination based on these three DD algorithms are implemented and evaluated in the reference software of JMVC8.0 respectively. Simulation results indicate that RDD based one has no performance lost and reduce 25% runtime on average. ODD and sODD based ones can save 54.3% and 43.6% runtime respectively and have a acceptable R-D performance lost.

P4-8
TitleFast Stereo Matching with Predictive Search Range
AuthorYu-Cheng Tseng, Po-Hsiung Hsu, Tian-Sheuan Chang (Dept. of Electronics Engineering & Institute of Electronics, National Chiao Tung University, Taiwan)
Pagepp. 506 - 509
Keyworddisparity estimation, stereo matching
AbstractLocal stereo matching could deliver accurate disparity maps by the associated method, like adaptive support-weight, but suffers from the high computational complexity, O(NL), where N is pixel count in spatial domain, and L is search range in disparity domain. This paper proposes a fast algorithm that groups similar pixels into super-pixels for spatial reduction, and predicts their search range by simple matching for disparity reduction. The proposed algorithm could be directly applied to other local stereo matching, and reduce its computational complexity to only 8.2%-17.4% with slight 1.5%-3.2% of accuracy degradation.

P4-9
TitleInfluences of Frame Delay and Packet Loss between Left and Right Frames in Stereoscopic Video Communications
AuthorShuliang Lin, Yuichiro Sawa, Norishige Fukushima, Yutaka Ishibashi (Nagoya Institute of Technology, Japan)
Pagepp. 510 - 513
Keywordstereo video, network delay, packet loss, subjective assessment, inter-media synchronization
AbstractThis paper analyzes the influences of frame delay and packet loss in stereoscopic vision when stereoscopic video transferred over a IP network. We employ live action videos which are transferred to a head-mount-display (HMD) and do the assessment on stereoscopic perception. As a result, we found that speed and movement direction of the attention object play a great role on the deterioration when frame delay and packet loss occurs.


[Beyond H.264/MPEG-4 AVC and related topics]

P4-10
TitleA Novel Inloop Filter for Video-Compression Based on Temporal Pixel Trajectories
AuthorMarko Esche, Andreas Krutz, Alexander Glantz, Thomas Sikora (Technische Universität Berlin, Germany)
Pagepp. 514 - 517
Keywordvideo-compression, temporal inloop filter, pixel trajectories, deblocking
AbstractThe objective of this work is to investigate the performance of a new inloop filter for video compression, which uses temporal rather than spatial information to improve the quality of reference frames used for prediction. The new filter has been integrated into the H.264/AVC baseline encoder and tested on a wide range of sequences. Experimental results show that the filter achieves a bit rate reduction of up to 12% and more than 4% on average without increasing the complexity of either encoder or decoder significantly.

P4-11
TitleFast Rate-Distortion Optimized Transform for Intra Coding
AuthorXin Zhao (Institute of Computing Technology, Chinese Academy of Sciences; Graduate University of Chinese Academy of Sciences, China), Li Zhang, Siwei Ma, Wen Gao (Institute of Digital Media, Peking University, China)
Pagepp. 518 - 521
Keywordrate-distortion optimization (RDO), Intra coding, mode-dependent directional transform (MDDT), rate-distortion optimized transform (RDOT)
AbstractIn our previous work, rate-distortion optimized transform (RDOT) is introduced. The proposed RDOT achieves remarkable coding gain for KTA Intra coding, but the computational complexity is increased drastically at the encoder. To solve this problem, we propose a fast RDOT scheme using macroblock- and block-level R-D cost thresholding. With the proposed methods, unnecessary mode trials can be efficiently skipped from the encoding process. Experimental results show that, with negligible performance degradation, about 88.9% of the total encoding time is saved.

P4-12
TitleA Hierarchical Variable-Sized Block Transform Coding Scheme for Coding Efficiency Improvement on H.264/AVC
AuthorBumshik Lee, Jaeil Kim, Sangsoo Ahn, Munchurl Kim (Korea Advanced Institute of Science and Technology, Republic of Korea), Hui Yong Kim, Jongho Kim, Jin Soo Choi (Electronic Telecommunications Research Institute (ETRI), Republic of Korea)
Pagepp. 522 - 525
KeywordQuadtree Transform, Variable Block-size Transform, Discrete Cosine Transform, H.264/AVC
AbstractIn this paper, a rate-distortion optimized variable block transform coding scheme is proposed based on a quadtree-structured transform for macroblock (MB) coding with a set of the order-4 and -8 integer cosine transform (ICT) kernels of H.264/AVC as well as a new order-16 ICT kernel. The set of order-4, -8 and -16 ICT kernels are applied for inter-predictive coding in square (4x4, 8x8 or 16x16) or non-square (16x8 or 8x16) transform for each MB in a quadtree structured manner. The proposed quadtree-structured variable block transform scheme using the order-16 ICT kernel achieves significant bitrate reduction up to 15%, compared to the High profile of H.264/AVC. Even if the number of candidates for the transform types increases, the encoding time can be reduced to average 4-6% over the H.264/AVC.

P4-13
TitleEnhanced Region-Based Adaptive Interpolation Filter
AuthorShohei Matsuo, Yukihiro Bandoh, Seishi Takamura, Hirohisa Jozawa (NTT Cyber Space Laboratories, NTT Corporation, Japan)
Pagepp. 526 - 529
Keywordmotion compensation, adaptive interpolation filter, region-division, edge, image locality
AbstractAdaptive interpolation filter (AIF) was proposed to improve motion compensation. The conventional AIF optimizes the filter coefficients on a frame-by-frame basis. However, when the image is divided into multiple regions, each of which has different characteristics, the coding efficiency could improve by performing optimization on a region-by-region basis. In this paper, we proposed a region-based AIF (RBAIF). Simulation results showed that RBAIF offered about 0.43% and 5.05% higher coding gain than the conventional AIF and the H.264/AVC filter, respectively.

P4-14
TitleFractional-Sample Motion Compensation Using Generalized Interpolation
AuthorHaricharan Lakshman, Benjamin Bross, Heiko Schwarz, Thomas Wiegand (Fraunhofer Institute for Telecommunications, Heinrich Hertz Institute, Germany)
Pagepp. 530 - 533
Keywordvideo coding, motion-compensated prediction, reference picture upsampling, B-splines
AbstractTypical interpolation methods in video coding perform filtering of reference picture samples using FIR filters for motion compensated prediction. This process can be viewed as a signal decomposition using basis functions which are restricted by the interpolating constraint. Using the concept of generalized interpolation provides a greater degree of freedom for selecting basis functions. We implemented generalized interpolation using a combination of IIR and FIR filters. The complexity of the proposed scheme is comparable to that of an 8-tap FIR filter. Bit rate savings up to 20% compared to the H.264/AVC 6-tap filter are shown.


[Image/video coding and related topics]

P4-15
TitleImage Coding Approach Based On Image Decomposition
AuthorYunhui Shi, Yanli Hou, Baocai Yin, Wenpeng Ding (Beijing University of Technology, China)
Pagepp. 534 - 537
Keywordimage decomposition, texture synthesis, region selection, image coding
AbstractTextures in many images or video scenes are difficult to code because of the large amount of visible detail. This paper proposes an image coding approach to solve this problem, in which we incorporate image decomposition and texture synthesis technology into the image coding framework. The key idea of our approach is to first decompose the original image into cartoon component u and texture component v with different basic characteristics, and then to synthesize the selected texture regions in texture component v. The cartoon component u and the non-synthetic regions in texture component v are compressed by JPEG. Experimental results show bit-rate savings of over 30% compared with JPEG at similar visual quality levels.

P4-16
TitleA Real-Time System of Distributed Video Coding
AuthorKazuhito Sakomizu, Takahiro Yamasaki, Satoshi Nakagawa, Takashi Nishi (Oki Electric Industry Co., Ltd., Japan)
Pagepp. 538 - 541
Keyworddistributed video coding, real-time system, Slepian-Wolf theorem, Wyner-Ziv theorem
AbstractThis paper presents a real-time system of distributed video coding (DVC). The decoding process of DVC is normally complex, which causes difficulty in real-time implementation. To address this problem, we propose a new configuration of DVC with three methods. Then we implement the system with parallelization techniques. Experimental results show that the encoder on i.MX31 400 MHz could operates at about CIF 13 fps, and the decoder on Core 2 Quad 2.83 GHz operates at more than CIF 30 fps.

P4-17
TitleBlock-Based Second Order Prediction on AVS-Part 2
AuthorBinbin Yu, Shangwen Li, Lu Yu (Institute of Information and Communication Engineering, Zhejiang University, China)
Pagepp. 542 - 545
KeywordBlock-based Second Order Prediction, Motion-compensated prediction, Mode prediction, Directional operators, AVS
AbstractAVS-Part 2 is a mainstream video coding standard with high compression efficiency similar to H.264/AVC. A technique named Second Order Prediction (SOP) has been presented based on H.264/AVC to decrease the signal correlation after motion-compensated prediction. To achieve better coding performance, this paper presents a method named Block-based Second Order Prediction (BSOP) to ameliorate SOP to adapt to the features of the motion-compensation in AVS-P2 with analysis and demonstration in detail. Experimental results show that the proposed BSOP can outperform AVS-P2 P-picture coding by 3.99% bit-rate saving (0.126dB BD-PSNR gain) on average, and performs better than SOP implemented on AVS by 1.81% bit-rate saving.

P4-18
TitleImproved Local PDF Estimation in the Wavelet Domain for Generalized Lifting
AuthorJulio C. Rolon (National Polytechnic Institute, Mexico), Philippe Salembier (Technical University of Catalonia, Spain)
Pagepp. 546 - 549
KeywordGeneralized lifting, wavelets, pdf estimation, lossy image coding
AbstractGeneralized Lifting has been studied for lossy image compression in [2,3]. It has been demonstrated that the method achieves a significant reduction of the wavelet coefficients energy and entropy. The definition of the GL relies on an estimation of the pdf of the pixel to encode conditioned to a surrounding context. The objective of this paper is to present an improved method for the estimation of the pdf at the local level. We follow the idea of self similarity proposed in [1] for denoising, and propose to estimate the pdf using all the causal contexts within a window. Experimental results show an important increment in the energy and entropy gains when compared to previous strategies [2,3].

P4-19
TitleImage Coding by Using Non-Linear Texture Decomposition and Image Summarization
AuthorChihiro Suzuki, Takamichi Miyata, Yoshinori Sakai (Tokyo Institute of Technology, Japan)
Pagepp. 550 - 553
KeywordTexture Synthesis, TV-regularization, Bidirectional Similarity, Image Coding
AbstractTV-regularization can be used to decompose any natural image into a structure image(S) and a texture image(T). We proposed a novel image coding method that is coding separately these images. We make a compaction image that has all texture patterns. Then the compaction image is divided into S and T. Then, the encoder sends the compaction’s S, T and input’s S to the decoder. At the decoder, original-size T is synthesized from compaction’s S by matching the compaction’s S and the original-size S.

P4-20
TitleCoding Efficiency Improvement Using Inter-Picture Processing of Quantization Error
AuthorKenji Sugiyama, Naoya Sagara, Masao Arizumi (Seikei University, Japan)
Pagepp. 554 - 557
KeywordCompatibility, I-picture coding, B-picture coding, Motion compensation, Quantization error
AbstractAs standard video encoder techniques have matured, their rate of improvement has slowed. As an alternative, a new coding concept with semi-compatibility has been proposed, and an enhancement to I-pictures efficiency has been discussed. This applied method reduces the quantization error using motion compensated inter-picture processing. In this report, we apply this method to P-pictures to improve the efficiency of B-pictures. The quantization error component of the prediction signal is canceled by averaging the bi-directional prediction. Experiments using MPEG-4 show significant improvement in the coding efficiency using the proposed method. The maximum PSNR gain reaches 2.3 dB in a static sequence. At least 0.5 dB can be achieved in a high motion sequence.

P4-21
TitleMultiple Description Video Transcoding with Temporal Drift Control
AuthorPedro Correia, Pedro Assunção, Vitor Silva (Instituto de Telecomunicações, Portugal)
Pagepp. 558 - 561
KeywordMultiple Description Transcoding, Multiple Description Scalar Quantisation, Drift Distortion
AbstractThis paper proposes a multiple description (MD) transcoding scheme capable of preventing drift by distortion accumulation in temporally predicted motion compensated slices. Drift compensation is achieved by generating a controlled amount of side information to be used for decoding whenever a description fails to reach the end user terminal. The side information is generated by re-encoding the transcoding residue with an independent quantisation parameter which also controls redundancy. A simplified architecture is devised to reduce transcoding complexity in regard to the number of processing functions and buffer requirements. The experimental results show that temporally predicted frames do not suffer from drift and their quality is significantly improved at reduced redundancy cost in comparison with a classic MD transcoding scheme.

P4-22
TitleH.264/AVC to Wavelet-Based Scalable Video Transcoding Supporting Multiple Coding Configurations
AuthorEduardo Peixoto, Toni Zgaljic, Ebroul Izquierdo (Queen Mary, University of London, U.K.)
Pagepp. 562 - 565
KeywordTranscoding, Scalable Video Coding
AbstractScalable Video Coding (SVC) enables low complexity adaptation of the compressed video, providing an efficient solution for video content delivery through heterogeneous networks and to different displays. However, legacy video and most commercially available content capturing devices use conventional non-scalable coding, e.g., H.264/AVC. This paper proposes an efficient transcoder from H.264/AVC to a wavelet-based SVC to exploit the advantages offerend by the SVC technology. The proposed transcoder is able to cope with different coding configurations in H.264/AVC, such as IPP or IBBP with multiple reference frames. To reduce the transcoder's complexity, motion information and presence of the residual data extracted from the decoded H.264/AVC video are exploited. Experimental results show a good performance of the proposed transcoder in terms of decoded video quality and system complexity.

P4-23
TitleEdge-Adaptive Transforms for Efficient Depth Map Coding
AuthorGodwin Shen, Woo-shik Kim, Sunil Kumar Narang, Antonio Ortega (University of Southern California, U.S.A.), Jaejoon Lee, HoCheon Wey (Samsung Advanced Institute of Technology, Republic of Korea)
Pagepp. 566 - 569
KeywordMultiview plus depth (MVD), Depth coding, Rate-distortion optimization
AbstractIn this work a new set of edge-adaptive transforms (EATs) is presented as an alternative to the standard DCTs used in image and video coding applications. These transforms avoid filtering across edges in each image block, thus, they avoid creating large high frequency coefficients. These transforms are then combined with the DCT in H.264/AVC and a transform mode selection algorithm is used to choose between DCT and EAT in an RD-optimized manner. These transforms are applied to coding depth maps used for view synthesis in a multi-view video coding system, and provides up to 29% bit rate reduction for a fixed quality in the synthesized views.

P4-24
TitleDirection-Adaptive Hierarchical Decomposition for Image Coding
AuthorTomokazu Murakami, Keita Takahashi, Takeshi Naemura (The University of Tokyo, Japan)
Pagepp. 570 - 573
KeywordImage decomposition, directional prediction, L1 norm, directional transform, wavelet
AbstractA new model of decomposing an image hierarchically into direction-adaptive subbands using pixel-wise direction estimation is presented. For each decomposing operation, an input image is divided into two parts: a base image subsampled from the input image and subband components. The subband components consist of residuals of estimating the pixels skipped through the subsampling, which ensures the invertibility of the decomposition. The estimation is performed in a direction-adaptive way, whose optimal direction is determined by a L1 norm criterion for each pixel, aiming to achieve good energy compaction that is suitable for image coding. Furthermore, since the L1 norms are obtained from the base image alone, we do not need to retain the directional information explicitly, which is another advantage of our model. Experimental results show that the proposed model can achieve lower entropy than conventional Haar or D5/3 discrete wavelet transform in case of lossless coding.


[Image/video processing and related topics]

P4-25
TitleA Robust Video Super-Resolution Algorithm
AuthorXinfeng Zhang (Institute of Computing Technology, Chinese Academy of Sciences, China), Ruiqin Xiong, Siwei Ma, Wen Gao (School of Electronic Engineering and Computer Science, Peking University, China)
Pagepp. 574 - 577
Keywordsuper-resolution, kernel regression, irregular interpolation
AbstractIn this paper, we proposed a robust video super-resolution reconstruction method based on spatial-temporal orienta-tion-adaptive kernel regression. First, we proposed a robust registration efficiency model to reflect the temporal information reliability. Second, we proposed a spatial-temporal steering kernel considering motions between frames and structures in each low resolution frame. Simula-tion results demonstrate that our new super-resolution method substantially improves both the subjective quality and objective quality than other resolution enhancement methods.

P4-26
TitleAn Efficient Method for the Detection of Ringing Artifacts and De-Ringing in JPEG Image
AuthorShen-Chuan Tai, Bo-Jhih Chen, Mankit Choi (Institute of Computer and Communication Engineering, Department of Electrical Engineering, National Cheng Kung University, Taiwan)
Pagepp. 578 - 581
KeywordJPEG, ringing artifacts, image compression
AbstractJPEG standard is commonly used for the still image compression. However, the DCT-based coding of JPEG is one of the lossy compression tools and introduces most artifacts in the decompressed image such as blocking and ringing artifacts. In this paper, our proposed method focuses on how to efficiently detect the blocks which cause ringing artifacts. Moreover, these ringing blocks will be further filtered and the texture region will be able to preserve as well as the smooth region. Simulation results show that our proposed method can outperform the related algorithms in subjectively and objectively. By comparing JPEG compressed image, the decompressed image using our algorithm can achieve the better PSNR as well as the visual performance, especially at the lower quality coding (the higher compression rate).

P4-27
TitleLow Delay Distributed Video Coding Using Data Hiding
AuthorKrishna Rao Vijayanagar, Bowen Dan, Joohee Kim (Illinois Institute of Technology, U.S.A.)
Pagepp. 582 - 585
KeywordDistributed Video Coding, Data hiding, Low Delay DVC
AbstractDistributed Video Coding (DVC) is a popular topic in the research community and the past years have seen several different implementations. DVC has been proposed as a solution for applications that have limited battery resources and low hardware complexity, thus necessitating a low complexity encoder. An ideal application would be in remote surveillance/monitoring or live video conferencing. However, current solutions use iteratively decodable channel codes like LDPCA or Turbo codes that have large latencies. In order to make real-time communication possible. The proposed architecture makes efficient use of Skip blocks to reduce the bitrate, eliminates the iterative decoding nature of the Wyner-Ziv (WZ) channel and uses a simple data-hiding based compression algorithm. This drastically cuts down on the time complexity of the decoding procedure while still maintaining an rate-distortion performance better than that of H.264/AVC Intra coding and other current DVC solutions.

P4-28
TitleFFT-Based Full-Search Block Matching Using Overlap-Add Method
AuthorHidetake Sasaki, Zhen Li, Hitoshi Kiya (Tokyo Metropolitan University, Japan)
Pagepp. 586 - 589
Keywordblock matching, FFT, overlap-add method, pattern recognition, motion estimation
AbstractOne category of fast full-search block matching algorithms (BMAs) is based on the fast Fourier transformation (FFT). In conventional methods in this category, the macroblock size must be adjusted to the search window size by zero-padding. In these methods, the memory consumption and computational complexity heavily depend on the size difference between the macroblock and the search window. Thus, we propose a novel FFT-based BMA to solve this problem. The proposed method divides the search window into multiple sub search windows to versatilely control the difference between the macroblock and the search window sizes. Simulation results show the effectiveness of the proposed method.


[Quality, system, applications, and other topics]

P4-29
TitleTemporal Inconsistency Measure for Video Quality Assessment
AuthorSongnan Li, Lin Ma, Fan Zhang, King Ngi Ngan (The Chinese University of Hong Kong, Hong Kong)
Pagepp. 590 - 593
Keywordvideo quality assessment, spatial visual quality measure, temporal inconsistency measure
AbstractVisual quality assessment plays a crucial role in many vision related signal processing applications. In the literature, more efforts have been spent on spatial visual quality measure. Although a large number of video quality metrics have been proposed, the methods to use temporal information for quality assessment are less diversified. In this paper, we propose a novel method to measure the temporal impairments. The proposed method can be incorporated into any image quality metric to extend it into a video quality metric. Moreover, it is easy to apply the proposed method in video coding system to incorporate with MSE for rate-distortion optimization.

P4-30
TitleThe Dependence of Visual Noise Perception on Background Color and Luminance
AuthorMakoto Shohara, Kazunori Kotani (Japan Advanced Institute of Science and Technology, Japan)
Pagepp. 594 - 597
KeywordVisual system, Noise measurement, Color measurement, Noise generators, Shot noise
AbstractThis paper describes the dependency of noise perception on background color and luminance of noise quantitatively. We use the luminance and chromatic noise models derived from shot noise model. And we conduct subjective and quantitative experiments using a modified gray scale method. The subjective experiment results show the perceived color noise depends on the background color, but the perceived luminance noise does not. In addition, the perceived chromatic noise level is about 8 times smaller than calculated color noise.

P4-31
TitleAn Adaptive Low-Complexity Global Motion Estimation Algorithm
AuthorMd Nazmul Haque, Moyuresh Biswas, Mark R. Pickering, Michael R. Frater (The University of New South Wales, Australia)
Pagepp. 598 - 601
Keywordglobal motion estimation, video coding, image registration, gradient-descent optimization
AbstractA limitation of current global motion estimation approaches is the additional complexity of the gradient-descent optimization that is typically required to calculate the optimal set of global motion parameters. In this paper we propose a new low-complexity algorithm for global motion estimation. The complexity of the proposed algorithm is reduced by performing the majority of the operations in the gradient-descent optimization using logic operations rather than full-precision arithmetic operations. This use of logic operations means that the algorithm can be implemented much more easily in hardware platforms such as field programmable gate arrays (FPGAs). Experimental results show that the execution time for software implementations of the new algorithm is reduced by a factor of almost four when compared to existing fast implementations without any significant loss in registration accuracy.

P4-32
TitleScalable Multiple Description Video Coding Using Successive Refinement of Side Quantizers
AuthorMuhammad Majid, Charith Abhayaratne (The University of Sheffield, U.K.)
Pagepp. 602 - 605
KeywordMultiple description coding, quality scalability, resilience
AbstractIn this paper, we present a new method for scalable multiple description video coding based on motion compensated temporal filtering and multiple description scalar quantizer with successive refinement. In our method quality scalability is achieved by successively refining the side quantizers of a multiple description scalar quantizer. The rate of each description is allocated by considering different refinement levels for each spatio-temporal subband. The performance of the proposed scheme under lossless and lossy channel conditions are presented and compared with single scalable description video coding.

P4-33
TitleBit-Plane Compressive Sensing with Bayesian Decoding for Lossy Compression
AuthorSz-Hsien Wu (Electronic Engineering, National Chiao Tung University, Taiwan), Wen-Hsiao Peng (Computer Science, National Chiao Tung University, Taiwan), Tihao Chiang (Electronic Engineering, National Chiao Tung University, Taiwan)
Pagepp. 606 - 609
KeywordCompressive Sensing, Bayesian estimation, Bit-plane
AbstractThis paper addresses the problem of reconstructing a compressively sampled sparse signal from its lossy and possibly insufficient measurements. The process involves estimations of sparsity pattern and sparse representation, for which we derived a vector estimator based on the Maximum a Posteriori Probability (MAP) rule. By making full use of signal prior knowledge, our scheme can use a measurement number close to sparsity to achieve perfect reconstruction. It also shows a much lower error probability of sparse pattern than prior work, given insufficient measurements. To better recover the most significant part of the sparse representation, we further introduce the notion of bit-plane separation. When applied to image compression, the technique in combination with our MAP estimator shows promising results as compared to JPEG: the difference in compression ratio is seen to be within a factor of two, given the same decoded quality.

P4-34
TitleA Reduced-Reference Metric Based on the Interest Points in Color Images
AuthorMichael Nauge, Mohamed-Chaker Larabi, Christine Fernandez (University of Poitiers, France)
Pagepp. 610 - 613
KeywordMetric, Interest point, Reduced reference, quality, saliency
AbstractIn the last decade, an important research effort has been dedicated to quality assessment from the subjective and the objective points of view. The focus was mainly on Full Reference (FR) metrics because of the ability to compare to an original. Only few works were oriented to Reduced Reference (RR) or No Reference (NR) metrics, very useful for applications where the original image is not available such as transmission or monitoring. In this work, we propose an RR metric based on two concepts, the interest points of the image and the objects saliency on color images. This metric needs a very low amount of data (lower than 8 bytes) to be able to compute the quality scores. The results show a high correlation between the metric scores and the human judgement and a better quality range than well-known metrics like PSNR or SSIM. Finally, interest points have shown that they can predict the quality of color images.