Academia.edu no longer supports Internet Explorer.
To browse Academia.edu and the wider internet faster and more securely, please take a few seconds to upgrade your browser.
2009
abstract In this paper we present a new method for superresolution of depth video sequences using high resolution color video. Here we assume that the depth sequence does not contain outlier points which can be present in the depth images. Our method is based on multiresolution decomposition, and uses multiple frames to search for a most similar depth segments to improve the resolution of the current frame. First step is the wavelet decomposition of both color and depth images.
Circuits Systems and Signal Processing, 2000
Superresolution produces high-quality, high-resolution images from a set of degraded, low-resolution images where relative frame-to-frame motions provide different looks at the scene. Superresolution translates data temporal bandwidth into enhanced spatial resolution. If considered together on a reference grid, given low-resolution data are nonuniformly sampled. However, data from each frame are sampled regularly on a rectangular grid. This special type of nonuniform sampling is called interlaced sampling. We propose a new wavelet-based interpolation-restoration algorithm for superresolution. Our efficient wavelet interpolation technique takes advantage of the regularity and structure inherent in interlaced data, thereby significantly reducing the computational burden. We present one-and two-dimensional superresolution experiments to demonstrate the effectiveness of our algorithm.
2014
I wish to thank my advisor Evgeny Strekalovskiy for answering my numerous doubts during the elaboration of the thesis. I wish also to thank Prof. Dr. Cremers and the chair of Computer Vision to allow me to do a Master Thesis with them at the Tecnische Universität München. I would also like to express my special thanks of gratitude to my brother without him I would not have discovered the pleasures of informatics, as well as to Raúl who served me as inspiration of how to manage a Master Thesis.
2007
ABSTRACT It is believed by many that three-dimensional (3D) television will be the next logical development toward a more natural and vivid home entertaiment experience. While classical 3D approach requires the transmission of two video streams, one for each view, 3D TV systems based on depth image rendering (DIBR) require a single stream of monoscopic images and a second stream of associated images usually termed depth images or depth maps, that contain per-pixel depth information.
Superresolution produces high-quality, high-resolution images from a set of degraded, low-resolution images where relative frame-to-frame motions provide differ- ent looks at the scene. Superresolution translates data temporal bandwidth into enhanced spatial resolution. If considered together on a reference grid, given low-resolution data are nonuniformly sampled. However, data from each frame are sampled regularly on a rectangular grid. This special type of nonuniform sampling is called interlaced sampling. We propose a new wavelet-based interpolation-restoration algorithm for superresolution. Our efficient wavelet interpolation technique takes advantage of the regularity and structure inherent in interlaced data, thereby significantly reducing the computational burden. We present one- and two-dimensional superresolution experiments to demonstrate the effec- tiveness of our algorithm.
Lecture Notes in Computer Science, 2013
We use multi-frame super-resolution, specifically, Shift & Add, to increase the resolution of depth data. In order to be able to deploy such a framework in practice, without requiring a very high number of observed low resolution frames, we improve the initial estimation of the high resolution frame. To that end, we propose a new data model that leads to a median estimation from densely upsampled low resolution frames. We show that this new formulation solves the problem of undefined pixels and further allows to improve the performance of pyramidal motion estimation in the context of super-resolution without additional computational cost. As a consequence, it increases the motion diversity within a small number of observed frames, making the enhancement of depth data more practical. Quantitative experiments run on the Middlebury dataset show that our method outperforms state-of-the-art techniques in terms of accuracy and robustness to the number of frames and to the noise level.
2010
The joint usage of low- and full-resolution images in multiview systems provides an attractive opportunity for data size reduction while maintaining good quality in 3D applications. In this paper we present a novel application of a super-resolution method for usage within a mixed resolution multiview setup. The technique borrows high-frequency content from neighboring full resolution images to enhance particular low-resolution views. Occlusions are handled through matching of low-resolution images. Both the stereo and the more general multiview cases are considered using the multiview video-plus-depth format. Results demonstrate significant gains in PSNR and in visual quality for test sequences.
2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2015
This paper proposes to enhance low resolution dynamic depth videos containing freely non-rigidly moving objects with a new dynamic multi-frame super-resolution algorithm. Existent methods are either limited to rigid objects, or restricted to global lateral motions discarding radial displacements. We address these shortcomings by accounting for non-rigid displacements in 3D. In addition to 2D optical flow, we estimate the depth displacement, and simultaneously correct the depth measurement by Kalman filtering. This concept is incorporated efficiently in a multi-frame super-resolution framework. It is formulated in a recursive manner that ensures an efficient deployment in real-time. Results show the overall improved performance of the proposed method as compared to alternative approaches, and specifically in handling relatively large 3D motions. Test examples range from a full moving human body to a highly dynamic facial video with varying expressions.
Symmetry
The super-resolution (SR) technique reconstructs a high-resolution image from single or multiple low-resolution images. SR has gained much attention over the past decade, as it has significant applications in our daily life. This paper provides a new technique of a single image super-resolution on true colored images. The key idea is to obtain the super-resolved image from observed low-resolution images. A proposed technique is based on both the wavelet and spatial domain-based algorithms by exploiting the advantages of both of the algorithms. A back projection with an iterative method is implemented to minimize the reconstruction error and for noise removal wavelet-based de-noising method is used. Previously, this technique has been followed for the grayscale images. In this proposed algorithm, the colored images are taken into account for super-resolution. The results of the proposed method have been examined both subjectively by observation of the results visually and objectively...
Image Reconstruction from Incomplete Data III, 2004
In the last two decades a variety of super-resolution (SR) methods have been proposed. These methods usually address the problem of fusing a set of monochromatic images to produce a single monochromatic image with higher spatial resolution. In this paper we address the dynamic and color SR problems of reconstructing a high-quality set of colored super-resolved images from low-quality mosaiced frames. Our approach includes a hybrid method for simultaneous SR and demosaicing, this way taking into account practical color measurements encountered in video sequences. For the case of translational motion and common space-invariant blur, the proposed method is based on a very fast and memory efficient approximation of the Kalman filter. Experimental results on both simulated and real data are supplied, demonstrating the presented algorithm, and its strength.
Robust depth map refining using color image, 2024
Depth maps are essential for various applications, providing spatial information about object arrangement in a scene. They play a crucial role in fields such as computer vision, robotics, augmented and virtual reality, autonomous systems, and medical imaging. However, generating accurate, high-quality depth maps is challenging due to issues like texture-copying artifacts, edge leakage, and depth edge distortion. This study introduces a novel method for refining depth maps by integrating information from color images, combining structural and statistical techniques for superior results. The proposed approach employs a structural method to calculate affinities within a regularization framework, utilizing minimum spanning trees (MST) and minimum spanning forests (MSF). Superpixel segmentation is used to prevent MST construction across depth edges, addressing edge-leaking artifacts while preserving details. An edge inconsistency measurement model further reduces texture-copying artifacts. Additionally, an adaptive regularization window dynamically adjusts its bandwidth based on local depth variations, enabling effective handling of noise and maintaining sharp depth edges. Experimental evaluations across multiple datasets show the method's robustness and accuracy. It consistently achieves the lowest mean absolute deviation (MAD) compared to existing techniques across various upsampling factors, including 2×, 4×, 8×, and 16×. Visual assessments confirm its ability to produce depth maps free of texture-copying artifacts and blurred edges, yielding results closest to ground truth. Computational efficiency is ensured through a divide-and-conquer algorithm for spanning tree computations, reducing complexity while maintaining precision. This research underscores the importance of combining structural and statistical information in depth map refinement. By overcoming the limitations of existing methods, the proposed approach provides a practical solution for improving depth maps in applications requiring high precision and efficiency, such as robotics, virtual reality, and autonomous systems. Future work will focus on real-time applications and integration with advanced depth-sensing technologies.
Wavelet Applications in Industrial Processing V, 2007
In this paper we present a new method for joint denoising of depth and luminance images produced by time-of-flight camera. Here we assume that the sequence does not contain outlier points which can be present in the depth images. Our method first performs estimation of noise and signal covariance matrices and then performs vector denoising. Luminance image is segmented into similar contexts using k-means algorithm, which are used for calculation of covariance matrices. Denoising results are compared with the ground truth images obtained by averaging of the multiple frames of the still scene.
2007
abstract In this paper we present a new method for joint denoising of depth and luminance images produced by time-of-flight camera. Here we assume that the sequence does not contain outlier points which can be present in the depth images. Our method first performs estimation of noise and signal covariance matrices and then performs vector denoising. Luminance image is segmented into similar contexts using k-means algorithm, which are used for calculation of covariance matrices.
3DTV Conference, 2007, 2007
In this paper we present a new method for joint denoising of depth and luminance images produced by time-of-flight camera. Here we assume that the sequence does not contain outlier points which can be present in the depth images. Our method first performs estimation of noise and signal covariance matrices and then performs vector denoising. Two versions of the algorithm are presented, depending on the method used for the classification of the image contexts. Denoising results are compared with the ground truth images obtained by averaging of the multiple frames of the still scene.
2012
ABSTRACT The recent development of low-cost and fast time-of-flight cameras enabled measuring depth information at video frame rates. Although these cameras provide invaluable information for many 3D applications, their imaging capabilities are very limited both in terms of resolution and noise level. In this paper, we present a novel method for obtaining a high resolution depth map from a pair of a low resolution depth map and a corresponding high resolution color image.
EURASIP Journal on Advances in Signal Processing, 2011
This article proposes an efficient wavelet-based depth video denoising approach based on a multihypothesis motion estimation aimed specifically at time-of-flight depth cameras. We first propose a novel bidirectional block matching search strategy, which uses information from the luminance as well as from the depth video sequence. Next, we present a new denoising technique based on weighted averaging and wavelet thresholding. Here we take into account the reliability of the estimated motion and the spatial variability of the noise standard deviation in both imaging modalities. The results demonstrate significantly improved performance over recently proposed depth sequence denoising methods and over state-of-the-art general video denoising methods applied to depth video sequences.
IEEE transactions on image processing : a publication of the IEEE Signal Processing Society, 2017
Accurate and high-quality depth maps are required in lots of 3D applications, such as multi-view rendering, 3D reconstruction and 3DTV. However, the resolution of captured depth image is much lower than that of its corresponding color image, which affects its application performance. In this paper, we propose a novel depth map super-resolution (SR) method by taking view synthesis quality into account. The proposed approach mainly includes two technical contributions. First, since the captured low-resolution (LR) depth map may be corrupted by noise and occlusion, we propose a credibility based multi-view depth maps fusion strategy, which considers the view synthesis quality and interview correlation, to refine the LR depth map. Second, we propose a view synthesis quality based trilateral depth-map up-sampling method, which considers depth smoothness, texture similarity and view synthesis quality in the up-sampling filter. Experimental results demonstrate that the proposed method outp...
IEEE Signal Processing Letters, 2013
Multimedia Tools and Applications
Different from the existing super-resolution (SR) reconstruction approaches working under either the frequency-domain or the spatial-domain, this paper proposes an improved video SR approach based on both frequency and spatial-domains to improve the spatial resolution and recover the noiseless high-frequency components of the observed noisy low-resolution video sequences with global motion. An iterative planar motion estimation algorithm followed by a structure-adaptive normalised convolution reconstruction method are applied to produce the estimated low-frequency sub-band. The discrete wavelet transform process is employed to decompose the input low-resolution reference frame into four sub-bands, and then the new edge-directed interpolation method is used to interpolate each of the high-frequency sub-bands. The novelty of this algorithm is the introduction and integration of a nonlinear soft thresholding process to filter the estimated high-frequency sub-bands in order to better preserve the edges and remove potential noise. Another novelty of this algorithm is to provide flexibility with various motion levels, noise levels, wavelet functions, and the number of used low-resolution frames. The performance of the proposed method has been tested on three well-known videos. Both visual and quantitative results demonstrate the high performance and improved flexibility of the proposed technique over the conventional interpolation and the state-of-the-art video SR techniques in the wavelet-domain.
Computer Vision – ACCV 2016
Depth image super-resolution is an extremely challenging task due to the information loss in sub-sampling. Deep convolutional neural network have been widely applied to color image super-resolution. Quite surprisingly, this success has not been matched to depth super-resolution. This is mainly due to the inherent difference between color and depth images. In this paper, we bridge up the gap and extend the success of deep convolutional neural network to depth super-resolution. The proposed deep depth super-resolution method learns the mapping from a lowresolution depth image to a high resolution one in an end-toend style. Furthermore, to better regularize the learned depth map, we propose to exploit the depth field statistics and the local correlation between depth image and color image. These priors are integrated in an energy minimization formulation, where the deep neural network learns the unary term, the depth field statistics works as global model constraint and the colordepth correlation is utilized to enforce the local structure in depth images. Extensive experiments on various depth super-resolution benchmark datasets show that our method outperforms the stateof-the-art depth image super-resolution methods with a margin.
EURASIP Journal on Advances in Signal Processing, 2012
View-plus-depth is a scene representation format where each pixel of a color image or video frame is augmented by per-pixel depth represented as gray-scale image (map). In the representation, the quality of the depth map plays a crucial role as it determines the quality of the rendered views. Among the artifacts in the received depth map, the compression artifacts are usually most pronounced and considered most annoying. In this article, we study the problem of post-processing of depth maps degraded by improper estimation or by block-transformbased compression. A number of post-filtering methods are studied, modified and compared for their applicability to the task of depth map restoration and post-filtering. The methods range from simple and trivial Gaussian smoothing, to in-loop deblocking filter standardized in H.264 video coding standard, to more comprehensive methods which utilize structural and color information from the accompanying color image frame. The latter group contains our modification of the powerful local polynomial approximation, the popular bilateral filter, and an extension of it, originally suggested for depth super-resolution. We further modify this latter approach by developing an efficient implementation of it. We present experimental results demonstrating high-quality filtered depth maps and offering practitioners options for highest-quality or better efficiency.