# Efficient registering of color and range images

- Juan D Gomez
^{1}Email author, - Guido Bologna
^{1}and - Thierry Pun
^{1}

**2013**:41

**DOI: **10.1186/1687-5281-2013-41

© Gomez et al.; licensee Springer. 2013

**Received: **8 November 2012

**Accepted: **10 July 2013

**Published: **19 July 2013

## Abstract

We present a simple yet highly efficient method to register range and color images. This method does not rely upon calibration parameters nor does it use visual features analysis. Our assumption is that if the transformation that registers the images is a mathematical function, we can approximate with little number of samples. To this end, thin-plate spline-based interpolations are used in this paper. Therefore, the registration of one point in our method takes only the solving of a nonlinear function. Drastically enhanced performances in the computational processing are attained under this condition. In fact, we show that ultimately our computational algorithm is independent of the complexity of the mathematical model underlying it. Finally, this paper reports on the results of experiments conducted with various range camera models that endorse the proposed method. Eventually, three key features can be derived from our method: practicality, accuracy, and wide applicability.

### Keywords

Time-of-flight Registration Color images Range images Spline## Introduction

With the increasing use of 3D entertainment and multipurpose representation of virtual environments, range cameras continue to gain in popularity as prices are getting lower. While generally promising, there are shortcomings to the use of these sensors, which need to be resolved. Particularly, these cameras lack for color and some do not even provide a gray level or intensity image. This fact dramatically diminishes their scope for expansion into computer vision where image intensity is essential.

Although a number of interesting ideas emerged from this problem, when it comes to couple two camera systems, image registration is perhaps the most affordable approach. Yet classic registration methods yield no suitable results in this particular case. Much is known about intensity images registration; however, there are still many open questions about registering an intensity image and a surface that lacks color and geometric features. In this spirit, the work here introduced presents a general method to register range and red, green, and blue (RGB) digital images. Unlike any other, our approach needs neither calibration of the cameras nor estimation of visual image descriptors.

Further, this paper reports evidence to endorse three features of our method: practicality, accuracy, and wide applicability. Overall, we show that it is a nonlinear function that underlies the registration of depth-color image pairs. In our method, samples of this function are manually taken to construct an interpolated surface model. We use this model to find corresponding shifts that are needed to register points of one image into the other. Also in the general case, nonlinear models are required because images are often corrupted by distortion. Hence, we study as well how to cope with the computation of nonlinear models with no loss of efficiency. In this way, we attain fairly important performance gains in computational terms. In fact, our method might be regarded as an approach to correct distortion in range images, an issue that remains challenging. Finally, when the need of accuracy permits it, our method could use a linear model that yields acceptable results in undistorted cameras such as Kinect. We therefore annex to this paper relevant information in this regard (Appendix).

This article is organized as follows. Section ‘Background’ describes our approach to the problem of registering images from two close-positioned cameras. A step-by-step description of the method proposed in this paper is given in Section ‘A new approach’. Section ‘The shifting basis function’ deals with the mathematical deduction of the basis function underlying our method for range-color image registration. Finally, Section ‘Algorithmic performance, experiments, and comparisons’ presents an overview of practical sides of our approach, namely, algorithmic performance, distortion issues, experiments, and comparisons. Relevant discussion concludes this article with Section ‘Conclusions’.

### State of the art

The registration of RGB and range images of a same scene aims at matching color images and surfaces which lack color [1]. This problem remains largely unexplored in computer vision. Nonetheless, its applicability is fairly well defined. As examples it is worth to mention the following: 3D-laser extrinsic parameters estimation [2], color improvement in depth-color camera pairs [3], and joint-depth and color calibration [4]. Particularly, extrinsic calibration of colorless ToF (time-of-flight) cameras or 3D lasers is a relentless challenge that is usually approached in more refined ways: Hirzinger et al. [5] describe a multi-spline model that requires a robotic arm to know the exact pose of the intended sensor. Zhu et al. [6] describe a high-cost algorithm for fusing stereo-based depth and ToF cameras via triangulation. Unfortunately, a method that is easy-to-use, accurate, and applicable to a wide range of sensor has largely been missing.

An assessment of the general problem of image registration might be useful. In general, a vast range of techniques exists in the literature. Yet, more needs to be done to progress toward general solutions, if any. In 2003 Zitová and Flusser [7] published a complete review of the classic and recent image registration methods. Following them, Deshmukh et al*.* widened the spectrum of solutions by including updated advances in a more recent review in 2011 [8]. In all these works the image registration problem is defined as the matching of images of a scene taken from different sources, viewpoints and/or times. Yet, the former condition (inter-source) is limited to the variability of RGB sources only. Therefore, registration methods such as the one proposed by Yang [9] using artificial neural networks, or others that use belief propagation strategies as is the case of Sun et al*.* in [10], are likely to fail. This is mostly the case because they rely on the matching of color-based visual features common (or mappable) in both images.

In mainstream applications of computer vision, depth and color together as complementary cues about the scene are highly desirable [4, 11–14]. Yet, while low resolution of the ToF camera is enough to segment depth-based areas, higher-resolution RGB camera allows for accurate image processing. In this spirit, Huhle et al*.*
[1] present a novel registration method that combines geometry and color information in order to couple a PMD (photonic mixer device) camera with an external color camera. The alignment carried out in this work is based on the normal distributions transform [15] of the range images and a scale invariant feature transform [16] feature detector applied to the high-resolution color images. Thus, the authors claim to combine the robustness of the globally applicable feature-based approach and the precise local fitting via NDT. More recently, in 2011, Van den Bergh and Van Gool [4] combined a digital camera and SwissRange ToF sensor using a regular calibration method for stereo systems [17]. The key idea of this approach was treating the output of the range sensor as though it was a RGB image.

In [18] the authors conduct a comparative study of some of the most important depth-and-color calibration algorithms. This work includes implementations as well as performance comparisons in both real-world experiments and simulation. Two algorithms stand out in this study: first, Zhang's method [19] that presents a maximum likelihood solution. This method uses checkboards and relies on co-planar assumptions. Also, manual correspondences need to be specified to improve calibration accuracy. Second is Herrera's method [20], in which authors claim to achieve features such as accuracy, practicality, and applicability. The method requires planar surface to be imaged from various poses and presents a new depth distortion model for the depth sensor. One more method called DCCT (depth-camera calibration toolbox) is studied in [18], though the article about this method is to date unpublished. Also, the authors in [19] have not shared their code so we only use [20] for comparisons later in this work.

Finally, as discussed by Han et al*.* in [7], the ‘parallax’ is perhaps the most challenging problem when it comes to image registration. Algorithms suffer from this problem by virtue of the assumption that the scene can be regarded as approximately planar. This is of course not satisfied by large depth variation in the images with raised objects [21]. Paulson et al*.*
[22] presented in 2011 an outstanding idea to cope with the parallax problem by leveraging approximated depth information. Basically, their idea was to recover the depth in the image region with high-rise objects to build accurate transform function for image registration. The drawbacks from which this method suffers are fourfold: motion camera parameters are vital, significant manual work is needed, inaccurate approximations based on heuristics are very likely, and no real time. It is worth noticing that by feeding from a depth source (ToF sensor), the parallax phenomenon is no longer an issue in our method.

### Background

*d*

_{1}and

*d*

_{2}). Furthermore, the following observations may be made on the same figure:

- a.
Only the

*x*-axis is prone to have disparity. - b.
The disparity decreases as the distance of the planes (

*d*_{i}) augments. - c.
Disparity is constant for all the points into parallel planes.

It is worth noticing that (a) will be met provided that the stereo images are first rectified [6]. In other words, both images are rotated (until their epipolar lines get aligned [23]) to allow for disparities in only the horizontal direction (i.e., there is no disparity in the *y* coordinates). Regarding (b), while the relation there pointed might be a common-place observation, the rate at which it is given will be further studied in this section (i.e., ‘Disparity vs. depth’). In fact, it will be shown that this relation is nonlinear. Finally for (c), we want to stress that this is very much expected in an ideal system and this is the reason why we hypothesize about it. Nonetheless, this might not be the case in manually assembled camera rig studied in this paper. Our method, however, performs efficiently in any case.

Ultimately, the actual registration of two images demands a functional description of the displacements (disparities) between parallel planes across the depth. Thus, objects lying on an image plane (left) shall be accurately shifted to their counterparts on the parallel image plane (right). *Our idea is to sample as many pairs of points as possible* (*blue-red pairs in Figure *
2
*) in as many parallel planes as possible, too*. Thus, we can interpolate the function that describes twofold information: firstly, the variation of the disparity between parallel planes, if any; secondly, the variation of the disparities with depth, which is expected to be nonlinear. Before we go any further with this idea, however, some important aspects need to be studied in order to endorse the assumptions made so far. This will be of help later in formulating our algorithm.

### Disparity vs. depth

*P*(a point in the space) from its projections

*p*and

*p′*, we need to consider similar triangles (Δ

*PO*

_{R}

*O*

_{T}and Δ

*Ppp′*):

*x*

_{R}

*− x*

_{T}is the disparity (

*d*),

*Z*is the depth of

*P*, and

*B*represents the distance between the two cameras. The fixate location (

*f*) is known as the distance in which the planes of projections are fixed.

*O*

_{R}or

*O*

_{T}) by a depth sensor in Figure 3, few considerations are only needed: The range map is to be regarded as a regular image within which disparities with its colored peer may be encountered. Also,

*Z*turns into a known variable accessible from the range map itself. This being so, (1) still holds when a color camera is replaced and should describe the relation ‘disparity (

*d*) vs. depth (

*Z*)’ as a nonlinear hyperbolic function. Finally, Figure 4a,b depicts the process for manually assessing of disparity over a point within a depth map and its peer in a color image given a specific depth distance.

## A new approach

- 1.
To sample as many planes as possible within the range of depth, several objects are placed at different distances in front of the cameras.

- 2.
To capture nearly the same scene with the two cameras (color and range camera), two images (

*I*_{c}and*I*_{d}) are taken as synchronized as possible (Figure 5). - 3.
Sufficient landmarks are selected in

*I*_{c}along with their peers in*I*_{d}. For each landmark, threefold information is assessed: - a)
The

*x*and*y*coordinates of the landmark in*I*_{c}, namely (*x*_{c},*y*_{c}) - b)
The

*x*and*y*coordinates of the landmark in*I*_{d}, namely (*x*_{d},*y*_{d}) - c)
The depth of the landmark, namely

*D*.

*D*is accessible likewise from

*I*

_{d}and pinpoints the distance plane on which the landmark was observed. Thus, ∆ = (

*x*

_{d},

*y*

_{d})

*−*(

*x*

_{c},

*y*

_{c}) is but an example of the shifting of the images at distance

*D*and not elsewhere

*.*In general, each landmark provides evidence of the offset of the images at a specific distance

*.*In practice, taking as many distinct landmarks as possible for a given distance

*D*is advisable at all (as many distances as possible). As noted in the previous section, the shifting ∆ behaves linearly at

*D*(i.e., disparity varies linearly into parallel planes).

- 4.Now the landmarks are used as a set of samples on which a global shifting function (
**∆**) can be interpolated. Eventually, this function can be regarded as a 2D vector flow describing the offset of the images. Hence, one function per coordinate is finally estimated, and ∆ can be reformulated as follows:$\Delta =\left(\mathit{\Delta x}\left({x}_{d},D\right),\mathit{\Delta y}\left({y}_{d},D\right)\right)$(2)

The resulting function **∆** is now vector-valued: it maps each point (*x*
_{d}, *y*
_{d}) in *I*
_{d} to its shifted homolog (*x*
_{c}, *y*
_{c}) in *I*
_{c} so that *x*
_{d} + Δ*x* = *x*
_{c} and *y*
_{d} + Δ*y* = *y*
_{c} for any given *D*. Yet, only few samples of this function are still known. Next section deals with the estimation of the model that best fits these samples. Also, this model will let us interpolate the function in its entirety.

### The shifting basis function ∆

*n*data points (landmarks) do not show a linear distribution; therefore, we must use nonlinear fitting models to approximate ∆. Also, image distortion makes apparent the need of nonlinear models for ∆. This is because cameras suffering from distortion are known to wrap the image with nonlinear aspect [3]. Although in this paper we use an adaptable class of splines [25], there is no constraint in this regard. The thin-plate smoothing spline

*f*used in this work to approximate

*∆μ*(

*μ*= {x | y}) given a set of

*n*data points or landmarks $\left({x}_{\mathrm{d}}^{j},\phantom{\rule{0.5em}{0ex}}{y}_{\mathrm{d}}^{j},{D}^{j}\right)\phantom{\rule{0.5em}{0ex}}\cup \phantom{\rule{0.5em}{0ex}}\left({x}_{\mathrm{c}}^{j},\phantom{\rule{0.5em}{0ex}}{y}_{\mathrm{c}}^{j},{D}^{j}\right)$

*,*∀

*j*∈ {1, … ,

*n*} can be regarded as a unique minimizer of the weighted sum:

with $E\left(f\right)={\displaystyle {\sum}_{j}}\phantom{\rule{0.25em}{0ex}}{\left|\Delta {\mu}^{j}\phantom{\rule{0.2em}{0ex}}-\phantom{\rule{0.3em}{0ex}}f\left({x}_{\mathrm{d}}^{j},\phantom{\rule{0.5em}{0ex}}{y}_{\mathrm{d}}^{j},{D}^{j}\right)\right|}^{2}\phantom{\rule{0.75em}{0ex}}$ as the error measure, and *R*(*f*) = ∫(|∂_{1}∂_{1}
*f*|^{2} + |∂_{2} ∂_{2}
*f*|^{2}) the roughness measure. Here, the integral is taken over all of **R**^{2}, |*z*|^{2} denotes the sum of squares of all the entries of *z*, and ∂_{
i
}
*f* denotes the partial derivative of *f* with respect to its *i* th argument. The smoothing parameter κ in (3) is derived from preprocessing of the set of data.

*f*be the shifting function [also known as ∆ in (2)] so that

*f*maps (

*x*

_{d},

*y*

_{d}) → (

*x*

_{c},

*y*

_{c}) for a given

*D*. The general equation for

*f*is given as follows:

*, f*may be also expressed in terms of its components as

*x*and

*y*coordinates is described independently [i.e.,

*f*

_{ x }

*=*∆

*x*and

*f*

_{ x }

*=*∆

*y*in (2)]. In Equation 4,

*n*is the number of samples (landmarks) we shall use to interpolate

*f*. Whereas

*a*

_{1},

*a*

_{ x, }

*a*

_{ y },

*a*

_{ D, }and all

*w*

_{ i }are the unknown coefficients we need to calculate. As for

*U*, this is a special function underlying the thin-spline [25] defined as

*U*(

*x,y*)

*= U*(

*r*)

*= r*

^{ 2 }log(

*r*

^{ 2 }), with

*r*being the distance $\sqrt{{x}^{2}+{y}^{2}}$ from the Cartesian origin. Now, for the calculation of the unknown coefficients in Equation 4, we need to consider ${r}_{j,i}=\left|\left({x}_{d}^{i},\phantom{\rule{0.5em}{0ex}}{y}_{d}^{i},{D}^{i}\right)-\left({x}_{d}^{j},\phantom{\rule{0.5em}{0ex}}{y}_{d}^{j},{D}^{j}\right)\right|$, ∀

*j*, and ∀

*i*∈ {1, … ,

*n*}. Therefore,

^{T}is the matrix transpose operator and

**0**is a 4 × 4 matrix of zeros. Then let

**Y**= (

**V**| 0000)

^{T}be a vector of length

*n*+ 4. Finally, define

*W =*(

*w*

_{1},

*w*

_{2},…,

*w*

_{ n }) and the coefficients

*a*

_{1},

*a*

_{ x },

*a*

_{ y },

*a*

_{ D }by the equation:

the solution of *L*^{−1}
*Y* gives all the necessary information to construct *f*. Note that this is a matrix of size of 2 × (*n* + 4), with each row providing *w*
_{
i
}, *a*
_{1}, *a*
_{
x
}, *a*
_{
y,
} and *a*
_{
D
} for both, *f*
_{
x
} and *f*
_{
y
}.

### Algorithmic performance, experiments, and comparisons

In this section, our algorithm for color-range calibration is outlined. Further, its computational performance is assessed too. Eventual concerns regarding nonlinearity along with efficient solutions are introduced and treated here in Subsections ‘Practical test’ and ‘Efficient use of splines’. Finally, comparisons with related methods are conducted in this section. It is worth noticing though that our method is proposed as a general framework to couple any depth-color camera pair. We have limited the comparisons in the Section ‘Evaluation of the method’ to a specific case where our algorithm may be specifically applied as well i.e., internal Kinect calibration. The approaches whose efficiency is compared to that of our method in Subsection ‘Evaluation of the method’ are threefold:

A-1. Calibration of Kinect (mapping of depth data onto the RGB images) using typical checkerboard-based stereo calibration [4, 26] i.e., assuming the range camera as digital

A-2. Calibration of Kinect using the drivers provided by manufacturer (PrimeSense, Tel Aviv, Israel)

A-3. Herrera's method [20] that uses a new depth distortion model to calibrate depth and color sensors

### Algorithm

- i.
- ii.
Construct matrix

**L**, using**K**,**P**and**P**^{T}as described in Equation 7. - iii.
Let (

**V**| 0000)^{T}be the vector termed Y. - iv.
Solve

**L**^{−1}**Y**= (*W*|*a*_{1}*a*_{ x }*a*_{ y }*a*_{ D })^{T}to obtain*a*_{1},*a*_{ x },*a*_{ y },*a*_{ D, }and all*w*_{ i }in Equation 4. - v.
Get

*I*_{c}and*I*_{d}from corresponding sensors. - vi.
Find ∆

*x*for each (*x*_{d},*y*_{d},*D*) solving*f*_{ x }described in Equation 4. - vii.
Find ∆

*y*for each (*x*_{d},*y*_{d},*D*) solving*f*_{ y }described in Equation 4. - viii.
Shift all (

*x*_{d},*y*_{d}) towards (*x*_{c},*y*_{c})*.*

Note that steps (i) to (iv) are related to the calculation of the spline model and are performed offline only once. Furthermore, if the system is ever decoupled, no recalculation of these steps is needed when recoupling. One can do the readjustment of the cameras by hand until acceptable matching of the images is attained. Also, we can see that the calculation of these offline steps is not computationally heavy. Typically, **L**^{−1}
**Y** = (*W* | *a*
_{1}
*a*
_{
x
}
*a*
_{
y
})^{T} is solved as a system of linear equations of the kind of **A*****x** = **B**. The computation of this linear system requires around *O*(*d*^{
3
} *+ d*^{
2
}
*n*) computations (where *n* is the number vectors or landmarks and *d* indicates their dimension). Theoretically, in our method only 3 three-dimensional landmarks are needed (three points are enough to calculate a plane). In practice, however, the typical number of landmarks is approximately 20.

On the other hand, steps (v) to (viii) make up the workflow to be performed online. Particularly, we are concerned with steps (vi) to (viii) which are actually in the core of our computational approach. Having *a*
_{1,}
*a*
_{
x
}, *a*
_{
y
}, *a*
_{
D,
} and *w*
_{
i
} as constant data resulting from the offline phase, the solving of spline equation [steps (vi) and (vii)] requires moderate number of elemental operations [27]. This is perhaps the densest part of our algorithm. In next section, however, we will see how this part in this paper is further lowered in computational terms (*efficient use of splines*). Finally, step (viii), in turn, is but a constant array assignation. Overall, the complexity of our online algorithm is linear with the size of the images *N* [i.e., *O*(*N*)]. For images as *I*
_{c} and *I*
_{d} that usually do not exceed the order of megabytes [28], the complexity is noticeably low.

### Practical test

*x*and ∆

*y*were estimated for each scene. Each scene’s landmarks (Figure 5) provided, then, two data sets to be interpolated. A total of twelve data sets were collected and twelve interpolations performed in this test. Nonlinear fitting models (thin-plate splines) were used; also all these models were cross-validated

*.*Figure 7 shows the mean error for these interpolations.

Figure 7 reveals that in the general case [as expected from (1)], nonlinear fitting models are suitable enough to interpolate the ∆*x* and ∆*y* functions. In average, an error of just 2.3 pixels affected these interpolations. This error was totally expected as a consequence of the manual markup to which this method is subjected. In the first four cases (RGB-Kinect), the interpolating splines presented small roughness parameter (κ ≈ 0, Equation 3). In other words, the interpolating surfaces were nearly flat (planes). In these cases and of course, depending on the needs of accuracy, we observed that ∆*x* and ∆*y* could be acceptably fitted by planes. This has to do with the range of depth in which the cameras are intended to be registered. When it comes to short ranges, it was a commonplace observation in our workflow that (1) could be tolerably approximated by linear surfaces (Appendix). Kinect, for instance, has a limited range rather small in comparison to other depth sensors, likewise, whether or not the images are undistorted matters as well. It is well known that Kinect also meets this feature since its images are internally preprocessed. This, of course, was not the case in the last eight interpolations were κ was rather large, making the nonlinearity apparent.

### Efficient use of splines

As observed in (1), planar regressions would fail to keep accurate fitting of the functions ∆*x* and ∆*y*. Also according to Figure 7, splines happen to be a very precise method to model the nonlinearity inherent to the problem. In general, ToF cameras also suffer from distortions both on the measured depth and on the ray direction [3]. This makes even more apparent the need of nonlinear models in this problem. Although Kinect is not an exception, this sensor is calibrated during manufacturing. The calibration parameters come internally stored and are used by the official drivers. This might partially explain why linear models could be moderately acceptable in this case. Yet in the general case (including Kinect), the use of splines yields optimal results, which is otherwise unachievable. Unfortunately, the use of nonlinear spline models [steps (vi) and (vii) of Subsection ‘Algorithm’] could raise concerns regarding computational performance. This fact makes a lineal model more desirable for the proposed methods. Nevertheless, we will see that the use of a nonlinear model does not affect at all the actual efficiency of the proposed method.

*f*

_{ x }and

*f*

_{ y }described in Equation 4 to find ∆

*x*and ∆

*y*. In lieu of this, we pick values from the surfaces saved in memory. Thus, the online process becomes independent of the mathematical model (either linear or nonlinear), as we simply access the memory to read intended values. Eventually, steps (vi) and (vii) are lowered elementally, which enhance even more the performance of our method. Figure 8 roughly summarizes this idea.

### Evaluation of the method

Using Kinect, three patterns which the edges are known to be lines are imaged from multiple views. A set of 20 pairs (depth and color) of raw images is gathered in the end. A manual segmented version of all the color images serves as ground truth. Three areas are segmented from each image i.e., the three patterns. Moreover, we register the 20 pairs of images by shifting the depth maps using A-1, A-2, A-3, and, the method being described in this paper. These shifted maps are automatically segmented in three areas as well. Finally, we compare the common areas between these segmented maps (sm) and those of the ground truth (gt). Common areas must overlap exactly each other under the assumption of perfect registration. Thus, for each pair of overlapped areas $\left({a}_{i}^{\mathrm{gt}},{a}_{i}^{\mathrm{sm}}\right)$ we assess its intersection $\left({a}_{i}^{\mathrm{gt}}\cap {a}_{i}^{\mathrm{sm}}\right)$, ∀*i* ∈ {1, 2, 3}.

*y*$\frac{1}{6}{\displaystyle \sum}_{i=1}^{4}{{\displaystyle}}^{\mathrm{c}}\phantom{\rule{0.25em}{0ex}}\left|\frac{{\mathit{a}}_{\mathit{i}}^{\mathrm{gt}}{\displaystyle \cap}{\mathit{a}}_{\mathit{i}}^{\mathrm{sm}}}{{\mathit{a}}_{\mathit{i}}^{\mathrm{sm}}}+\frac{{\mathit{a}}_{\mathit{i}}^{\mathrm{gt}}{\displaystyle \cap}{\mathit{a}}_{\mathit{i}}^{\mathrm{sm}}}{{\mathit{a}}_{\mathit{i}}^{\mathrm{gt}}}\right|$. Notice that Acc is expected to be 1 for images successfully registered and below in other cases. We also measure the time elapsed during the registration of two images using the four methods. Figure 9 and Table 1 summarize the results of this section.

**Computational performance**

Our method | A1 | A2 | A3 | |
---|---|---|---|---|

Time (s) | 0.021 | 0.036 | 0.027 | 0.033 |

Potential fps | 46 | 27 | 37 | 30 |

*Acc*equal to 1 was not expected for any of the methods. This is mostly the case because segmented areas in range images are known to present highly noisy edges (Figure 10, bottom row). Thus, flawless intersection with the areas in the ground truth (Figure 10, middle row) is unlikely. As a consequence, the accuracy of the manufacturer (A2) can be regarded as a baseline. By showing no substantial difference with this baseline, our method roughly reaches the maximum expectation of accuracy in this experiment. Moreover, having a standard deviation slightly smaller, results obtained with A2 may be regarded as a more consistent. Nonetheless, this very fact allows our method to achieve better accuracy than A2 in some cases (not outliers). As for A3, much better accuracy than A1 was noticeably reached, although the method certainly failed to surpass the threshold of 90% accuracy. This leads our method to a slightly better performance with nearly 92%. It is worth noticing that A1, A2, and A3 are methods that require extrinsic and intrinsic parameters of both cameras. Hence, use of extensive calibration techniques with checkboards and heavily manual work is unavoidable. The efficiency of our method suppresses these procedures, as well as maintains an average accuracy for otherwise unreachable.

Finally, A2 is inextensible to the general problem of color and range images registration. This calibration is conducted during manufacturing and internally stored into the official drivers. Therefore, coupling the Kinect range sensor with an external color camera using A2 turns out to be of no use. On the other hand, A1 method does apply to the general problem. There is no apparent reason, however, to expect better accuracy by varying either of the cameras. The problem here relies on the treatment of noisy range images (with not even visual features) as highly defined color images. With regard to Table 1, it is worth stressing that both our method and A1 were implemented in Matlab, whereas, method A2 is an internal routine of the Kinect driver written is C++. Therefore, drastically better performance is expected for a binary compiled version of our algorithm.

## Conclusions

In this work, a method for registration of color and range images was presented. We showed that the proposed method is practical, accurate, and widely applicable. Firstly, the problem of registering two close-positioned cameras and the underlying nonlinearity was studied in this paper. Following, an itemized outline of our method was given as well as a mathematical description of the nonlinear basis function. Furthermore, we presented a computational approach that preserves efficiency and accuracy regardless of the underlying mathematical model. This leads our method to be useful in correcting range camera distortion, a problem that remains challenging.

Overall, our method may be suitable in several applications: be it for Kinect color improvement, coupling of range and color cameras, or the calibration of the depth-color camera pair of Kinect. As for the latter, we presented an evaluation of our method that revealed as much accuracy as that of the manufacturer when matching the internal depth-color image pairs. Besides, our computational performance was fairly better compared to those of manufacture and other method aiming the same goal. However, our method is not yet fully automatic due to user markup required in the initial calculation of the shifting function. This is perhaps the major limitation of the work here presented. Therefore, future works will be focused on this line of research. We would like to explore automatic markup based on saliency points or corner detection. The problem to be tackled is that automatic algorithms will fail to detect, for instance, corners in range images due to boundaries' discontinuities caused by physical issues of range measurement hardware.

Unlike the others, the approach here proposed does not require intrinsic calibration of neither of the cameras. Also, the simplicity of its implementation leading to fairly important performance gains relies on the novel approach based on splines here presented. This method was motivated by the fact that cutting-edge depth-color cameras present low-quality RGB images. Even worse, some ToF sensors (e.g., CamCube) provide not even an intensity image, which dramatically diminishes their scope for expansion into computer vision. Additionally, this camera, among others, presents distortion as significant as to make calibration a must in the workflow. Typical checkboard-based calibration, however, is of no use when not even gray-level information is provided. Eventually, this problem may be solved if after registering the color and the depth images, the distortion is corrected in the former and then extrapolated to the latter.

## Appendix

In this paper, we compared our spline-based approach with planar fitting models (linear approach). Although the general case is ruled by (1), in exceptional cases, planes yield acceptable interpolations (Section ‘Practical test’). Therefore, in this section we will show how those planes were built to interpolate both, ∆*x* and ∆*y*. In this view, the problem can reduce to a linear regression in a three-dimensional space.

*μ*be either of the variables

*x*,

*y*so that

*∆μ*denotes either of the functions ∆

*x*, ∆

*y*. Given a set χ of

*n*data points (landmarks) (

*μ*

_{d}

^{(1)},

*D*

^{(1)},

*∆μ*

^{(1)}), (

*μ*

_{d}

^{(2)},

*D*

^{(2)},

*∆μ*

^{(2)}),…,(

*μ*

_{d}

^{(n)},

*D*

^{(n)},

*∆μ*

^{(n)}). We want to find the equation of the plane that best fits our set. A vector version of the equation of this plane can be formulated as follows:

where *a* is a normal vector of the plane and *b* is a vector that results from the product of *a* and the mean of the set of data points (i.e. $b={a}^{\wedge}\frac{1}{n}{\displaystyle \sum}_{i=1}^{n}\left({\mu}_{\mathrm{d}}^{i},{D}^{i},\phantom{\rule{0.5em}{0ex}}{\Delta}_{\mu}^{i}\right)$. Therefore, *a* happens to be the only variable unknown.

Principal components analysis can be used to calculate a linear regression that minimizes the perpendicular distances from the data to the fitted model [29]. In other words, given the three data vectors *μ*
_{d}, *D*, and *∆μ*, one can fit a plane that minimizes the perpendicular distances from each of the points (*μ*
_{d}
^{(i)}, *D*^{(i)}, *∆μ*^{(i)}) to the plane, ∀*i* ∈ {1, … , *n*}. In short, the first two principal components of χ define the plane; the third is orthogonal to them and defines the normal vector of the plane [30], namely *a*.

## Declarations

## Authors’ Affiliations

## References

- Huhle B, Jenke P, Strasser W: On-the-fly scene acquisition with a handy multi-sensor system.
*Int. J. Intelligent Syst. Tech. Appl.*2008, 5(3):255-263.Google Scholar - Scaramuzza D, Harati A, Siegwart R: Extrinsic self calibration of a camera and a 3D laser range finder from natural scenes. In
*Proceedings of the International Conference on Intelligent Robots and Systems (IROS), 2007*. San Diego, California; 2007:4164-4169.View ArticleGoogle Scholar - Herrera D, Kannala J, Heikkila J: Joint depth and color camera calibration with distortion correction.
*IEEE Trans Pattern Anal Mach Intell*2012, 34(10):2058-2064.View ArticleGoogle Scholar - Van den Bergh M, Van Gool L: Combining RGB and ToF Cameras for Real-time 3D Hand Gesture Interaction. In
*IEEE Workshop on Applications of Computer Vision (WACV)*. Kona, Hawaii; 2011:66-72.Google Scholar - Fuchs S, Hirzinger G: Extrinsic and depth calibration of ToF cameras. In
*Proceedings of the International Conference on Computer Vision and Pattern Recognition (CVPR)*. Anchorage, Alaska; 2008:1-6.Google Scholar - Zhu J, Wang L, Yang R, Davis J: Fusion of time-of-flight depth and stereo for high accuracy depth maps. In
*Proceedings of the International Conference on Computer Vision and Pattern Recognition (CVPR)*. Anchorage, Alaska; 2008:11-18.Google Scholar - Zitová B, Flusser J: Image, Registration methods: a survey.
*Elsevier Image and Vision Computing*2003, 21(11):977-1000.View ArticleGoogle Scholar - Deshmukh M, Bhosle U: A survey of image registration.
*Int. J. Image Process. (IJIP)*1992, 5(3):245-269.Google Scholar - Yang SH: Neural network based stereo matching algorithm utilizing vertical disparity. In
*Proceedings of the Annual Conference on IEEE Industrial Electronics Society (IECON)*. Glendale, AZ; 2010:1155-1160.Google Scholar - Sun J, Zheng N, Shum H: Stereo matching using belief propagation.
*IEEE Trans Pattern Anal Mach Intell*2002, 25(7):787-800.Google Scholar - Han B, Paulson C, Wang J, Wu D: Depth-based image registration. In
*Proceedings of the Algorithms for Synthetic Aperture Radar Imagery*. Honolulu; 2010:542-562.Google Scholar - Staranowicz A, Morbidi F, Mariottini G-L: Depth-camera calibration toolbox (dcct): accurate, robust, and practical calibration of depth cameras. In
*Proceedings of the British Machine Vision Conference (BMVC)*. Leeds; 2012.Google Scholar - Tilak D: Evaluation of the Kinect sensor for 3-D kinematic measurement in the workplace.
*Appl Ergon*2011, 43(4):645-649.MathSciNetGoogle Scholar - Andersen M, Jensen T, Lisouski P, Mortensen A, Hansen M, Gregersen T, Ahrendt P:
*Kinect depth sensor evaluation for computer vision applications*. Technical Report (Aarhus University, Department of Engineering); 2012.Google Scholar - Biber P, Straßer W: The normal distributions transform: a new approach to laser scan matching. In
*Proceedings of the International Conference on Intelligent Robots and Systems (IROS)*. Las Vegas; 2003.Google Scholar - Lowe DG: Distinctive image features from scale-invariant keypoints.
*J Comput Vision*2004, 60(2):91-110.View ArticleGoogle Scholar - Golub H, Van L:
*Matrix Computations*. 3rd edition. Baltimore: John Hopkins University; 1996.Google Scholar - Staranowicz A, Mariottini G-L: A comparative study of calibration methods for Kinect-style cameras. In
*Proceedings of the International Conference on Pervasive Technologies Related to Assistive Environments (PETRA)*. Crete, Greece; 2012.Google Scholar - Zhang C, Zhang Z: Calibration between depth and color sensors for commodity depth cameras. In
*Proceedings of the Workshop on Hot Topics in 3D (ICME)*. Barcelona, Spain; 2011.Google Scholar - Herrera D, Kannala J, Heikkila J: Accurate and practical calibration of a depth and color camera pair. In
*Proceedings of the 14th International Conference on Computer Analytical Images (CAIP)*. Sevilla; 2011.Google Scholar - Crispell D, Mundy J, Taubin G: Parallax-Free Registration of Aerial Video. In
*Proceedings of the British Machine Vision Conference (BMVC)*. Leeds; 2008:245-250.Google Scholar - Han B, Paulson C, Wu D: Depth based image registration via 3D geometric segmentation.
*J. Visual Communication and Image Representation (JVCIR)*2012, 22(5):421-431.MathSciNetView ArticleGoogle Scholar - Marr D, Poggio T: A computational theory of human stereo vision.
*Proc R Soc London, Ser B*1979, 204: 301-328.View ArticleGoogle Scholar - Gomez J, Bologna G, Pun T: A virtual ceiling mounted depth-camera using orthographic Kinect. In
*Proceedings of the International Conference on Computer Vision (ICCV)*. Barcelona; 2011:50-55.Google Scholar - Bookstein FL: Principal warps: thin-plate splines and the decomposition of deformations.
*IEEE Trans Pattern Anal Mach Intell*1989, 11(6):567-585.View ArticleGoogle Scholar - Bouguet J-Y:
*Matlab Camera Calibration Toolbox*. 2010. . Accessed 2 March 2012 http://www.vision.caltech.edu/bouguetj/calib_doc/ Google Scholar - Guo Z: Reduced complexity Schnorr-Euchner decoding algorithms for MIMO systems.
*IEEE Communication Lett.*2004, 8(5):286-288.View ArticleGoogle Scholar - Anderson J, Mohan S: Sequential coding algorithms: a survey and cost analysis.
*IEEE Trans Commun*1994, 32(6):169-176.Google Scholar - Moore B: Principal component analysis in linear systems: controllability, observability, and model reduction.
*IEEE Trans Autom Control*1981, 26(1):17-32.View ArticleGoogle Scholar - Oja E: Neural networks, principal components, and subspaces.
*Int J Neural Syst*1989, 1(1):48-71.View ArticleGoogle Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.