Image Processing Reference
In-Depth Information
accurate version of this model by assuming a different affine transform, and thus a different d , for
each facet of the mesh. This approximation neglects depth variation across individual facets, rather
than across the whole surface. Under this assumption, the projection of a 3D point q i lying on facet
f of the mesh can be expressed as
d f u i
v i
A I 2 × 2
0 q i +
0 ,
=
(3.2)
where d f accounts for the average depth of facet f . Here, without loss of generality, we expressed
the 3D point in the camera referential, and therefore replaced R with the first two rows of the
3
3 identity matrix and the translation with a zero vector. Note that this does not prevent us from
accounting for camera motion. It simply means that it will be interpreted as a rigid motion of the
object of interest.
Under the full perspective model, the projection of a 3D point q i
×
is written as
u i
v i
1
A I 3 × 3 q i +
0 ,
=
d i
(3.3)
where the matrix of internal camera parameters A is now a 3
×
3 matrix, and each point i has a
different depth factor d i .
3.2
3D-TO-2DCORRESPONDENCES
Detecting feature points in images has received enormous attention in the Computer Vision commu-
nity. For most template-based approaches, feature points are typically detected with either the SIFT
keypoints detector Lowe [ 2004 ] or Harris's corners detector Harris and Stephens [ 1988 ]. Once fea-
ture points have been detected in two images, they need to be matched to produce correspondences.
When using SIFT, this can be done by a simple dot-product between specific vector representa-
tions of the feature points. For Harris's corners, methods based on Randomized Trees have proved
efficient Lepetit and Fua [ 2006 ]. From a large set of views obtained by applying random affine trans-
formations to a reference image, a tree that models the relationships between neighboring keypoints
is built. Each leaf-node of the tree then corresponds to a specific keypoint, and matching can be
done by dropping the feature points of a new image down the tree.
Another way to establish correspondences is to first tackle the 2D non-rigid image regis-
tration problem. Non-rigid image registration aims at finding a transformation between two im-
ages of the same surface undergoing different deformations. Different parameterizations have been
proposed to represent the transformation, such as RBFs Bartoli and Zisserman [ 2004 ], thin-plate
splines Bookstein [ 1989 ], or 2D meshes Pilet et al. [ 2008 ]. Since the resulting warp is defined over
the entire image, discrete correespondences can then be obtained by sampling it. Note that, while
2D non-rigid registration can be thought of as related to 3D non-rigid reconstruction, we believe
 
Search WWH ::




Custom Search