You are on page 1of 41

Alignment problem

In alignment, we will fit the parameters of some transformation


according to a set of matching feature pairs (“correspondences”).

xi
xi '

1
Kristen Grauman
Basic Set of 2D planar Transformation
Table 6.1 Jacobians of the 2D coordinate transformations x ′ = f(x; p) shown in Table 2.1, where we
have re-parameterized the motions so that they are identity for p = 0.
2D Alignment using Least Square
2D Alignment using Least Square
2D Alignment using Least Square
2D Alignment using Least Square
2D Alignment using Least Square
A simple linear regression (linear least squares problem) can be formulated as
The minimum can be found by solving the symmetric positive definite (SPD) system of normal equations
• Let tj be the location of the jth image coordinate frame in the global
composite frame and xij be the location of the ith matched feature in
the jth image. In order to align the images, we wish to minimize the
least squares error
Applications : panography
Iterative algorithms
Iterative algorithms
New Parametric Equations
The Jacobian is
Robust least squares and RANSAC
3D Alignment
Pose Estimation
• Pose estimation is a computer vision technique that predicts and tracks
the location of a person or object. This is done by looking at a
combination of the pose and the orientation of a given person/object
• pose estimation as the problem of determining the position and
orientation of a camera relative to a given person or object.
• his is typically done by identifying, locating, and tracking a number
of keypoints on a given object or person. For objects, this could be
corners or other significant features. And for humans, these keypoints
represent major joints like an elbow or knee
Tracking keypoints on a person playing ping
pong.
Pose Estimation
• The problem of recovering pose from three correspondences, which
is the minimal amount of information necessary, is known as the
perspective-3-point-problem (P3P), with extensions to larger numbers
of points collectively known as PnP.
• Linear algorithms
• Iterative algorithms
Applications
• In addition to tracking human movement and activity, pose estimation
opens up applications in a range of areas, such as:
• Augmented reality
• Animation
• Gaming
• Robotics
Pose Estimation- Linear algorithms
Camera Calibration
• Geometric camera calibration, also referred to as camera
resectioning, estimates the parameters of a lens and image
sensor of an image or video camera.
• You can use these parameters to correct for lens distortion,
measure the size of an object in world units, or determine the
location of the camera in the scene.
These tasks are used in applications such as machine vision to
detect and measure objects. They are also used in robotics, for
navigation systems, and 3-D scene reconstruction.
• Camera parameters include intrinsics, extrinsics, and distortion
coefficients.
• To estimate the camera parameters, you need to have 3-D
world points and their corresponding 2-D image points.
• You can get these correspondences using multiple images of a
calibration pattern, such as a checkerboard.
• Using the correspondences, you can solve for the camera
parameters. After you calibrate a camera, to evaluate the
accuracy of the estimated parameters. 
The pinhole camera parameters are represented in a 4-by-3 matrix called the camera matrix.

This matrix maps the 3-D world scene into the image plane.

The calibration algorithm calculates the camera matrix using the extrinsic and intrinsic parameters.

The extrinsic parameters represent the location of the camera in the 3-D scene.

The intrinsic parameters represent the optical center and focal length of the camera.

        
Camera Calibration Parameters
The calibration algorithm calculates the camera matrix using the extrinsic
and intrinsic parameters.

The extrinsic parameters represent a rigid transformation from 3-D world


coordinate system to the 3-D camera’s coordinate system.

The intrinsic parameters represent a projective transformation from the 3-D


camera’s coordinates into the 2-D image coordinates.

            
The world points are transformed to camera coordinates using the extrinsics
parameters. The camera coordinates are mapped into the image plane using the
intrinsics parameters.
Extrinsic Parameters
The extrinsic parameters consist of a rotation, R, and a translation, t. The origin of the camera’s
coordinate system is at its optical center and its x- and y-axis define the image plane.

            
Intrinsic Parameters
The intrinsic parameters include the focal length, the optical
center, also known as the principal point, and the skew
coefficient. The camera intrinsic matrix, K, is defined as:
The pixel skew is defined as:

     
[cx, cy] — Optical center (the principal point), in pixels.

(fx,fy) — Focal length in pixels.


fx=F/px
fy=F/py
F — Focal length in world units, typically expressed in millimeters.
(px,py) — Size of the pixel in world units.

s — Skew coefficient, which is non-zero if the image axes are not perpendicular.
s=fx tanα

You might also like