An important problem in robotic vision is moving a camera so that the view it sees matches the view we want it to have. To achieve this we exploit knowledge about how an image changes as a camera moves. Then we invert that and compute how the camera should move so the image changes in […]
Search Results for: visual odometry
The relationship between world coordinates, image coordinates and camera spatial velocity is elegantly summed up by a single matrix equation that involves what we call the image Jacobian.
Let’s recap the important points from the topics we have covered about human depth perception, display of 3D images and estimating 3D scene structure using stereo and other types of sensors.
The human eye is quite amazing, let’s look at its various components including the light sensitive rod and cone cells.
The relationship between world coordinates, image coordinates and camera spatial velocity has some interesting ramifications. Some very different camera motions cause identical motion of points in the image, and some camera motions leads to no change in the image at all in some parts of the image. Let’s explore at these phenomena and how we […]
We summarise the important points from this lecture.
We learn how to describe the 2D pose of an object by a 3×3 homogeneous transformation matrix which has a special structure. Try your hand at some online MATLAB problems. You’ll need to watch all the 2D “Spatial Maths” lessons to complete the problem set.
A problem arises when using three-angle sequences and particular values of the middle angle leads to a condition called a singularity. This mathematical phenomena is related to a problem that occurs in the physical world with mechanical gimbal systems. Note that in Robotics, Vision & Control (second edition) and RTB10.x the default definition of roll-pitch-yaw […]
We learn a method for succinctly describing the structure of a serial-link manipulator in terms of its Denavit-Hartenberg parameters, a widely used notation in robotics.