Building a highly accurate robot is not trivial yet we can perform fine positioning tasks like threading a needle using hand-eye coordination. For a robot we call this visual servoing.
Search Results for: computer vision
Let’s look at some recent research results that vividly show how information from many 2D images taken from many different locations can be combined to form a detailed 3D model of the world.
Vision is useful to us and to almost all forms of life on the planet, perhaps robots could do more if they could also see. Robots could mimic human stereo vision or use cameras with superhuman capability such as wide angle or panoramic views.
Given two images of a scene taken from slightly different viewpoints, a stereo image pair, it’s possible to determine the disparity for every pixel using template matching. The disparity image is one where the value of each pixel is inversely related to the distance between that point in the scene and the camera.
Let’s recap the important points from the topics we have covered about human depth perception, display of 3D images and estimating 3D scene structure using stereo and other types of sensors.
We will compare and contrast the terms image processing, computer vision and robotic vision — they have much in common but there are some subtle but important distinctions. When it comes to interpreting an image we typically try to find and describe regions, lines and interest points.
An important problem in robotic vision is moving a camera so that the view it sees matches the view we want it to have. To achieve this we exploit knowledge about how an image changes as a camera moves. Then we invert that and compute how the camera should move so the image changes in […]
Visual servoing is concerned with the motion of points in the world. How can we reliably detect such points using computer vision techniques.