Home/ Technology/ Sensing & Perception

How Does a Robot
Actually See?

A camera is the easy part. The hard part is everything that happens next — turning a flood of raw light data into something a machine can actually use to understand where it is, what surrounds it, and what to do about it.

Robot vision sensor array
A Boston Dynamics Spot robot navigating an unstructured environment. The sensors visible on its body are not cameras in any conventional sense — each one is solving a different piece of the perception problem simultaneously.
The Problem

Cameras have existed since the 1800s. Engineers have been trying to make machines understand what a camera sees for over sixty years. The gap between those two facts tells you something about how hard the problem is.

Capturing light is not the challenge. The challenge is understanding what the light means. Turning a stream of raw image data into a useful picture of the world, one that tells a robot where it is, what is around it, and what those things are, is a problem that resisted solution for most of computing history. It is only now becoming tractable.

The camera is not the hard part. Understanding what the camera sees is the entire problem.

Why a Camera Alone Is Not Enough

A standard camera captures light and turns it into an image. That image is flat. It is a two-dimensional record of what was in front of the lens at a given moment. It has color, texture, and contrast, but it has no depth. It cannot tell you how far away anything is.

For a human looking at a photograph, depth is not a problem. Our brains fill it in from visual cues, the way objects overlap, how they get smaller with distance, shadows and perspective. We have spent a lifetime learning to read these cues without thinking about it.

A robot has no such experience to draw on. And for a machine that needs to navigate physical space, to walk down a corridor, pick an object off a table, or avoid hitting a person, knowing how far away things are is not optional. It is the most important piece of information there is.

This is why modern robotic vision systems use multiple sensors. Each one solves a different part of the problem. Together they build a picture that none of them could produce alone.

The Sensors

LiDAR

LiDAR fires pulses of laser light, thousands of them every second, in all directions around the sensor. Each pulse travels outward until it hits a surface and bounces back. The sensor measures how long that return journey takes. Because light always travels at the same speed, that time measurement translates directly into a distance.

Do this hundreds of thousands of times per second and you build a three-dimensional map of everything around the sensor. A cloud of millions of individual distance measurements, each one a precise point in space. Engineers call this a point cloud.

LiDAR point cloud visualization
A LiDAR point cloud of an indoor environment. Every dot is an individual laser measurement. The robot knows the precise distance to every one of those points, but nothing about what they are.

A good LiDAR sensor can measure distances up to 100 meters with an error of just a centimeter or two. It works in the dark. Shadows and changes in lighting do not affect it, though heavy rain and direct sunlight can introduce errors.

Its limitation is that it cannot tell you what anything is. A point cloud shows shapes and distances but nothing about color, texture, or meaning. To LiDAR, a white wall and a white wardrobe look identical. Both are flat surfaces at a certain distance. Distinguishing between them requires something else.

How it works

Why LiDAR is so precise

Light travels at approximately 300,000 kilometers per second. A LiDAR sensor can measure time intervals of just a few nanoseconds, billionths of a second. At those timescales, even tiny differences in travel time correspond to measurable differences in distance. This is why LiDAR can resolve objects centimeters apart at distances of 50 meters or more.

Cameras

While LiDAR maps the geometry of a space, a camera captures what that geometry looks like. It records color. It captures texture, the difference between a smooth floor and a carpet, between glass and wood, between a person's face and the wall behind them. It sees patterns and markings that LiDAR cannot.

A single camera still cannot measure depth reliably on its own. But two cameras mounted side by side, the way your two eyes are, can. The slight difference between what each one sees lets the robot calculate depth geometrically, the same way your brain does when it uses both eyes to judge distance. This is called stereo vision.

Stereo vision works well up to a few meters. Beyond that, the difference between the two images becomes too small to measure accurately. For longer distances, LiDAR takes over. The two sensors are not competing. They are covering each other's weaknesses.

Depth cameras

A depth camera projects a pattern of infrared dots onto the environment and measures how that pattern is distorted by the surfaces it lands on. From the distortion it calculates distance. The result is an image where every pixel has not just a color value but a depth value.

Depth cameras are cheaper and more compact than LiDAR but less precise and less effective outdoors where sunlight interferes with the infrared signal. They are common in indoor robots. The face recognition on many modern smartphones uses the same technology to map the contours of a face in three dimensions.

From sensing to understanding

Sensor Fusion

A robot running all of these sensors simultaneously is receiving an enormous amount of raw data every second. LiDAR is generating a new point cloud ten times per second. Cameras are producing thirty images per second. None of this is yet useful on its own. It is just numbers.

The process of combining all of it into a single coherent picture is called sensor fusion. When the camera spots a chair and LiDAR confirms a solid object at exactly that location and distance, the robot does not process these as two separate pieces of information. It understands them as one. The chair has a precise location in space and an identity. That combination, where something is and what it is, is what the robot needs to act on.

For fusion to work, every sensor on the robot must be carefully calibrated. Its exact position, angle, and timing synchronized with all the others so that the data lines up correctly in space. A camera pointing two degrees off from where the robot thinks it is pointing produces errors that cascade through everything that follows.

From Data to Understanding

Even with perfectly fused sensor data, the robot still does not understand its environment in any deep sense. Before higher-level thinking can happen, it needs to solve a more basic problem: it needs to know where it is.

This is handled by a technique called SLAM, Simultaneous Localization and Mapping. The robot builds a map of its surroundings from sensor data while simultaneously tracking its own position within that map. It is doing both things at once, each one helping to refine the other. SLAM is covered in depth in a separate article, but it is worth knowing that without it, a robot navigating a corridor has no reliable sense of where it started, where it has been, or where it is now.

Once the robot knows where it is, the next layer can begin. Modern robots use neural networks to identify what their sensors are seeing. The network has been shown millions of labeled images during training: a chair labeled "chair," a door labeled "door," a person labeled "person." After enough examples, it learns to recognize these things in any lighting, at any angle, partially hidden, near or far.

When sensor data flows into the network, it does not just output a label. It outputs a confidence level. "That object is a chair, and I am 94% confident." Run this across an entire scene and the robot gets a complete picture: what is in the room, where everything is, and how certain it is about each one.

Beyond identification, the most capable systems are starting to understand context. A hand reaching toward a cup suggests movement is coming. A stack of boxes is something to navigate around. A flat surface at table height is somewhere to place something. The robot is not just cataloguing what it sees. It is starting to read the situation. That is still limited compared to human perception, but it is a meaningful step beyond pattern recognition.

Why This Is Happening Now

All of these sensors have existed in some form for decades. LiDAR was developed in the 1960s. Stereo vision was being researched in the 1980s. The reason robotic vision has only recently become capable is not the sensors. It is the computing power to process their output, and the machine learning techniques to make sense of it.

The processors now embedded in robots can handle the required calculations in real time. The neural networks running on them have been trained on datasets containing hundreds of millions of labeled images. What once required a room full of expensive computers now runs on compact hardware inside a mobile robot.

That is why robotics demonstrations today look so different from five years ago. The sensors are better, the processing is faster, and the software interpreting the output has improved considerably. The gap between what a robot sees and what it understands is still real. But it is closing faster than most people expect.

Stay current

The field moves fast. Keep up.

Monthly dispatches on robotics research, industry developments, and educational resources — for curious minds at every level.