One sensor is never enough
Close one eye and try to judge the distance to something
across the room. You can do it, roughly, using depth cues in the visual scene.
But it is noticeably harder than with both eyes. Now imagine trying to drive a
car with one eye, at night, in rain. Suddenly you are acutely aware of what you
are missing.
Physical AI systems face a more severe version of this
problem. Any single sensor type captures a partial, limited picture of the
physical world. Cameras see color and texture but struggle with depth and low
light. Lidar measures depth and geometry but cannot detect color or see through
glass. Radar can work in darkness and rain but lacks fine-grained spatial
resolution. Force sensors measure contact but cannot see at all.
The solution is sensor fusion: combining the outputs of
multiple sensor types so that each compensates for the limitations of the
others. A camera-lidar pair sees both color and depth. Adding radar handles
low-visibility conditions. Adding force sensing handles contact uncertainty.
The system that fuses all of these has a much richer, more robust understanding
of its physical environment than any single sensor could provide.
This is not a cutting-edge feature reserved for the most
advanced physical AI systems. For any robot or autonomous system that needs to
reliably operate in a real environment, sensor fusion is table stakes.
The challenge is not the hardware
Physical AI teams sometimes assume that sensor fusion is
primarily a hardware problem: get the right sensors, mount them correctly,
write the drivers. The harder challenge is often the data problem that comes
after.
Each sensor has its own coordinate system, its own
timestamp format, its own data rate, and its own noise characteristics. A
camera captures images at 30 frames per second. A lidar scanner produces a
complete point cloud every tenth of a second. A radar sensor updates at a
different rate still. When you want to reason about what all three sensors saw
at the same physical moment in time, you first need to align them all to that
moment.
This alignment has two components. Temporal calibration
ensures that sensor readings from the same physical moment in time are matched
together correctly. Spatial calibration ensures that a point in the lidar
coordinate system and a pixel in the camera image that correspond to the same
physical object in the real world are correctly associated with each other.
Both of these calibrations require careful, ongoing
attention. Calibration shifts over time as sensors vibrate, are bumped, or are
replaced. Even small calibration errors compound: a camera-lidar extrinsic
calibration error of two centimeters means that a lidar point cloud label for
an object at five meters range corresponds to the wrong location in the camera
image by a visually significant margin.
Sensor fusion is only as reliable as the calibration
between sensors. And calibration drifts. This is not a one-time problem to
solve before deployment. It is an ongoing operational concern.
Why this complicates annotation significantly
If you are annotating data from a single sensor, the
annotation task is bounded. Draw the box around the object in the image.
Segment the region of the point cloud that belongs to this object. The
annotator is reasoning within a single coordinate system and a single data
representation.
For fused multi-sensor data, annotation is more complex.
An object that the annotator labels in the lidar point cloud must be
consistently labeled in every camera view that captured it simultaneously. A
pedestrian's position needs to be consistent across camera, lidar, and radar
representations at the same timestamp. The label must be not just locally
correct in one sensor view but globally consistent across all of them.
Cross-sensor annotation consistency is something that most
annotation platforms and most annotation training programs are not designed
for. The tools were built for single-sensor tasks. Annotators are often trained
on single-sensor workflows and then asked to apply them to multi-sensor data,
which requires a fundamentally different spatial reasoning process.
The consequence is systematic inconsistency in fused
dataset annotations: labels that are correct in one sensor view and imprecise
in others, temporal alignments that are slightly off, and object identities
that fragment and merge inconsistently across sensor modalities. Each of these
is an invisible error that the model learns from.
Time synchronization: the detail that matters more than it
seems
A physical AI system operating in a dynamic environment
records data from multiple sensors that are all changing rapidly. A robot
moving its arm at half a meter per second covers five centimeters in a tenth of
a second. A vehicle traveling at city speeds moves over a meter in the same
interval.
If sensor streams are not tightly synchronized, the label
placed on a frame at time T in the camera is being associated with the lidar
scan captured at time T plus a hundred milliseconds. For a slowly moving scene,
this might not matter. For a fast-moving manipulation task or a vehicle
traveling through traffic, that temporal offset puts the annotation in the
wrong physical location.
Tight temporal synchronization requires hardware trigger
signals or software timestamping at the driver level. It requires storing
timestamps with enough precision to detect and correct for latency differences
between sensor pipelines. And it requires annotation workflows that use
synchronized multi-sensor views rather than labeling each sensor stream
independently.
Teams that are not rigorous about temporal synchronization
often discover the problem only when they train a model and notice that learned
object representations are blurry or inconsistent in ways they cannot initially
explain.
What well-designed sensor fusion annotation looks like in
practice
The benchmark for good multi-sensor annotation is that a
label placed on an object should be consistent in every sensor view
simultaneously. An annotator working on a pedestrian detection dataset should
see the pedestrian's 3D bounding box correctly aligned with the lidar point
cloud, correctly projected into every camera image, and correctly associated
with the radar return, all at the same timestamp.
Achieving this requires annotation tooling that displays
synchronized multi-sensor views side by side, allows annotators to adjust a
label in one view and see the change propagate to all others, and flags
inconsistencies where a label is well-aligned in one view but misaligned in
another.
It requires annotators trained specifically for
multi-sensor tasks, not just annotators who have been given access to new
tooling. The spatial reasoning process is different, the error modes are
different, and the quality checking requirements are different.
And it requires quality control that checks cross-sensor
consistency explicitly, not just annotation quality within individual sensor
streams. An annotation program that checks whether bounding boxes are tight in
the camera images but never verifies that those boxes align with the lidar
ground truth will produce a dataset with systematic cross-sensor errors that
only become visible when model performance falls short of what the individual
sensor annotation quality would predict.
The physical world is multi-dimensional. Physical AI
training data needs to reflect that fully, not piecemeal. Sensor fusion is not
a feature to add later; it is the foundation from the start.
One sensor is never enough
Close one eye and try to judge the distance to something across the room. You can do it, roughly, using depth cues in the visual scene. But it is noticeably harder than with both eyes. Now imagine trying to drive a car with one eye, at night, in rain. Suddenly you are acutely aware of what you are missing.
Physical AI systems face a more severe version of this problem. Any single sensor type captures a partial, limited picture of the physical world. Cameras see color and texture but struggle with depth and low light. Lidar measures depth and geometry but cannot detect color or see through glass. Radar can work in darkness and rain but lacks fine-grained spatial resolution. Force sensors measure contact but cannot see at all.
The solution is sensor fusion: combining the outputs of multiple sensor types so that each compensates for the limitations of the others. A camera-lidar pair sees both color and depth. Adding radar handles low-visibility conditions. Adding force sensing handles contact uncertainty. The system that fuses all of these has a much richer, more robust understanding of its physical environment than any single sensor could provide.
This is not a cutting-edge feature reserved for the most advanced physical AI systems. For any robot or autonomous system that needs to reliably operate in a real environment, sensor fusion is table stakes.
The challenge is not the hardware
Physical AI teams sometimes assume that sensor fusion is primarily a hardware problem: get the right sensors, mount them correctly, write the drivers. The harder challenge is often the data problem that comes after.
Each sensor has its own coordinate system, its own timestamp format, its own data rate, and its own noise characteristics. A camera captures images at 30 frames per second. A lidar scanner produces a complete point cloud every tenth of a second. A radar sensor updates at a different rate still. When you want to reason about what all three sensors saw at the same physical moment in time, you first need to align them all to that moment.
This alignment has two components. Temporal calibration ensures that sensor readings from the same physical moment in time are matched together correctly. Spatial calibration ensures that a point in the lidar coordinate system and a pixel in the camera image that correspond to the same physical object in the real world are correctly associated with each other.
Both of these calibrations require careful, ongoing attention. Calibration shifts over time as sensors vibrate, are bumped, or are replaced. Even small calibration errors compound: a camera-lidar extrinsic calibration error of two centimeters means that a lidar point cloud label for an object at five meters range corresponds to the wrong location in the camera image by a visually significant margin.
Sensor fusion is only as reliable as the calibration between sensors. And calibration drifts. This is not a one-time problem to solve before deployment. It is an ongoing operational concern.
Why this complicates annotation significantly
If you are annotating data from a single sensor, the annotation task is bounded. Draw the box around the object in the image. Segment the region of the point cloud that belongs to this object. The annotator is reasoning within a single coordinate system and a single data representation.
For fused multi-sensor data, annotation is more complex. An object that the annotator labels in the lidar point cloud must be consistently labeled in every camera view that captured it simultaneously. A pedestrian's position needs to be consistent across camera, lidar, and radar representations at the same timestamp. The label must be not just locally correct in one sensor view but globally consistent across all of them.
Cross-sensor annotation consistency is something that most annotation platforms and most annotation training programs are not designed for. The tools were built for single-sensor tasks. Annotators are often trained on single-sensor workflows and then asked to apply them to multi-sensor data, which requires a fundamentally different spatial reasoning process.
The consequence is systematic inconsistency in fused dataset annotations: labels that are correct in one sensor view and imprecise in others, temporal alignments that are slightly off, and object identities that fragment and merge inconsistently across sensor modalities. Each of these is an invisible error that the model learns from.
Time synchronization: the detail that matters more than it seems
A physical AI system operating in a dynamic environment records data from multiple sensors that are all changing rapidly. A robot moving its arm at half a meter per second covers five centimeters in a tenth of a second. A vehicle traveling at city speeds moves over a meter in the same interval.
If sensor streams are not tightly synchronized, the label placed on a frame at time T in the camera is being associated with the lidar scan captured at time T plus a hundred milliseconds. For a slowly moving scene, this might not matter. For a fast-moving manipulation task or a vehicle traveling through traffic, that temporal offset puts the annotation in the wrong physical location.
Tight temporal synchronization requires hardware trigger signals or software timestamping at the driver level. It requires storing timestamps with enough precision to detect and correct for latency differences between sensor pipelines. And it requires annotation workflows that use synchronized multi-sensor views rather than labeling each sensor stream independently.
Teams that are not rigorous about temporal synchronization often discover the problem only when they train a model and notice that learned object representations are blurry or inconsistent in ways they cannot initially explain.
What well-designed sensor fusion annotation looks like in practice
The benchmark for good multi-sensor annotation is that a label placed on an object should be consistent in every sensor view simultaneously. An annotator working on a pedestrian detection dataset should see the pedestrian's 3D bounding box correctly aligned with the lidar point cloud, correctly projected into every camera image, and correctly associated with the radar return, all at the same timestamp.
Achieving this requires annotation tooling that displays synchronized multi-sensor views side by side, allows annotators to adjust a label in one view and see the change propagate to all others, and flags inconsistencies where a label is well-aligned in one view but misaligned in another.
It requires annotators trained specifically for multi-sensor tasks, not just annotators who have been given access to new tooling. The spatial reasoning process is different, the error modes are different, and the quality checking requirements are different.
And it requires quality control that checks cross-sensor consistency explicitly, not just annotation quality within individual sensor streams. An annotation program that checks whether bounding boxes are tight in the camera images but never verifies that those boxes align with the lidar ground truth will produce a dataset with systematic cross-sensor errors that only become visible when model performance falls short of what the individual sensor annotation quality would predict.
The physical world is multi-dimensional. Physical AI training data needs to reflect that fully, not piecemeal. Sensor fusion is not a feature to add later; it is the foundation from the start.