top of page
Blue Argus Demo
10:56
Blue Argus Demo
Learn about Blue Sky Robotics' Computer Vision Package: Blue Argus!
Features: Houston
00:33
Features: Houston
Blue Sky Robotics' low-code automation platform
Features: Analytics Dashboard
00:56
Features: Analytics Dashboard
Blue Sky Robotics' control center analytics dashboard
Meet the "Hands" of your robot!
00:30
Meet the "Hands" of your robot!
Meet the "Hands" of your robot! 🤖 End effectors are how robotic arms interact with their world. We’re breaking down the standard UFactory gripper—the versatile go-to for most of our automation tasks. 🦾✨ #UFactory #xArm #Robotics #Automation #Engineering #TechTips #shorts Learn more at https://f.mtr.cool/jenaqtawuz

3D Sensing Technology: How It Works and Why It's the Foundation of Modern Robot Automation

  • Apr 8
  • 5 min read

Every meaningful advance in robotic automation over the past decade traces back to a single capability: the ability of a robot to perceive its environment in three dimensions. Not just a flat image. Not just presence or absence. Full spatial awareness, depth, geometry, orientation, surface texture, captured in real time and translated into motion.


That capability is 3D sensing technology. It is the foundation on which bin picking, vision-guided palletizing, inline dimensional inspection, and precision assembly automation are built. Without it, robots are limited to fixed, controlled environments where nothing ever changes. With it, robots can handle the variability that defines real manufacturing and logistics operations.


This post explains the core technologies behind 3D sensing, how each one works, where each one fits, and how to match the right technology to a specific application.


What 3D Sensing Technology Measures


All 3D sensing technology does the same fundamental thing: it captures the distance from a sensor to surfaces in a scene, producing spatial data with three coordinates, X, Y, and Z, for every measured point. The collection of those points is a point cloud.


What distinguishes one 3D sensing technology from another is the method used to measure that distance. Each method involves different physics, different hardware, and different tradeoffs in accuracy, speed, cost, and performance on challenging surfaces.


The three technologies that dominate industrial robotics applications are structured light, stereo vision, and Time-of-Flight. Understanding what each one actually does is the most direct path to choosing correctly.


Structured Light: The Industrial Standard for Precision


Structured light sensing works by projecting a known pattern of light, typically a grid, a series of stripes, or a more complex coded pattern, onto the target scene. A camera captures the projected pattern. Because the system knows exactly what the pattern should look like, it can calculate depth by measuring how the pattern deforms as it conforms to the shapes of objects in the scene.


The deformation of the pattern encodes depth. A flat surface produces an undistorted pattern. A curved surface distorts it. An edge produces a sharp discontinuity. Software processes these deformations and reconstructs a dense, accurate 3D point cloud.


Structured light produces the highest quality point clouds of the three main technologies. It handles a wide range of surface conditions including reflective metals, dark materials, and objects with complex geometric features that defeat simpler sensing approaches. Mech-Mind's Mech-Eye industrial cameras use structured light, as do most industrial-grade 3D cameras deployed in demanding bin picking and precision inspection applications.


The tradeoff is speed. Structured light systems typically require the scene to be relatively still during acquisition. For high-speed conveyor applications where objects are moving continuously, this can be a limitation.


Stereo Vision: Accessible and Versatile


Stereo vision sensing uses two cameras offset from each other, similar to how human eyes work, to calculate depth from the difference between the two images. Each camera captures the scene from a slightly different angle. For any given point visible in both images, the horizontal shift between where it appears in the left image versus the right image (called disparity) encodes how far away it is. More disparity means closer; less disparity means farther.


Stereo vision produces point clouds that are less dense and less accurate than structured light, particularly on surfaces that lack texture or on reflective materials where the two cameras capture inconsistent information. But it is significantly more affordable, more compact, and fast enough for most manipulation tasks.


The Intel RealSense D435 and Luxonis OAK-D-Pro-PoE are the most widely deployed stereo cameras in cobot applications. UFactory's open-source vision SDK natively supports both cameras across the xArm and Lite 6 lineup, including hand-eye calibration examples and Python-based integration code.


For entry-level bin picking, flexible pick and place, and machine tending with standard industrial parts under reasonable lighting conditions, stereo vision is often the right choice. The performance is sufficient and the deployment cost is substantially lower than structured light.


Time-of-Flight: Speed and Coverage at Scale


Time-of-Flight sensing works by emitting pulses of infrared or laser light and precisely measuring how long each pulse takes to return from the scene. Since light travels at a known speed, the round-trip time directly encodes distance. The sensor builds a depth map by measuring return times across its entire field of view simultaneously.


ToF sensors produce depth maps in real time at high frame rates, often 30 frames per second or faster, which makes them well suited for fast-moving applications where the scene changes continuously. They maintain reliable performance across variable lighting conditions, including bright factory floors where structured light systems can struggle with ambient interference.


The tradeoff is accuracy and resolution. ToF sensors typically produce lower-resolution depth maps with less point cloud density than structured light, and their absolute accuracy at close ranges can be lower. For applications where the robot needs to monitor a large area or track fast-moving objects, ToF excels. For applications requiring sub-millimeter precision on specific part features, structured light is the better choice.


Matching Technology to Application


The decision framework maps cleanly to application requirements.

For precision inspection of small features, reflective metal parts, or complex geometries, structured light is the required technology. The density and accuracy of the point cloud are what enable reliable measurement at the tolerances these applications demand.


For general-purpose bin picking, pick and place, and machine tending with standard parts, stereo vision provides sufficient accuracy at a fraction of the cost. It is the right starting point for teams building their first vision-guided cell.


For fast-moving conveyor applications, large-area monitoring, or environments with variable lighting, Time-of-Flight delivers the frame rate and robustness that neither structured light nor stereo can match at comparable cost.


Many production cells combine technologies: a stereo camera for general guidance and a structured light camera or laser profiler for precision inspection at a dedicated station.


Connecting 3D Sensing Technology to a Complete Cell


The sensor is only the perception layer. Blue Sky Robotics' Blue Argus platform ships camera, compute, mount, and vision software as a complete kit, eliminating the custom integration work that typically separates a capable sensor from a working robot cell. Blue Argus uses pre-trained vision models that recognize novel objects without per-SKU training, which means the system works on day one regardless of what parts arrive.


For the robot arm layer, the UFactory Lite 6 ($3,500) is the most accessible entry point for stereo vision-guided applications. The Fairino FR5 ($6,999) covers the widest range of production applications, and the Fairino FR10 ($10,199) handles heavier bin picking and palletizing tasks alongside industrial structured light cameras.


Getting Started


Use our Cobot Selector to match an arm to your sensing application, or the Automation Analysis Tool to model the ROI. Browse our full UFactory lineup and Fairino cobots, or book a live demo.


FAQ


What is 3D sensing technology?

3D sensing technology refers to sensors and systems that capture the spatial geometry of a scene in three dimensions, producing depth data alongside standard image information. In robotics, 3D sensing gives robot arms the spatial awareness they need to locate, grasp, and interact with objects in variable positions and orientations.


Which 3D sensing technology is most accurate?

Structured light produces the most accurate and dense point clouds, making it the standard for demanding inspection and bin picking applications. For applications requiring micron-level measurement accuracy, laser profiler sensors achieve Z repeatability as precise as 0.2 micrometers.


Is stereo vision good enough for bin picking?

For standard industrial parts with sufficient surface texture under controlled lighting, stereo vision can support bin picking effectively. For reflective metal parts, dark materials, or applications requiring high pick success rates on complex geometries, structured light produces significantly more reliable results.

bottom of page