We introduce an innovative concept for 3D imaging that utilizes a structured light principle. While our design is specifically tailored for collaborative scenarios involving mobile transport robots, it is also applicable to similar contexts. Our system pairs a standard camera with a projector that employs a diffractive optical element (DOE) and a collimated laser beam to generate a coded light pattern. This allows a three-dimensional measurement of objects from a single camera shot. The main objective of the 3D-sensor is to facilitate the development of automatic, dynamic and adaptive logistics processes capable of managing diverse and unpredictable events. The key novelty of our proposed system for triangulation-based 3D reconstruction is the unique coding of the light pattern, ensuring robust and efficient 3D data generation, even within challenging environments such as industrial settings. Our pattern relies on a perfect submap, a matrix featuring pseudorandomly distributed dots, where each submatrix of a fixed size is distinct from the others. Based on the size of the working space and known geometrical parameters of the optical components, we establish vital design constraints like minimum pattern size, uniqueness window size, and minimum Hamming distance for the design of an optimal pattern. We empirically examine the impact of these pattern constraints on the quality of the 3D data and compare our proposed encoding with some single-shot patterns found in existing literature. Additionally, we provide detailed explanations on how we addressed several challenges during the fabrication of the DOE, which are crucial in determining the usability of the application. These challenges include reducing the 0th diffraction order, accommodating a large horizontal field of view, achieving high point density, and managing a large number of points. Lastly, we propose a real-time processing pipeline that transforms an image of the captured dot pattern into a high-resolution 3D point cloud using a computationally efficient pattern decoding methodology.
This paper presents a principle for scene-related camera calibration in Manhattan worlds. The proposed estimation of extrinsic camera parameters from vanishing points represents a useful alternative to the traditional target-based calibration methods, especially in large urban or industrial environments. We analyse the effects of errors in the calculation of camera poses and derive general restrictions for the use of our approach. In addition, we present methods for calculating the position and orientation of several cameras to a world coordinate system and discuss the effect of imprecise or incorrectly calculated vanishing points. Our approach was evaluated with real images of a prototype for human-robot collaboration installed at ZBS e.V. The results were compared with a perspective n-Point (PnP) method.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.