Open Access Presentation + Paper
9 September 2021 High-speed 3D imaging with digital fringe projection techniques
Author Affiliations +
Abstract
Fascinated by the marvelous work that Professor Jim Wyant has done in both academy and industry, I decided to pursue a career in the field of 3D optical metrology in my early graduate study. I took advantage of the modern digital technology for 3D optical metrology to address the challenges in high-speed, high-resolution 3D optical metrology and optical information processing. This paper summarizes what I have done to achieve speed breakthroughs in the field of 3D optical metrology using digital fringe projection techniques, to achieve real-time 3D video streaming by compressing enormous 3D video data size, and to solve some practical problems using the techniques we have developed including biomechanics, robotics, forensics, and entertainment.
Conference Presentation

1.

INTRODUCTION

As a graduate student, I had enormous challenges when started my research on high-speed 3D imaging with fringe projection techniques. I got stuck for more than three years without much progress. It was Professor Wyant successes in both both academy and industry that kept me believe that I could pursue a career in this field. Even though I have never met Professor Wyant in person, but his generous help and kind support were critical to reach my current career status.

I took advantage of the modern digital technology for 3D imaging to address the challenges in high-speed, highresolution 3D imaging and optical information processing. As a graduate student, I have developed a system that simultaneously captures, processes and displays 3D geometries at 30 Hz with over 300,000 measurement points per frame, which was unprecedented at that time (a decade ago). My current research focuses on achieving speed breakthroughs by developing the binary defocusing techniques; and exploring novel means to store enormously large 3D geometric data by innovating geometry/video compression methods. The binary defocusing methods coincide with the inherent operation mechanism of the digital-light-processing (DLP) technology, permitting tens of kHz 3D measurement speed at camera pixel spatial resolution. The novel methods of converting 3D data to regular 2D counterparts offer us the opportunity to leverage mature 2D data compression platform, achieving extremely high compression ratios without reinventing the whole data compression infrastructure.

This paper summarizes what we have done to achieve speed breakthroughs in the field of 3D optical metrology using digital fringe projection techniques, to achieve real-time 3D video streaming by compressing enormous 3D video data size, and to solve some practical problems using the techniques we have developed including biomechanics, robotics, forensics, and entertainment.

2.

HIGH-SPEED 3D IMAGING

Figure 1 shows the basic schematics of a digital fringe projection (DFP) system. A projector projects the computergenerated fringe patterns onto an object surface, the object surface geometry deforms the fringe patterns, a camera captures the deformed fringe patterns from a different perspective, and the software algorithm analyzes the captured fringe patterns for 3D reconstruction. The fringe deformation is reflected in carrier phase as well. Phase-shifting algorithm can be used to analyze the phase information for 3D reconstruction after system calibration.

Figure 1:

Schematics of a digital fringe projection (DFP) system.2

00225_psisdg11813_118130v_page_2_1.jpg

2.1

Three-step phase-shifting algorithm

Phase-shifting algorithms are extensively employed partially because they are robust to noise, and simultaneously achieve high measurement speed and accuracy. Three-step phase-shifting algorithm is often used for high-speed 3D imaging because it only requires three fringe patterns for phase extraction,

00225_psisdg11813_118130v_page_2_2.jpg
00225_psisdg11813_118130v_page_2_3.jpg
00225_psisdg11813_118130v_page_2_4.jpg

where I’(x, y) is the average intensity, I”(x, y) is the intensity modulation and ϕ (x, y) is the phase to be solved for. The ratio of I"(x, y) and I’(x, y) reflects the fringe contrast to environmental light. Solving these equations simultaneously gives

00225_psisdg11813_118130v_page_2_5.jpg

here φ(x, y) is the wrapped phase whose value ranges from -π and π with 2π discontinuities, and a spatial or temporal phase algorithm is required to obtain a continuous phase map by removing 2π discontinuities,

00225_psisdg11813_118130v_page_2_6.jpg

where k(x, y) is an integer number for each pixel. Zhang1 summarized the state-of-art absolute phase unwrapping methods for DFP systems.

In addition, averaging these three images washes out fringe stripes and gives a texture image, a photograph of the object

00225_psisdg11813_118130v_page_2_7.jpg

2.2

DFP system calibration

To convert phase to 3D informtion, structured light system calibration is required. The calibration process could be simple or complex, and the achievable accuracy level could be different.2 The digital fringe projection system consists of a camera and a projector. The mathematical model for the camera and the projector is identical and straightforward,3

00225_psisdg11813_118130v_page_2_8.jpg

describing the transformation from the world coordinate system (xw,yw, zw) to the lens coordinate system and the projection from the lens coordinate system to the image plane (u, v). Here T denotes matrix transpose; A is a 3 × 3 intrinsic matrix that models the focal length and the principle point (intrinsic parameters matrix); R is a 3 × 3 rotation matrix and t is a 3 × 1 translation vector, and matrix [R, t] describes the transformation from the world coordinate system to the lens coordinate system (extrinsic parameters matrix).

The camera calibration was well studied at that time, and but DFP system calibration was difficult since calibrating projector was not easy. The projector calibration is difficult because the projector cannot capture images like a camera. Zhang and Huang4 published a method that enables projector to capture images like a camera. Essentially, we project two sets of fringe patterns, and use the carrier phase to establish the corresponding point between camera and the projector. The known correspondence provides a mapping from the camera image to the projector pixel by pixel. Once the projector can capture images, its calibration problem becomes well-studied camera calibration problem, and the DFP system calibration is a standard stereo vision system calibration. This is one of the most extensively adopted methods in the field of structured light 3D imaging.

2.3

Real-time 3D imaging with digital-light-processing (DLP) projector

Started in graduate school, my research works on high-speed 3D imaging. I first tried the single-chip DLP projector for real-time 3D imaging. It took me 4 years to develop hardware system and software algorithms. The hardware system development involves with modifying the projector, encoding three phase-shifted patterns into three color channels, and precisely capturing these channel images in grayscale. The software algorithms includes fast phase wrapping,5 rapid phase unwrapping,6 and 3D reconstruction and visualization on a single personal computer (PC). We took advantage of the single-chip DLP projector’s unique projection mechanism to naturally three phase-shifted fringe images at a high speed.7

Figure 2 shows an image of the real-time 3D imaging system we developed. We were able to achieve 25 Hz 3D imaging on a single computer with dual central processing units (CPUs). The right image shows the subject being captured and the left right shows the real-time 3D image created on the computer screen.The quality of data may not be the best in today’s standard, but it was done in 2004.8

Figure 2:

Results captured by one of the earliest real-time 3D imaging system based on the digital fringe projection technique.8

00225_psisdg11813_118130v_page_3_1.jpg

2.4

Superfast 3D imaging with binary defocusing technique

We then spent many years to achieve speed breakthrough without drastically increasing the hardware cost. We ultimately found that the DLP technology had the potential. DLP technology creates grayscale images by flipping the micro mirror ON/OFF at an extremely high speed.

We came out the binary defocusing technology9 to fully take advantage of the binary image generation nature of DLP technology. As the computer generated binary structured patterns being more and more blurred by defocusing the lens, pseudo sinusoidal patterns can be generated. This pseudo sinusoidal patterns can be analyzed using phase-shifting algorithms for 3D imaging.

DLP Discovery 4100 can refreshes 8-bit grayscale images at approximately 300 Hz, but binary images at over 30 kHz. Since only three images are required for each 3D reconstruction, this platform allows for 3D imaging at a speed over 10 kHz. Zhang et al.10 developed a 3D imaging system that achieved 667 Hz with a three-step phase-shifting algorithm and a DLP Discovery Development kit; Gong and Zhang11 achieved 4,000 Hz rate by using a commercial DLP projector; Li et al.12 achieved 5 kHz with a modified Fourier transform algorithm; and Zuo et al.13 achieved 10 kHz with a modified Fourier transform algorithm.

Figure ?? shows the 3D imaging results of a flying bird robot bird. The robotic bird (Model: XTIM Bionic Bird Avitron V2.0) that we used in this research has a beat frequency of approximately 25 cycles per second with both wings made of inextensible thin membranes. The total span of a single wing is about 150 mm (L) × 70 mm (W). Precisely synchronized with the fringe projection, the camera captures images also at a rate of 5,000 Hz with an image resolution of 800 × 600 pixels.

Figure 3:

3D imaging results of a flying bird robot captured at 5,000 Hz.12 (a) - (c) Three sample frames of 2D images; (d) - (f) three sample frames of reconstructed 3D geometries.

00225_psisdg11813_118130v_page_4_1.jpg

3.

3D IMAGE/VIDEO COMPRESSION AND COMMUNICATIONS

When 3D imaging speed increases, recording highly dynamic events becomes easier and easier. At the same time, storing 3D image data becomes increasingly challenging. As such, we had to find ways to compress 3D data to save storage space.

The approach we use is to create a virtual fringe projection system that virtually projects fringe patterns onto the 3D model and the rendered image can be “see” as the camera image.14 Since the virtually created fringe projection system can be precisely controlled, we can encode all 3D and 2D texture information into a standard RGB 2D image. The difference between the encoded and original image is pretty much random noise.

Streaming 3D videos is very interesting but is even more challenging. For example, it is impossible to stream relatively low-resolution 3D video (800 × 600) through even 5G networks. Fortunately, our compression algorithm converts 3D images to standard 2D images. We can further leverage 2D video compression methods for 3D video compression.15 Leveraging standard video compression methods such as H.264, we could stream 3D video through standard low-bandwidth networks if the median quality 3D video is acceptable.

Figure 4 shows some examples decoded from the data stored using various H.264 qualities. The lossless encoding achieves a compression ratio of 129:1. When using H.264 to lossy encode video at various levels, higher compression ratios can be achieved, requiring slower networks for real-time streaming. This figure future shows that even at the very low bitrate of 4.8 Mbps the decoded 3D images are still of great quality in both geometry and color texture.

Figure 4:

Compressing encoded 3D data with the H.264 video codec at different qualities.16 First row shows the decoded images from the video stored with lossless H.264 (compression ratio 129:1); second row shows the decoded images from H.264 video using 4:2:0 subsampling and a constant rate factor of 6 (compression ratio 551:1); and third row shows the decoded images using the same encoding instead with a constant rate factor of 12 (compression ratio 1,602:1).

00225_psisdg11813_118130v_page_5_1.jpg

We developed a 3D video streaming technology called Holostream that enables real-time 3D video capture, compression, streaming, decompression, and visualization.16 This was probably the first technology that enables high-fidelity 3D video streaming using standard cellphone networks. Figure 5 shows the successful implementation of our holostream system. The left image shows the Acquisition and Compression Modules. The right image shows multiple users on their mobile devices receiving and interacting with the live 3D video stream delivered across a standard wireless network.

Figure 5:

Successful implementation of the Holostream system.16

00225_psisdg11813_118130v_page_5_2.jpg

3.1

Other capabilities

We developed a real-time multimodal system that can simultaneously record 3D geometry, 2D texture and 2D surface temperature information.17 We found a way to calibrate the thermal camera and normal camera in the same coordinate system. Once the system is calibrated, we can map the temperature map to 3D geometry pixel by pixel for each frame.

We developed technologies that can capture microstructures at a high speed as well.18 For example, we developed a 500 Hz 3D imaging system. This system achieved 15 μm spatial resolution and 2 μm depth resolution.19

The silicon based DLP technology has light spectrum limitation, and the light illumination limitation. Since only binary patterns can be used for 3D imaging, we developed a mechanical projector for 3D imaging that successfully achieved 10 kHz with such a technology.20

Recently, one of the problems we are working on autofocusing. Autofocusing is a mature technology for 2D imaging but difficult for 3D imaging. We developed a technology that uses two projectors for triangulation, and a camera with electrically tunable lens for fringe image capture.21 Since depth field of the projectors is very large, the 2D autofocusing method can be used for 3D imaging.

Another problem we are working on is to achieve real-time auto-exposure time control. We have developed optimal exposure time determination method because of the flexibility of binary defocusing method.22 We further developed a method that can quickly determine the best exposure time regardless whether the initial image is too bright or too dark.23

4.

SOME APPLICATIONS

3D imaging has been numerous applications ranging from healthcare for overall diagnosis offered by the third dimension, aerospace and defense for mapping, imaging and quality control of complex part manufacturing, along with others. The rapid growth of 3D sensor industry was driven by rising demands in consumer electronics for devices enhanced by 3D sensors (e.g., smartphones, tablets). In this section, we only pick a few surprising applications that we are luckily involved with.

The forensics community is interested in the high-resolution 3D imaging technology for crime scene documentation. Our prior research has demonstrated that high-resolution 3D imaging outperforms the current practices (i.e., 2D photography and casting).24 This community needs a fully automated 3D camera that does not require any training for examiners and practitioners to operate. It remains a challenge to develop such smart 3D imaging technology for this field.

Figure 6 shows the test of shoe impression in clay with grain size being smaller than 0.01 mm. The red circles outline locations that correspond to test cuts in the shoe outsole and the orange circles outline locations that correspond to small rockholds in the shoe outsole. Figure 6(a) shows the evidence quality 2D photograph: resolved 10 test cuts and the two small rockholds; Figure 6(b) photograph of the conventional physical cast: resolved two small rockholds and all 10 test cuts, but the small bubbles produced in the casting process made visualization difficult for several of the cuts; Figure 6(c) shows the low-resolution 3D image (approximately 137 dots per inch, dpi) of the impression that resolved 9 out of the 10 test cuts, and two small rockholds; Figure 6(d) shows high-resolution 3D image(approximately 400 dpi) of the impression that resolved all 10 test cuts and two small rockholds.

Figure 6:

Test of shoe impression in clay. The red circles outline locations that correspond to test cuts in the shoe outsole and the orange circles outline locations that correspond to small rockholds in the shoe outsole.24 (a) Evidence quality 2D photograph; (b) photograph of the conventional physical cast; (c) low-resolution 3D image of the impression; (d) high-resolution 3D 3D image of the impression.

00225_psisdg11813_118130v_page_6_1.jpg

The real-time 3D imaging we developed caught some attentions. But the most surprising early interest came from the entertainment industry. Radiohead, and English rock band used our technology the created this 3D music video “house of cards”.25

The superfast 3D imaging was very useful in biomedical engineering field. For example, working with our collaborators, we developed a multimodal imaging system that can simultaneously record 3D geometry, 2D texture image, and 2D fluorescence images at kHz range.26 The recorded information can be used to analyze the mechanics of heart (such as strain field on the bottom left video), as well as diagnose heard diseases.

5.

SUMMARY

Thank you, Professor Jim Wyant, for what you have done for the community and being my hero. I am so grateful for your generous support to someone like me whom you never met. Your successes in both academia and industry inspired me to pursue a career in 3D imaging and still motivate me to work hard every day.

REFERENCES

[1] 

Zhang, S., “Absolute phase retrieval methods for digital fringe projection profilometry: a review,” Opt. Laser Eng., 107 28 –37 (2018). https://doi.org/10.1016/j.optlaseng.2018.03.003 Google Scholar

[2] 

Xu, J. and Zhang, S., “Status, challenges, and future perspectives of fringe projection profilometry,” Optics and Lasers in Engineering, 135 106193 (2020). https://doi.org/10.1016/j.optlaseng.2020.106193 Google Scholar

[3] 

Zhang, Z., “A flexible new technique for camera calibration,” IEEE Trans. Pattern Anal. Mach. Intell., 22 (11), 1330 –1334 (2000). https://doi.org/10.1109/34.888718 Google Scholar

[4] 

Zhang, S. and Huang, P. S., “Novel method for structured light system calibration,” Opt. Eng., 45 (8), 083601 (2006). https://doi.org/10.1117/1.2336196 Google Scholar

[5] 

Huang, P. S. and Zhang, S., “Fast three-step phase-shifting algorithm,” Appl. Opt., 45 (21), 5086 –5091 (2006). https://doi.org/10.1364/AO.45.005086 Google Scholar

[6] 

Zhang, S., Li, X., and Yau, S.-T., “Multilevel quality-guided phase unwrapping algorithm for real-time threedimensional shape reconstruction,” Appl. Opt., 46 (1), 50 –57 (2007). https://doi.org/10.1364/AO.46.000050 Google Scholar

[7] 

Zhang, S. and Huang, P. S., “High-resolution real-time three-dimensional shape measurement,” Opt. Eng, 45 (12), 123601 (2006). https://doi.org/10.1117/1.3516729 Google Scholar

[8] 

Zhang, S. and Huang, P., “High-resolution, real-time 3-d shape acquisition,” IEEE Comp. Vis. and Patt. Recogn. Workshop, 3 28 –37 (2004). Google Scholar

[9] 

Lei, S. and Zhang, S., “Flexible 3-d shape measurement using projector defocusing,” Opt. Lett., 34 (20), 3080 –3082 (2009). https://doi.org/10.1364/OL.34.003080 Google Scholar

[10] 

Zhang, S., van der Weide, D., and Oliver, J., “Superfast phase-shifting method for 3-d shape measurement,” Opt. Express, 18 (9), 9684 –9689 (2010). https://doi.org/10.1364/OE.18.009684 Google Scholar

[11] 

Gong, Y. and Zhang, S., “Ultrafast 3-d shape measurement with an off-the-shelf dlp projector,” Opt. Express, 18 (19), 19743 –19754 (2010). https://doi.org/10.1364/OE.18.019743 Google Scholar

[12] 

Li, B. and Zhang, S., “Novel method for measuring dense 3d strain map of robotic flapping wings,” Measurement Science and Technology, 29 (4), 045402 (2018). https://doi.org/10.1088/1361-6501/aaa4cc Google Scholar

[13] 

Zuo, C., Tao, T., Feng, S., Huang, L., Asundi, A., and Chen, Q., “Micro fourier transform profilometry (μftp): 3d shape measurement at 10,000 frames per second,” Optics and Lasers in Engineering, 102 70 –91 (2018). https://doi.org/10.1016/j.optlaseng.2017.10.013 Google Scholar

[14] 

Karpinsky, N. and Zhang, S., “Composite phase-shifting algorithm for three-dimensional shape compression,” Opt. Eng., 49 (6), 063604 (2010). https://doi.org/10.1117/1.3456632 Google Scholar

[15] 

Karpinsky, N. and Zhang, S., “Holovideo: Real-time 3d video encoding and decoding on gpu,” Opt. Laser Eng., 50 (2), 280 –286 (2012). https://doi.org/10.1016/j.optlaseng.2011.08.002 Google Scholar

[16] 

Bell, T., Allebach, J., and Zhang, S., “Holostream: High-accuracy, high-speed 3d range video,” in IS&T International Conference on Electronic Imaging, (2018). Google Scholar

[17] 

An, Y. and Zhang, S., “High-resolution, real-time 3d surface geometry and temperature measurement,” Opt. Express, 24 (13), 14552 –14563 (2016). https://doi.org/10.1364/OE.24.014552 Google Scholar

[18] 

Li, B. and Zhang, S., “Flexible calibration method for microscopic structured light system using telecentric lens,” Opt. Express, 23 (20), 25795 –25803 (2015). https://doi.org/10.1364/OE.23.025795 Google Scholar

[19] 

Li, B. and Zhang, S., “Microscopic structured light 3d profilometry: binary defocusing technique vs sinusoidal fringe projection,” Opt. Laser Eng., 96 117 –123 (2017). https://doi.org/10.1016/j.optlaseng.2016.06.009 Google Scholar

[20] 

Hyun, J.-S., Chiu, G. T.-C., and Zhang, S., “High-speed and high-accuracy 3d surface measurement using a mechanical projector,” Opt. Express, 26 (2), 1474 –1487 (2018). https://doi.org/10.1364/OE.26.001474 Google Scholar

[21] 

Zhong, M., Hu, X., Chen, F., Xiao, C., Peng, D., and Zhang, S., “Autofocusing method for digital fringe projection system with dual projectors,” Optics Express, 28 (9), 12609 –12620 (2020). https://doi.org/10.1364/OE.392006 Google Scholar

[22] 

Ekstrand, L. and Zhang, S., “Auto-exposure for three-dimensional shape measurement with a digital-light-processing projector,” Opt. Eng, 50 (12), 123603 (2011). https://doi.org/10.1117/1.3662387 Google Scholar

[23] 

Zhang, S., “Rapid and automatic optimal exposure control for digital fringe projection technique,” Opt. Laser Eng., 128 106029 (2020). https://doi.org/10.1016/j.optlaseng.2020.106029 Google Scholar

[24] 

Liao, Y.-H., Hyun, J.-S., Feller, M., Bell, T., Bortins, I., Wolfe, J., Baldwin, D., and Zhang, S., “Portable highresolution automated 3d imaging for footwear and tire impression capture,” Journal of Forensic Sciences, 66 (1), 112 –128 (2021). https://doi.org/10.1111/jfo.v66.1 Google Scholar

[25] 

, “Radiohead, Radiohead - house of cards,” (2008) https://www.youtube.com/watch?v=8nTFjVm9sTQ Google Scholar

[26] 

Laughner, J. I., Zhang, S., Li, H., Shao, C. C., and Efimov, I. R., “Mapping cardiac surface mechanics with structured light imaging,” American Journal of Physiology - Heart and Circulatory Physiology, 303 (6), H712 –H720 (2002). https://doi.org/10.1152/ajpheart.00269.2012 Google Scholar
© (2021) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Song Zhang "High-speed 3D imaging with digital fringe projection techniques", Proc. SPIE 11813, Tribute to James C. Wyant: The Extraordinaire in Optical Metrology and Optics Education, 118130V (9 September 2021); https://doi.org/10.1117/12.2567675
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
3D video streaming

3D video compression

Projection systems

3D image processing

Image compression

Video

Video compression

RELATED CONTENT

Point cloud compression on the basis of 3D motion estimation...
Proceedings of SPIE (September 06 2019)
Field tests of error-resilient 3D SPIHT vs. MPEG-2
Proceedings of SPIE (November 21 2002)
3D video compression with the H.264 codec
Proceedings of SPIE (January 30 2012)
Predictive Coding of Depth Images Across Multiple Views
Proceedings of SPIE (March 05 2007)
A new video codec based on 3D DTCWT and vector...
Proceedings of SPIE (October 01 2011)

Back to Top