News Release

3-D scanning with water

3-D shape acquisition using water displacement as the shape sensor for the reconstruction of complex objects

Peer-Reviewed Publication

Association for Computing Machinery

3-D Scanning with Water

image: 3-D scanning using Dip Transform. The object is dipped in water (left) using a robot arm, acquiring a dip transform by which the object is reconstructed (right). The team's method produces a complete reconstruction of the complex shape, including its hidden and inner regions. view more 

Credit: Courtesy of ACM SIGGRAPH 2017

CHICAGO, July 21, 2017--A global team of computer scientists and engineers have developed an innovative technique that more completely reconstructs challenging 3D objects. An ancient scientific breakthrough serves as the foundation for the team's modern, innovative solution to remaining challenges in current 3D shape reconstruction. This new approach to 3D shape acquisition is based on the well-known fluid displacement discovery by Archimedes and turns modeling surface reconstruction into a volumetric problem. Most notably, their method accurately reconstructs even hidden parts of an object that typical 3D laser scanners are not able to capture.

The research, "Dip Transform for 3D Shape Reconstruction," is authored by a team from Tel-Aviv University, Shandong University, Ben-Gurion University and University of British Columbia. They will present their work at SIGGRAPH 2017 in Los Angeles, 30 July to 3 August. An annual conference, SIGGRAPH spotlights the most innovative in computer graphics research and interactive techniques worldwide.

Traditional 3D shape acquisition or reconstruction methods are based on optical devices, most commonly, laser scanners and cameras that successfully sample the visible shape surface. But this common approach tends to be noisy and incomplete. Most devices can only scan what is visible to them but hidden parts of an object remain inaccessible to the scanner's line of sight. For instance, a typical laser scanner cannot accurately capture the belly or underside of an elephant statue, which is hidden from its line of sight.

The team's dip transform to reconstruct complex 3D shapes utilizes liquid, computing the volume of a 3D object versus its surface. By following this method, a more complete acquisition of an object, including hidden details, can be reconstructed in 3D. Liquid has no line of sight; it can penetrate cavities and hidden parts, and it treats transparent and glossy materials identically to opaque materials, thus bypassing the visibility and optical limitations of optical and laser-based scanning devices.

For the study, the team implemented a low-cost 3D dipping apparatus--objects in the water tank were dipped via a robotic arm. By dipping an object in the liquid along an axis, they were able to measure the displacement of the liquid volume and form that into a series of thin volume slices of the shape. By repeatedly dipping the object in the water at various angles, the researchers were able to capture the geometry of the given object, including the parts that would have normally been hidden by a laser or optical 3D scanner.

The team's dip transform technique is related to computed tomography--an imaging method that uses optical systems for accurate scanning or to produce detailed pictures. However, the challenge with this more traditional method is that tomography-based devices are bulky and expensive and can only be used in a safe, customized environment. The team's approach is both safe and inexpensive, and a much more appealing alternative for generating a complete shape at a low-computational cost using an innovative data collection method.

In the study, they demonstrated the new technique on 3D shapes with a range of complexity, including a hand balled up into a fist, a mother-child hugging and a DNA double helix. Their results show that the dip reconstructions are nearly as accurate as the original 3D model, paving the way to a new world of non-optical 3D shape acquisition techniques.

###

The work was supported by the Beijing Government (through Beijing Film Academy) and the Israeli Science Foundation and Natural Science Foundation of China (through Shandong University). The research team is comprised of Kfir Aberman, Tel Aviv University, Advanced Innovation Center for Future Visual Entertainment; Oren Katzir, Tel Aviv University, Advanced Innovation Center for Future Visual Entertainment; Qiang Zhou, Shandong University; Zegang Luo, Shandong University; Andrei Sharf, Advanced Innovation Center for Future Visual Entertainment, Ben-Gurion University of the Negev; Chen Greif, The University of British Columbia; Baoquan Chen, Shandong University; and Daniel Cohen-Or, Tel-Aviv University.

For the full paper and video, visit http://irc.cs.sdu.edu.cn/3dshape/. To register for SIGGRAPH 2017 and hear from the authors themselves, visit http://s2017.SIGGRAPH.org.

About ACM, ACM SIGGRAPH and SIGGRAPH 2017

ACM, the Association for Computing Machinery, is the world's largest educational and scientific computing society, uniting educators, researchers, and professionals to inspire dialogue, share resources, and address the field's challenges. ACM SIGGRAPHis a special interest group within ACM that serves as an interdisciplinary community for members in research, technology, and applications in computer graphics and interactive techniques. SIGGRAPH is the world's leading annual interdisciplinary educational experience showcasing the latest in computer graphics and interactive techniques. SIGGRAPH 2017 , marking the 44th annual conference hosted by ACM SIGGRAPH, will take place from 30 July-3 August at the Los Angeles Convention Center in Los Angeles.


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.