Any Mobile Camera Can Be A 3D Scanner
Written by David Conrad   
Tuesday, 10 December 2013

Take a standard mobile phone with camera, add some software and take a 3D scan of any object. It is that easy and you don't need extra hardware - a problem for the Kinect?

A team at ETH Zurich has managed to get the necessary optical processing code to work out a 3D reconstruction from a set of 2D photos. All the user has to do is move the phone around the object and the software works out when to take pictures. It knows where it is by using the accelerometers to calculate its position relative to the first picture taken. The user also gets feedback on what parts of the object have been covered and can add views to, for example, see behind the object. 

Such computations previously needed a big computer, but now with optimization and the help of a GPU they can be done on a modest mobile phone. As no server is involved in the processing, the phone can be used to capture a 3D model even when not connected to the Internet. 



Although the ability to scan an object using a simple phone is a great tool, it isn't going to push out 3D cameras like the Kinect. Depth cameras provide a real time flow of depth information that can be used by robots and game players. The usefulness of this sort of software is in capturing isolated 3D models. The examples given in the video, i.e. scanning museum exhibits, sound great, but it isn't clear how well this would go down with curators who normally ban any sort of photography. Even so you can believe that there are reasons for wanting the app. 


What is more interesting is to speculate on where this single-sensor multiple view 3D reconstruction could go in the future. The reason that 3D cameras are useful in vision tasks is that the images they produce have far more redundancy. In a color image two close pixels of the same color aren't necessarily part of the same 3D object, but two close pixels at the same depth in a depth image are very likely to be part of the same 3D object. This makes implementing computer vision algorithms much easier. If you can extract 3D depth data from a single sensor moving to different points of view you can use the same depth algorithms to implement computer vision. It could be the breakthrough in computer vision we are looking for.

The big problem is that the ETH Zurich press release ends with the chilling line: 

The patent pending technology was developed exclusively by ETH Zurich and can run on a wide range of current smartphones.

This is a very clever technique but patentable? 



More Information

Your smartphone as a 3D scanner

Live Metric 3D Reconstruction on Mobile Phones pdf

Related Articles

 Megastereo - Panoramas With Depth

Imaging The Pulse

Google - We Have Ways Of Making You Smile

Computational Photography On A Chip

Super Seeing Software Ready To Download

Blink If You Don't Want To Miss it

Better 3D Meshes Using The Nash Embedding Theorem       

Light field camera - shoot first, focus later       


To be informed about new articles on I Programmer, install the I Programmer Toolbar, subscribe to the RSS feed, follow us on, Twitter, FacebookGoogle+ or Linkedin,  or sign up for our weekly newsletter.


raspberry pi books



or email your comment to:



Can C++ Be As Safe As Rust?

Herb Sutter is a well known and respected C++ champion and he thinks that the language only needs a few tweaks to make it as safe as Rust. Can this be true?

ZLUDA Ports CUDA Applications To AMD GPUs

ZLUDA is a translation layer that lets you run unmodified CUDA applications with near-native performance on AMD GPUs. But it is walking a fine line with regards to legality.

More News

Last Updated ( Monday, 09 December 2013 )