Computational Photography Moves Beyond The Camera
Written by David Conrad   
Sunday, 01 October 2017

Computational photography is the big second photographic revolution - the first being the move from film to digital. Now we have algorithms that can compose a photo after it has been taken and create effects that would require light to move in curved paths through lenses. We have moved far beyond the simple camera.

 

compzoom3

One of the marks of a professional photographer is the understanding of depth of field and perspective. By changing the focal length of the lens the entire feel of a photo can be altered. To the beginner it looks as if all that happens when you move from a wide angle, short focus, lens to a narrow angle, telephoto, lens is that you get closer to your subject and as a side issue the field of view changes. However, there is a much more subtle, and often more important effect, the perspective changes. This means the composition of the photo changes. If you use a wide angle lens then things in the background are small compared to the foreground. If you use a telephoto lens then there isn't so much difference and foreground and background are closer together. 

Until now it had always been a choice that you had to make at the time the photo was taken and you really had to pick either a short or a long focal length lens. Now researchers at UC Santa Barbara and NVIDIA have invented computational zoom. This allows the photographer to select the zoom after the photo has been taken and even blend together different zoom factors to create photos that couldn't be taken with a simple digital camera. 

The trick is a familiar one - take a stack of images, moving the camera closer to the scene between each shot and without changing the focal length of the lens. The stack of images is used to estimate depth and hence to construct a 3D model of the scene which can then be used to synthesize particular views, including some that couldn't have been taken with a traditional setup. 

 

compzoom1

 

compzoom2

 

To see what can be achieved take a look at the following video and concentrate on the relationship between the background and foreground:

 

The big problem with the current technique is the need to take multiple photos at different distances. An alternative approach using a single camera with multiple focal length lenses could make capturing a stack as easy as "click" - which is the one aspect of film photography that will never be replaced.

compzooom

 

More Information

Project Web Site

Computational Zoom: A Framework for Post-Capture Image Composition by Abhishek Badki, University of California, Santa Barbara and NVIDIA; Orazio Gallo, NVIDIA;  Jan Kautz, NVIDIA, Pradeep Sen, University of California, Santa Barbara

Related Articles

 Google Implements AI Landscape Photographer

Using Super Seeing To See Photosynthesis

Selfie Righteous - Corrects Selfie Perspective

Computational Photography Shows Hi-Res Mars

 

To be informed about new articles on I Programmer, sign up for our weekly newsletter, subscribe to the RSS feed and follow us on Twitter, Facebook or Linkedin.

 

Banner


Improved Code Completion With JetBrains Mellum
29/10/2024

JetBrains has launched Mellum, a proprietary large language model specifically built for coding. Currently available only with JetBrains AI Assistant, Mellum is claimed to provide faster, sm [ ... ]



Check Your APIs With Zuplo's Rate My OpenAPI
15/10/2024

Zuplo has launched a new suite of tools that rates the quality of your API, based on its OpenAPI specification. We put it through its paces and find it useful.


More News

espbook

 

Comments




or email your comment to: comments@i-programmer.info

Last Updated ( Sunday, 01 October 2017 )