Integral Photography was proposed by Gabriel Lippmann in 1908, well before the computer, holography, or even color photography as we know it. The method has greatly improved over the past 100 years, but has yet to become commercially viable. Now, with advances in digital capture and parallel image processing on the GPU, Lippmann’s methods are being revitalized. For the first time we have the computational power to assemble high quality images from integral photographs, and do it interactively, something that was impossible just a few years back! This computational power, and the availability of digital media to display dynamic and 3D content in real time lay the groundwork for the Plenoptic camera and the lightfield rendering techniques to become the 3D photography of the future.
In this presentation we will review the Plenoptic camera designs in versions 1.0 and 2.0. We will show mathematically and visually demonstrate the difference in resolution. Our presentation also introduces a number of algorithms, like Fourier Slice refocusing and the direct, advanced and refocused rendering and editing of 3D views from Plenoptic 2.0 radiance data. New developments, including superresolution, HDR and spectral color capture with the Plenoptic camera will be shown. As part of the presentation we will also take live pictures and render refocused 3D views in real time.