Technical Papers
Computational Sensing & Display
Tuesday, 12 August 3:45 PM - 5:15 PM | Vancouver Convention Centre, East Building, Exhibit Hall A Session Chair: Gordon Wetzstein, MIT Media Lab
Tuesday, 12 August 3:45 PM - 5:15 PM | Vancouver Convention Centre, East Building, Exhibit Hall A Session Chair: Gordon Wetzstein, MIT Media Lab
To generate graphics using levitated small objects, this method expands the acoustic manipulation method by changing the distribution of the acoustic-potential field. The approach makes available many expressions (for example, expression by materials and nondigital appearance), and the expressions are levitated in mid-air.
Yoichi Ochiai
The University of Tokyo
Takayuki Hoshi
Nagoya Institute of Technology
Jun Rekimoto
The University of Tokyo
This paper describes a technique for learning to estimate absolute, pixel-wise depth values from a conventional, monocular color camera, with minor hardware modifications. The algorithm converts any camera into a real-time depth camera, without the power consumption, bulk, and expense typical of current depth sensors.
Sean Ryan Fanello
Istituto Italiano di Tecnologia
Cem Keskin
Microsoft Research Cambridge
Shahram Izadi
Microsoft Research Cambridge
Pushmeet Kohli
Microsoft Research Cambridge
Jamie Shotton
Microsoft Research Cambridge
Antonio Criminisi
Microsoft Research Cambridge
David Kim
Microsoft Research Cambridge
David Sweeney
Microsoft Research Cambridge
Sing Bing Kang
Microsoft Research Redmond
This analysis of light propagation in an unknown scene using projectors and cameras that operate at transient timescales improves geometry reconstruction and light-in-flight imaging.
Matthew O'Toole
University of Toronto
Felix Heide
The University of British Columbia
Lei Xiao
The University of British Columbia
Matthias B. Hullin
Friedrich-Wilhelms-Universität Bonn
Wolfgang Heidrich
The University of British Columbia
Kiriakos N. Kutulakos
University of Toronto
A technique for achieving post-capture control of focus and aperture in a traditional camera by acquiring a carefully selected set of images (typically 5-10 images) and computationally reconstructing images corresponding to other focus and aperture settings.
Atsushi Ito
Sony Corporation
Salil Tambe
Rice University
Kaushik Mitra
Rice University
Aswin Sankaranarayanan
Carnegie Mellon University
Ashok Veeraraghavan
Rice University
A design for an optical see-through near-eye display that provides a wide field of view (110 degrees) and a compact form factor approaching eyeglasses. The approach uses tiled defocused point light sources coded through a transmissive spatial light modulator to project into the eye.
Andrew Maimone
University of North Carolina at Chapel Hill
Douglas Lanman
NVIDIA Research
Kishore Rathinavel
University of North Carolina at Chapel Hill
Kurtis Keller
University of North Carolina at Chapel Hill
David Luebke
NVIDIA Research
Henry Fuchs
University of North Carolina at Chapel Hill