About Me

I work on computer vision and machine learning problems, with a focus on applications in photography. My work at Google is the basis for Portrait Mode and Magic Eraser on the Google Pixel smartphone. Previously, I completed a Ph. D under the supervision of Bill Freeman at the Computer Science and Artificial Intelligence Laboratory (CSAIL) at MIT. I worked on ways to visualize tiny changes in images and videos by amplifying them.


News


October 2023 RealFill featured on Two Minute Papers!
May 2023 Named an Outstanding Reviewer for CVPR 2023
October 2021 Launched Magic Eraser on the Pixel 6.
December 2019 New blog post about Pixel 4's Portrait Mode.
September 2019 Papers on depth estimation at ICCV 2019 and camera synchronization at ICCP 2019.
November 2018 New blog post about Pixel 3's Portrait Mode.
June 2018 New paper at SIGGRAPH 2018 about the Google Pixel 2's Portrait Mode.
October 2017 New Paper in Proceedings of the National Academy of Sciences: Motion microscopy for visualizing and quantifying small motions.
October 2016 I am starting a job at Google Research in Marc Levoy's group.
January 2016 I graduated and submitted my Ph. D dissertation: Revealing and Analyzing Imperceptible Deviations in Images and Videos.
December 2015 I gave an invited talk at the Computer Vision for Affordable and Accessible Healthcare Workshop at the 2015 International Conference in Computer Vision (ICCV).
October 2015 I have a new paper at SIGGRAPH Asia titled Deviation Magnification: Revealing Departures from Ideal Geometries.
May 2015 I gave an invited talk at the National Board for Medical Examiners in Philadelphia, Pennsylvania.
June 2014 Awarded Best Demo at CVPR 2014 for Real-time Video Magnification.

Open Source Software

Phase Based Video Motion Processing code
Matlab code for implementing the new and improved phase-based motion magnification pipeline.

Motion Microscopy for Visualizing and Quantifying Small Motions code
Matlab code to quantify small motions in videos.

Dual-Pixel Capture App
Sample Android app showing how to capture dual-pixel data on Google Pixel phones.

Generative AI Publications

RealFill: Reference-Driven Generation for Authentic Image Completion

Luming Tang, Nataniel Ruiz, Qinghao Chu, Yuanzhen Li, Aleksander Holynski, David E. Jacobs, Bharath Hariharan, Yael Pritch, Neal Wadhwa, Kfir Aberman, Michael Rubinstein

arXiv, 2023.
[Paper] [Website] [Two Minute Papers]
Diffusion inpainting models hallucinate a plausible result when uncropping an image. In contrast, RealFill fills the image with what should have been there.
HyperDreamBooth: HyperNetworks for Fast Personalization of Text-to-Image Models

Nataniel Ruiz, Yuanzhen Li, Varun Jampani, Wei Wei, Tingbo Hou, Yael Pritch, Neal Wadhwa, Michael Rubinstein, Kfir Aberman

arXiv, 2023.
[Paper] [Website]
HyperDreamBooth is able to personalize a text-to-image diffusion model 25x faster than DreamBooth, by using a hyper network and fast finetuning.

Mobile Photography Publications

Defocus Map Estimation and Deblurring from a Single Dual-Pixel Image

Shumian Xin, Neal Wadhwa, Tianfan Xue, Jonathan T. Barron, Pratul Srinivasan, Jiawen Chen, Ioannis Gkioulekas, Rahul Garg

International Conference on Computer Vision (ICCV), 2021.
[Paper] [Website]
Multiplane images can be used to simultaneously deblur dual-pixel images, despite variable defocus due to depth variation in the scene.
Du2Net: Learning Depth Estimation from Dual-Cameras and Dual-Pixels

Yinda Zhang, Neal Wadhwa, Sergio Orts-Escolano, Christian Häne, Sean Fanello, Rahul Garg

European Conference on Computer Vision (ECCV), 2020.
[Paper]
Using both dual-cameras and dual-pixels for depth estimation yields superior depth to using either alone.
Learning to Autofocus

Charles Herrmann, Richard Strong Bowen, Neal Wadhwa, Rahul Garg, Qiurui He, Jonathan T. Barron, Ramin Zabih

Computer Vision and Pattern Recognition (CVPR), 2020.
[Paper] [Website]
A learning based approach to autofocus beats several learned and non-learned baselines.
Learning Single Camera Depth Estimation using Dual-Pixels

Rahul Garg, Neal Wadhwa, Sameer Ansari, Jonathan T. Barron

International Conference on Computer Vision (ICCV), 2019.
[Paper] [Github]
Taking into account the optics of dual-pixel cameras can improve monocular depth estimation.
Wireless Software Synchronization of Multiple Distributed Cameras

Sameer Ansari, Neal Wadhwa, Rahul Garg, Jiawen Chen

International Conference on Computational Photography (ICCP), 2019.
[Paper]
Smartphone cameras can be wirelessly synchronized to capture images within 250 microseconds of each other without extra hardware.
Synthetic Depth-of-Field with a Single-Camera Mobile Phone

Neal Wadhwa, Rahul Garg, David E. Jacobs, Bryan E. Feldman, Nori Kanazawa, Robert Carroll, Yair Movshovitz-Attias, Jonathan T. Barron, Yael Pritch, Marc Levoy

ACM Transactions on Graphics, Volume 37, Number 4 (Proc. SIGGRAPH), 2018.
[Paper] [Supplemental] [Blog Post]
Depth from dual-pixels and semantic segmentation can be used to produce shallow depth-of-field images. Basis for Portrait Mode on Google Pixel 2.
Aperture Supervision for Monocular Depth Estimation

Pratul P. Srinivasan, Rahul Garg, Neal Wadhwa, Ren Ng, Jonathan T. Barron

Proc. of Computer Vision and Pattern Recognition (CVPR), 2018.
[Paper]
Varying a camera's aperture provides a supervisory signal that can teach a neural network to do monocular depth estimation.
Bilateral Guided Upsampling

Jiawen Chen, Andrew Adams, Neal Wadhwa, Samuel W. Hasinoff

ACM Transactions on Graphics, Volume 35, Number 6 (Proc. SIGGRAPH Asia), 2016.
[Paper] [Code] [Video]
Black-box image processing operators can be accelerated by fitting a fast bilateral-space model.

Motion Magnification and Analysis Publications

Prospective Validation of Smartphone-based Heart Rate and Respiratory Rate Measurement Algorithms

Sean Bae, Silviu Borac, Yunus Emre, Jonathan Wang, Jiang Wu, Mehr Kashyap, Si-Hyuck Kang, Liwen Chen, Melissa Moran, John Cannon, Eric S. Teasley, Allen Chai, Yun Liu, Neal Wadhwa, Mike Krainin, Michael Rubinstein, Alejandra Maciel, Michael V. McConnell, Shwetak Patel, Greg S. Corrado, James A. Taylor, Jiening Zhan, Ming Jack Po

Nature Communications Medicine, 2022.
[Paper]
Smartphone camera-based techniques can accurately measure heart rate and respiration rate across groups with different skin tones and respiratory health status.
Motion Microscopy for Visualizing and Quantifying Small Motions

Neal Wadhwa, Justin G. Chen, Jonathan B. Sellon, Donglai Wei, Michael Rubinstein, Roozbeh Ghaffari, Dennis M. Freeman, Oral Buyukozturk, Pai Wang, Sijie Sun, Sung Hoon Kang, Katia Bertoldi, Frédo Durand, and William T. Freeman

Proceedings of the National Academy of Sciences, October 2017.
[Paper] [Webpage] [Code]
Motion magnification and analysis reveal phenomena in three different scientific domains.
Eulerian Video Magnification and Analysis

Neal Wadhwa, Hao-Yu Wu, Abe Davis, Michael Rubinstein, Eugene Shih, Gautham J. Mysore, Justin G. Chen, Oral Buyukozturk, John V. Guttag, William T. Freeman, Frédo Durand

Communications of the ACM, Volume 60, Number 1, January 2017.
[Paper]
An expository paper summarizing the various motion magnification and subtle motion analysis techniques we have developed.
Modal Identification of Simple Structures with High-Speed Video using Motion Magnification

Justin G. Chen, Neal Wadhwa, Young-Jin Cha, Frédo Durand, William T. Freeman, Oral Buyukozturk

Journal of Sound and Vibration Volume 345, 2015.
[Paper]
Motion magnification can be used to identify the modal shapes of simple structures such as pipes and beams.
Deviation Magnification: Revealing Departures from Ideal Geometries

Neal Wadhwa, Tali Dekel, Donglai Wei, Frédo Durand, William T. Freeman

ACM Transactions on Graphics, Volume 34, Number 6 (Proc. SIGGRAPH Asia), 2015.
[Paper] [Webpage]
A method to reveal subtle deviations of objects from ideal geometries, like lines or cirlces, in a single image by magnifying them.
Refraction Wiggles for Measuring Fluid Depth and Velocity from Video

Tianfan Xue, Michael Rubinstein, Neal Wadhwa, Anat Levin, Frédo Durand, William T. Freeman

Proc. of European Conference on Computer Vision (ECCV), 2014.
[Paper] [Webpage]
A method to recover the velocity and depth of hot air from video.
The Visual Microphone: Passive Recovery of Sound from Video

Abe Davis, Michael Rubinstein, Neal Wadhwa, Gautham Mysore, Frédo Durand, William T. Freeman

ACM Transactions on Graphics, Volume 33, Number 4 (Proc. SIGGRAPH), 2014.
[Paper] [Webpage]
A technique to recover sound from videos of objects subtly vibrating in response to sound.
Structural Modal Identification through High Speed Camera Video: Motion Magnification

Justin G. Chen, Neal Wadhwa, Young-Jin Cha, Frédo Durand, William T. Freeman, Oral Buyukozturk

Proceedings of the 32nd International Modal Analysis Conference (2014)
[Paper]
A validation that the motion magnified motions are indeed real and a way to compute the mode shapes of a cantilevered beam from video.
Riesz Pyramids for Fast Phase-Based Video Magnification

Neal Wadhwa, Michael Rubinstein, Frédo Durand, William T. Freeman

Computational Photography (ICCP), 2014 IEEE International Conference on
[Paper] [Webpage] [Pseudocode]
Provides the quality of the previous phase-based technique with the real-time speed of the original linear technique.
Phase-based Video Motion Processing

Neal Wadhwa, Michael Rubinstein, Frédo Durand, William T. Freeman

ACM Transactions on Graphics, Volume 32, Number 4 (Proc. SIGGRAPH), 2013.
[Paper] [Webpage] [BibTeX]
A new technique to amplify small motions that solves the noise amplification and intensity clipping artifacts of the previous linear method by manipulating the phase in sub-bands of videos.
Revealing Invisible Changes In The World

Michael Rubinstein, Neal Wadhwa, Frédo Durand, William T. Freeman

Science Vol. 339 No. 6119, Feb 1 2013
[Article in Science] [Video] [NSF SciVis 2012] [BibTeX]
An expository video showcasing our results and explaining our technique.