Remaining course logistics
1
- Gradescope submission site available.
- Gradescope submission site available.
- Tuesday 12/22, 1 – 2 pm, Grace Kuo (lensless cameras, diffuserCam, AR/VR displays).
https://docs.google.com/spreadsheets/d/1CVg7nUbI701pvZFPX3BR0uzKB76Y6PEl3tXmq1UF4AU/edit#gid=0
Class evaluation*s* – please take them!
2
Using incident lightfield for simulating GelSight
Arpit Agarwal
Objective
Capture spatially and directionally varying illumination and use this lighting to generate images with synthetic objects
Key Principle
Debevec, Paul. "Rendering synthetic objects into real scenes: Bridging traditional and image-based graphics with global illumination and high dynamic range photography." ACM SIGGRAPH 2008 classes. 2008. 1-10.
Capture
Render
Spatially varying incident lightfield
Unger, Jonas, et al. Capturing and rendering with incident light fields. UNIVERSITY OF SOUTHERN CALIFORNIA MARINA DEL REY CA INST FOR CREATIVE TECHNOLOGIES, 2003.
Motivation
2cm
2cm
Capture Method
GelSight
Capture Method
GelSight
Capture Method
GelSight
Rendering Method
Diffuse material
Precomputed
Rendering Method
Diffuse material
Scaled to [0,1]
Simulation setup
Simulation Results
GelSight data collection
GelSight Results
Summary
18
Deep High Dynamic Ranging of Dynamic Scenes
Presenter: Uma Arunachalam
Andrew ID: uarunach
Brief Overview
19
Approach
3 misaligned images
Optical Flow
3 misaligned images
CeLiu’s Classical Method
Approach
3 misaligned images
Learning based LDR merging
3 misaligned images
Network Architecture
Direct
Weight estimator
Fully Differentiable Architecture
Experimental Setup
Results: Error metric
Direct
WE
Results: Optical Flow
Results: Deep merging vs Tent
Results: WE vs Direct
Results
Metric | Tent merging | Deep merging |
Mean Square error | 0.0594 | 0.0037 |
PSNR | 12.27 | 24.31 |
Run time (merging)* | 9.58 s | 5.88 s |
* not profiled, averaged over 2 runs
Main References
Thank you!
Questions?
Real-time Cartoonization
Ricky Bao
What is NPR and cartoonization?
Example of cartoonization
How is cartoonization performed?
Real-time cartoonization
| Average (ms) | Min (ms) | Max (ms) |
Roof | 56.2 | 50.3 | 62.7 |
Android | 54.9 | 47.2 | 60.1 |
Pool | 58.0 | 50.6 | 66.9 |
Globe | 53.2 | 46.3 | 58.2 |
Sunset | 54.7 | 49.8 | 61.0 |
Resources
Multi-Flash Imaging for Depth Edge Detection and Photometric Stereo
Matthew Baron (mcbaron)
Setup
Depth Edge Detection
Photometric Stereo
References
[1] T. Papadhimitri and P. Favaro, “A New Perspective on Uncalibrated Photometric Stereo,” in 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, Jun. 2013, pp. 1474–1481, doi: 10.1109/CVPR.2013.194.
[2] R. Raskar, K.-H. Tan, R. Feris, J. Yu, and M. Turk, “Non-Photorealistic Camera: Depth Edge Detection and Stylized Rendering Using Multi-Flash Imaging,” p. 12.
[3] Y. Taguchi, “Rainbow Flash Camera: Depth Edge Extraction Using Complementary Colors,” p. 16.
[4] L. Wu, A. Ganesh, B. Shi, Y. Matsushita, Y. Wang, and Y. Ma, “Robust Photometric Stereo via Low-Rank Matrix Completion and Recovery,” in Computer Vision – ACCV 2010, vol. 6494, R. Kimmel, R. Klette, and A. Sugimoto, Eds. Berlin, Heidelberg: Springer Berlin Heidelberg, 2011, pp. 703–717.
[5] Y. Quéau, F. Lauze, and J.-D. Durou, “Solving Uncalibrated Photometric Stereo Using Total Variation,” J Math Imaging Vis, vol. 52, no. 1, pp. 87–107, May 2015, doi: 10.1007/s10851-014-0512-5.
Tracking Microscopic Motion Using Speckle Imaging
Yash Belhe
Setup
Optically Rough Surface
Laser
Bare Camera Sensor
Approx Co-located with Laser
Example with horizontal motion (with stutters)
Tracking Results
Learning to predict Depth and Autofocus using Focal and Aperture stack
Akankshya Kar and Anand Bhoraskar
{akankshk,abhorask}@andrew.cmu.edu
Contents
Problem Statement
Low Light Image
Bright Autofocus Image
Depth Map
AFI Image
Related Work
Learning to Autofocus1
Focal Stack
CNN
F
Auto-Focused Image
F
Datasets
DDFF7
Flowers Dataset8
Learning to Autofocus dataset1
Proposed Method
CNN
(Dark to light)
AFI
CNN
All in focus
Depth
Refocusing
CNN
F, A
Selected Patch
GT generation
Project scope
Ground Truth generation for Autofocus
f_min
f_max
Subset Focal stack
F (median depth)
Merging to get partial focused image
SSIM
(Grid search)
A, F(median)
Selected Patch
Selected subset using focus to disparity
AFI
Partial Focused
Results
GT generation
# of Patches: 19,311
Patches Size : 64 x 64
Stride : 100
Path | X_begin | Y_begin | Aperture size | Focus Distance |
cafeteria/LF_0001.npy | 32 | 96 | 7 | 3 |
cafeteria/LF_0001.npy | 32 | 128 | 7 | 3 |
cafeteria/LF_0001.npy | 32 | 160 | 7 | 3 |
cafeteria/LF_0001.npy | 32 | 192 | 7 | 3 |
cafeteria/LF_0001.npy | 64 | 32 | 7 | 3 |
cafeteria/LF_0001.npy | 64 | 64 | 7 | 3 |
cafeteria/LF_0001.npy | 64 | 96 | 7 | 3 |
cafeteria/LF_0001.npy | 64 | 128 | 7 | 3 |
cafeteria/LF_0001.npy | 64 | 160 | 7 | 2 |
cafeteria/LF_0001.npy | 64 | 192 | 7 | 2 |
cafeteria/LF_0001.npy | 96 | 32 | 7 | 2 |
cafeteria/LF_0001.npy | 96 | 64 | 7 | 2 |
cafeteria/LF_0001.npy | 96 | 96 | 7 | 2 |
cafeteria/LF_0001.npy | 96 | 128 | 7 | 2 |
Selected Patch in Full image
Depth Patch and cropped out patch
Depth Map Generation
AFI
Depth Map
CNN
Virtual Normal Loss2
Results
Depth from confocal stack
Ground Truth
Predicted
Auto-Focus
Refocusing
CNN
F, A
Selected Patch
Ordinal Regression Loss3,4
Future Work
References
Fast Separation of Direct/Global Images In A Pocket
Presenter: Zili Chai (zilic)
Hardware Setup
Sony IMX219
RPi 4B
DLPDLCR2000EVM
Camera Calibration
10 bit bayer data, BGGR ordering
AAAAAAAA BBBBBBBB CCCCCCCC DDDDDDDD AABBCCDD
Linearity ∝ exposure
ISO = 100
Dark Frame
Mean = 63.90 (on 100 dark images)
daylight result
1ms
100ms
Projector also needs time!
Global/Direct Images
Global Illumination
scattering, interreflection, shadows…
Separation In Real World
Global/Direct Images
Separation in Real World
Results
Checkerboard shift
max
min
direct
global
direct
global
Results
More Patterns
random pattern with phase shift
0.5+0.5·sinφ
+2pi/7, 4pi/7, 6pi/7, 8pi/7, 10pi/7,12pi/7
direct
global
Results
More Patterns
gray code
References
Sony, IMX219 product brief version 1.0, (Jan 2017).
Texas Instruments, TI DLP® LightCrafter™ Display 2000 EVM User's Guide, (Oct 2017).
Pagnutti, Mary A., et al. "Laying the foundation to use Raspberry Pi 3 V2 camera module imagery for scientific and engineering purposes." Journal of Electronic Imaging 26.1 (2017): 013014.
Nayar, Shree K., et al. "Fast separation of direct and global components of a scene using high frequency illumination." ACM SIGGRAPH 2006 Papers. 2006. 935-944.
Q&A
Kirchhoff Migration for NLOS Imaging
Dorian Chan
NLOS Imaging
Figure from:
David B. Lindell, Gordon Wetzstein, and Matthew O’Toole. 2019. Wave-based non-line-of-sight Imaging using fast f−k migration. ACM Trans. Graph. 38, 4, 116.
Traditional Approach: Backprojection
F-K Migration: a Wave-Based Theory
Fourier Transform
Resample and Filter
Inverse Fourier Transform
Supposedly more BRDF robust!
Kirchhoff Migration: another idea from seismology
Second Order Kirchhoff Migration
Relating F-K migration to volumetric albedo models
Questions?
Colorization through Optimization
Shruti Chidambaram | 15-463, Fall 2020
Colorization: Background
Acknowledgement
My project reimplements the algorithm presented in Colorization using Optimization
(Anat Levin, Dani Lischinski, and Yair Weiss, 2004)
Algorithm
Results
MARKED-UP GRAYSCALE
COLORIZED
GROUND TRUTH
Results: Experimenting with Varied User Input
MARKED-UP GRAYSCALE
COLORIZED
GROUND TRUTH
Experimenting with a Deep Learning Approach
Autocolorization results from demos.algorithmia.com/colorize-photos
paper: Colorful Image Colorization (Zhang, Isola, Efros; 2016)
OUTPUT
INPUT
GROUND
TRUTH
Non-Realistic Rendering App
Han Deng (handeng)
Heshan Liu (heshanl)
Motivation
Nowadays there are a lot of different types of filters you can choose when you are using camera app. We would like to make something different and slightly more complicated rather than just adjust the color of the image.
Original
Blending
Sketch
Watercolor
Goal
Environment Setup
Algorithms: Pencil Sketch
Input Image
Output 1
Output 2
Algorithms: WaterColor
Input Image
Output Image
Algorithms: Image Blending - Poisson Blending
Pérez, P., Gangnet, M., & Blake, A. (2003). Poisson image editing. In ACM SIGGRAPH 2003 Papers (pp. 313-318).
Algorithm and app Demo: Image blending
Possible Improvement
Thank you!
Uncertainty in Radiometric Calibration
15663 Course Project
Advait Gadhikar
Goal
Motivation
Standard Tone Map vs Learnt Tone Map
f is a polynomial and v is a linear transform, with g being a correction factor
If tone mapping is deterministic, Debevic and Malik’s method assumes this mapping
RMSE for Rendering Function (Raw to JPEG)
Sample Images in stack
Inverse Distribution for Linearization
HDR with Uncertainty
Estimated HDR image with uncertainty
Thank You
Questions!?
References
One-Shot to Vertigo: Novel View Synthesis using Light Field Cameras
Rohan Rao
(rgrao@andrew.cmu.edu)
Introduction
Single-camera Pipeline: Pre-processing
(from [1])
Single-camera Pipeline: Digital Zoom
(from [1])
Single-camera Pipeline: View Synthesis
(from [1])
Single-camera Pipeline: Image/Depth fusion
(from [1])
Single-camera Pipeline: Depth Hole Filling
(from [1])
Single-camera Pipeline: Image Inpainting
(from [1])
Single-camera Pipeline: Shallow Depth of Field (SDoF)
(from [1])
Intermediate steps towards final results
Post digital zoom (I1)
Without digital zoom (I2)
I1DZ
I2DZ
I1DZ
Final result
Extensions of prior work
Ideas for future work
Thanks for listening! Questions?
References
Epipolar scanning for shape-from-silhouette in scattering medium
Shirsendu S Halder
shirsenh
Shape from silhouette
Shape from silhouette
Collimated illumination
Orthographic camera
Object
Imaging through scattering medium
We want the transmissive paths (ballistic photons) as they sharpen the object image and also enhances contrast.
Imaging through scattering medium
Rows of an orthographic projector lit.
Camera captures ballistic paths from each row.
Combination of position cues, angle cues, and probing
Orthographic projector
Telecentric camera
Imaging through scattering medium
Rows of an orthographic projector lit sequentially.
Camera captures ballistic paths from each row.
Orthographic projector
Telecentric camera
Why telecentricity?
Image credit: https://www.edmundoptics.com/knowledge-center/application-notes/imaging/advantages-of-telecentricity/
Setup
Setup
Syncing electronics
Setup
Laser line
Object used for scanning
Scanning (in air)
Global
Scanning (in air)
Epipolar
Scanning (in water)
Global
Scanning (in water)
Epipolar
Scanning (milk + water - low concentration)
Global
Scanning (milk + water - low concentration)
Epipolar
Scanning (milk + water - medium concentration)
Global
Scanning (milk + water - medium concentration)
Epipolar
Scanning (milk + water - high concentration)
Global
Scanning (milk + water - high concentration)
Epipolar
Thank you
Questions?
HDRnet and Artistic Style Transfer for HDR replication
Brandon Hung
HDRNet
Artistic Style Transfer
HDRnet for Tonemapping
Input Output Learned
Style Transfer to Preserve Content
Combining the two?
Combining the two?
Unstructured Lightfields
Akshath Jain and Deepayan Patra
Inspiration
Assignment 4 - Planar Unstructured Lightfields
155
Technical Background
Goal: Generating arbitrary viewpoint rendering given unstructured, non-planar, input
Proposed Solution: Interpolate on a triangular mesh of input viewpoints with appropriate depth information to generate new image
156
Steps
Generate Input Poses
1
2
3
Triangulate Viewpoints
Rendering
157
Generating Poses
COLMAP estimates camera poses and a 3D reconstruction of the scene using structure from motion. This is broken down into:
158
Triangulating Viewpoints
From the new projected camera origins, we develop a Delaunay Triangulation of the viewpoints.
Each point within the input space has exactly three neighboring viewpoints to interpolate.
159
Rendering
Generate each novel viewpoint’s image as:
160
Results
Unstructured Images
Output
Credits
Our implementation was inspired by the Unstructured Light Fields paper by Davis et. al. We also relied on work done by Mildenhall et. al in LLFF to get camera poses from COLMAP and to identify ideal new sample viewpoints. Our full list of sources is as follows:
Davis et al. “Unstructured Light Fields.” Eurographics 2012.
Mildenhall et. al. “Local Light Field Fusion: Practical View Synthesis with Prescriptive Sampling Guidelines.” SIGGRAPH 2019.
Levoy, Marc. “Light Fields and Computational Imaging.” IEEE 2006.
Buehler et. al. “Unstructured Lumigraph Rendering.” SIGGRAPH 2001.
Isaksen et. al. “Dynamically Reparameterized Light Fields.” SIGGRAPH 2000.
Schönberger, Johannes Lutz and Frahm, Jan-Michael. “Structure-from-Motion Revisited.” CVPR 2016.
Pollefeys, Marc. “Visual 3D Modeling from Images” 2002.
164
15663: Computational Photography
Fall 2020
Using Spatio-Temporal Radiance Variations to Look Around Corners
Varun Jain
varunjai@andrew.cmu.edu
Idea[1]
Figure 1a: A and B represent two people hidden from the camera’s view by a wall.
Idea[1]
167
Figure 2: (top) the hidden scene with 2 actors, (bottom) penumbra as seen by the naked eye, as seen by the camera and the groundtruth trajectories over time.
Contributions
Results [1 person]
Fig: Observed video
Fig: 1-D angular projections of the hidden scene
Results [2 people]
Fig: Observed video
Fig: 1-D angular projections of the hidden scene
Comparative Results
Method for Background Subtraction | Error wrt Official Code (MSE) |
First Frame | 7.41 |
Mean Frame | 2.94 |
Windowed Average Frame | 20.16 |
Limitations
Contributions
Dataset and Methodology
Figure 3: (left) generated dataset, and, (right) model architecture.
References
175
Evaluating Confocal Stereo
Kyle Jannak-Huang
Relative Exitance Estimation
Aperture f3.5
Focus dist ~ 2cm
Aperture f7.1
Focus dist ~ 2cm
Relative Exitance Estimation
Aperture f3.5
Focus dist ~ 2cm
Aperture f3.5
Focus dist ~ 90cm
Image Alignment
Depthmap from Aperture-Focal Images
Depthmap from Aperture-Focal Images
Depthmap from Aperture-Focal Images
Fuzz dataset images
Denoised result
Discussion
Gradient-Domain Path Tracing
Zhi Jing (Zoltan), Ran Zhang (Ryan)
Gradient-Domain Path Tracing
Gradient-Domain Path Tracing
Standard Monte Carlo rendering
Gradient-Domain Path Tracing
Render result - primal image
Sample gradients
The most naive approach
Gradient image - naive approach
dx
dy
Reconstruction result - naive approach
primal
final
Reconnection
Gradient image - with reconnection
dx
dy
Gradient image - naive approach
dx
dy
Reconstruction result - with reconnection
primal
final
Paper’s approach - add MIS
dx
dy
Paper’s approach - add MIS
primal
final
Gradient-Domain Path Tracing
Image Reconstruction
Find an image that best fits the estimated primal and gradients
Primal
Gx
Gy
Poisson Reconstruction
L2-norm reconstruction
L1-norm reconstruction
Gradient term
Primal term
Poisson Reconstruction
A x = b
Solve the least squares minimization problem using Conjugate Gradient Descent
Naïve Implementation using CGD
Naïve Implementation using CGD
L2 norm
L1 norm
Re-implement and Experiment (L1 norm vs. L2 norm)
L2 Norm
L1 Norm
Experiment (L1 norm vs. L2 norm)
L2 Norm
L1 Norm
Unbiased but has some artifacts
Biased but has less artifacts
Experiment (different α values)
α = 0.2
α = 1.0
α = 10
Thank you!
References
Kettunen, Markus, et al. "Gradient-domain path tracing." ACM Transactions on Graphics (TOG) 34.4 (2015): 1-13.
Lehtinen, Jaakko, et al. "Gradient-domain metropolis light transport." ACM Transactions on Graphics (TOG) 32.4 (2013): 1-12.
Manzi, Marco, et al. "Gradient-Domain Bidirectional Path Tracing." EGSR (EI&I). 2015.
Pérez, Patrick, Michel Gangnet, and Andrew Blake. "Poisson image editing." ACM SIGGRAPH 2003 Papers. 2003. 313-318.
Manzi, Marco, Delio Vicini, and Matthias Zwicker. "Regularizing Image Reconstruction for Gradient‐Domain Rendering with Feature Patches." Computer graphics forum. Vol. 35. No. 2. 2016.
Ji, Hao, and Yaohang Li. "Block conjugate gradient algorithms for least squares problems." Journal of Computational and Applied Mathematics 317 (2017): 203-217.
View Synthesis using Neural Radiance Fields
Rohan Joshi
Problem Definition:
Given: Images of a scene from known camera poses,
Task: Render new views of the scene
Method: Scene Representation as Neural Radiance Fields
Algorithm for Volume Rendering
Forward Pass
Step 1: Generate Camera Rays
Step 2: Sample points along each ray (affects the resolution of output image )
Step 3: Get RGB and σ (Volume Density at the point) value
Algorithm for Volume Rendering
Error Calculation and Backpropagation
Step 1: Composite the pixel color along the ray to get the rgb image
Step 2: Minimise the mse for known images
Step 3: Backprop ...
W(σi)
Results: Performed on Synthetic Data
�
Training Set
Result: Rendered Views
References
Thank you
Color-Filtered Aperture For Image Depth & Segmentation
Leron Julian
Depth Estimation Techniques
Stereo
Depth From Disparity Using Color Filter (This Project)
Depth From Defocus
Constructing Depth Using Color Misalignment
How? (Red & Green Example)
Background
Camera Sensor
Object In Focus
Color Aperture
Sample Images
Depth Estimation
Depth Estimation
RGB Color Model
d = 1
d = 3
d = 5
Depth Map
Captured Images
Local Estimates
Estimate with Graph-Cuts
Trimap
Foreground
Unknown
Background
Matting
Matte Optimization Flow
Applications of Matting
References
Depth from Defocus in the Wild
Alice Lai, Adriana Martinez
Goal: Two-frame depth from defocus using tiny blur condition
Original Image
Image w/Blur
Idea: Combine local depth/flow estimation w/ �spline-based scene understanding
Idea: Combine local depth/flow estimation w/ �spline-based scene understanding
Local Depth/Flow Estimation (Bottom-up Likelihood)
Defocus Equalization Filters
Minimize sum-of-squared difference between two equalized images!
Ensures brightness constancy and equalizes the variances between the two images
Local Likelihood Q
Local Depth/Flow Estimation (Bottom-up Likelihood)
Local Prior Lqp
Local smoothness prior due to overlapping patches
Local Likelihood Q
Fit quadratic function to make it analytical for global optimization
Local Depth/Flow Estimation (Bottom-up Likelihood)
Qp evaluated at every pixel q for some (d,v) patch pixel p estimate (p denotes center pixel)
M x M array of values that encodes smoothness between every pixel w.r.t. every pixel
Minimize over LDFD loss to optimize (d,v) patch estimates using Markov Random Field (MRF)
Idea: Combine local depth/flow estimation w/ �spline-based scene understanding
Global DFD (Top-Down Likelihood)
Inputs:
Control Points, �Feature Map, �Patch Likelihoods
Update weight vectors
Update occlusion map
Update depth planes, 2D affine transformations, segment labels of control points
Update control points’ feature vectors
Output:
Pixel-based depth and flow
Patch-based depth and flow
Scene segmentation
Spline parameters
Control Point Cn: �- depth plane D�- 2D affine transformation U, V
Pixel q (including control points):
- weight vector
- feature vector
- segment label
Global DFD (Top-Down Likelihood)
Control Points
Scene Segmentation
Results
Arbitrary Lqp smoothness depth map
Ground truth depth
Original input
Patch-based Lqp smoothness depth map
Results
Arbitrary Lqp smoothness depth map
Ground truth depth
Original input
Patch-based Lqp smoothness depth map
Acknowledgements
Dr. Ioannis Gkioulekas for the tireless support of our project through many OH sessions and Slack messages
Dr. Kiriakos Kutulakos for sharing his research group’s dataset and contact information of first authors
References:
Huixuan Tang, Scott Cohen, Brian Price, Stephen Schiller and Kiriakos N. Kutulakos, Depth from defocus in the wild. Proc. IEEE Computer Vision and Pattern Recognition Conference, 2017. https://www.dgp.toronto.edu/WildDFD/
Questions?
A Reconstruction Framework for Time Series Thermal Traces using Mixed Stereo Photography
Arjun Lakshmipathy
Computational Photography (15-862)
Fall 2020
Context
Larger Thrust: Automated Design of Custom, Low-Cost Dextrous Hands from Demonstration
Principal Reference
The Approach
The Problems
3 problems to solve:
Materials and Setup
•1 Intel RealSense D415
•1 Flir C5
•2 tripods + 1 clamp mount
•?? Hand warmers
•1 dressed up turntable
•1 calibration target
•1 house lab space
Turntable and second tripod taken borrowed from CMU Motion Capture Lab
Process Step 1: Calibration
•Thermal intrinsics
•Stereo Calibration
•~ 10 images apiece
Process Step 2: Capture
Process Step 3: Object Pose Estimation
•Entirely in Depth CCS
•Load 3D model, convert to point cloud
•Segment scene to extract object
•Solve R and T using Iterative Closest Point Method (ICP) [2]
Process Step 4*: Alignment
Process Step 5*: Texture Mapping via Color Map Optimization [3]
Concluding Remarks
•Process: Calibration -> Capture -> Pose Estimation -> Alignment -> Texture Mapping
•Works reasonably…in principle. Still some items to resolve for full operation
•Next steps: work through bugs and nuances with each step, collect initial and final grasps for multiple objects from multiple subjects. Use as basis for next research thrust.
*Synthesized from dataset captures only / not my own yet
Thanks for listening! Questions?
References
[1] S. Brahmbhatt, C. Ham, C. C. Kemp, and J. Hays, “Contactdb: Analyzing and predicting grasp contact via thermal imaging,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8709–8719, 2019. �
[2] PJ Besl and Neil D McKay. A method for registration of 3- d shapes. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 14(2):239–256, 1992
[3] Qian-Yi Zhou and Vladlen Koltun. Color map optimization for 3d reconstruction with consumer depth cameras. ACM Transactions on Graphics (TOG), 33(4):155, 2014.
HDR Image Reconstruction using Hallucinated exposure stack
Shamit Lal(shamitl), Sajal Maheshwari(sajalm)
Problem
Traditional approach
Deep learning based methods
What if ?
Our approach
Naive encoder-decoder
Mapping based approach
Mapping based approach
0 | 0.2 | 0.4 | 0.7 | . | . | . | 1 |
Mapping based approach - Problem!
Mapping based approach - Results
Modified encoder-decoder
Modified encoder-decoder
Merging HDR
Unsupervised methods ?
References
1.Single-Image HDR Reconstruction by Learning to Reverse the Camera Pipeline, Yu-Lun Liu, Wei-Sheng Lai, Yu-Sheng Chen, Yi-Lung Kao, Ming-Hsuan Yang, Yung-Yu Chuang, Jia-Bin Huang (https://arxiv.org/abs/2004.01179)
2.Deep Reverse Tone Mapping, Yuki Endo, Yoshihiro Kanamori, and Jun Mitani(http://www.cgg.cs.tsukuba.ac.jp/~endo/projects/DrTMO/)
3.HDR image reconstruction from a single exposure using deep CNNs,Gabriel Eilertsen, Joel Kronander, Gyorgy Denes, Rafał K. Mantiuk, Jonas Unger(https://arxiv.org/abs/1710.07480)
4. Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement, Chunle Guo, Chongyi Lim Jichang Guo, Chen Change Loy, Junhui Hou, Sam Kwong, Runmin Cong(https://openaccess.thecvf.com/content_CVPR_2020/papers/Guo_Zero-Reference_Deep_Curve_Estimation_for_Low-Light_Image_Enhancement_CVPR_2020_paper.pdf )
5. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks , Jun-Yan Zhu, Taesung Park, Phillip Isola ,Alexei A. Efros(https://junyanz.github.io/CycleGAN/)
6. ExpandNet: A Deep Convolutional Neural Network for High Dynamic Range Expansion from Low Dynamic Range Content, Demetris Marnerides, Thomas Bashford-Rogers, Jonathan Hatchett, Kurt Debattista(https://arxiv.org/abs/1803.02266)
Shape Estimation
Qiqin Le, Qiao Zhang
Overview
Pipeline
Tao, Michael W., et al. "Shape estimation from shading, defocus, and correspondence using light-field angular coherence."
Image Source
http://lightfield.stanford.edu/lfs.html
Our images taken by the Lytro Illum camera
Defocus Depth
Defocus Depth
Depth Map & Confidence Map & Shape Estimation
Correspondence Depth
Correspondence Depth
Depth Map & Confidence Map
Combined Depth from Defocus and Correspondence
Combined Depth from Defocus and Correspondence
Depth Map & Confidence Map
Shading Depth
Albedo
Shading Depth
Spherical Harmonic
Defocus Depth
Depth Map & Confidence Map
Correspondence Depth
Depth Map & Confidence Map
Combined Depth from Defocus and Correspondence
Depth Map & Confidence Map
Defocus Depth
Depth Map & Confidence Map
Correspondence Depth
Depth Map & Confidence Map
Combined Depth from Defocus and Correspondence
Depth Map & Confidence Map
References
[1] Tao, Michael W., et al. "Shape estimation from shading, defocus, and correspondence using light-field angular coherence." IEEE transactions on pattern analysis and machine intelligence 39.3 (2016): 546-560.
[2] Tao, Michael W., et al. "Depth from combining defocus and correspondence using light-field cameras." Proceedings of the IEEE International Conference on Computer Vision. 2013.
Q & A
Normal Estimation for Transparent Objects
Amy Lee
302
Overview + Motivation
303
Pipeline
304
Image Set
305
Image Segmentation
306
Image
Mask
Final
Initial Silhouette-based Surface Normals
Key Ideas:
307
Initial Silhouette-based Surface Normals
308
Orientation Consistency w/ Reference Sphere
309
Normal Clustering
310
Normal Clustering
311
Finding Optimal Normal Clusters
Graph-cut cost minimization using max flow algorithm.
312
Final Calibrated Normals
313
Quality of results are heavily influenced by the initial normals
References
314
Thank you for your time!
315
Event-based Object Tracking-by-Detection
Jessica Lee
Why Event-Based Vision?
Event-based vision measures log-based changes in brightness for each pixel in the sensor independently
In comparison to a typical camera, which captures via rolling shutter synchronously [1].
Polarity - moving in/out of a scene
Data point: (ts, x, y, p)
Prior Work on Event-Based Object Detection
Image Reconstruction [2]
Event Tensors:
RED [3] (SoTA)
backbone
heads
Our Approach - Tracking by Detection
CenterTrack [4]
Event Volumes [5]
Volume t
Volume t-1
Tracks t-1
Tracks t
Sizes t
Offsets t
Inputs
Outputs
Current Results & Future Work
Results:
Future Work:
Citations
[1] E. Mueggler, B. Huber, and D. Scaramuzza, Event-based, 6-DOF Pose Tracking for High-Speed Maneuvers. IROS 2014
[2] H. Rebecq, R. Ranftl, V. Koltun, and D. Scaramuzza. High speed and high dynamic range video with an event camera. IEEE Transactions on Pattern Analysis and Machine Intelligence 2019.
[3] E. Perot, P. Tournemire, D. Nitti, J.Masci, and A. Sironi. Learning to Detect Objects with a 1 Megapixel Event Camera. NeurIPS 2020.
[4] X. Zhou, V. Koltun, and P. Krähenbühl. Tracking Objects as Points. ECCV 2020.
[5] A. Zhu, L. Yuan, K. Chaney, and K. Daniilidis, Unsupervised event-based learning of optical flow, depth, and egomotion. CVPR 2019.
Distortion-Free Wide-Angle Portraits on Camera Phones
Ji Liu, Zhuoqian Yang
12/17/2020
Introduction and
Method Overview
Introduction
Introduction
Before correction After correction
Method Overview
Results on Images of Our Own
Naive Stereographic Result
input
naive stereographic
flow
Subject Mask Segmentation Result
input
face mask
Correction Result
Before correction
After correction
Method - Subject Mask Segmentation
Method - Stereographic Projection
Stereographic projection can picture the 3D world onto a 2D plane with minimal conformal distortion, at the expense of changing the curvature of long lines.
We enforce the stereographic projection locally on face regions to correct perspective distortion.
Given the camera focal length f , we compute the stereographic projection from the input using a radial mapping:
Method - Mesh Placement
Before and after naive stereographic, we can observe strong artifacts around the face regions.
Method - Local Face Undistortion
We minimize the following energy function to determine an optimal mesh.
where Et is the weighted sum of several energy terms including Face Objective Term, Line-Bending Term and Regularization Term.
Method - Regularization Term
we regularize the mesh by encouraging smoothness between 4-way adjacent vertices using a regularization term
Method - Line-Bending Term
On the boundary between the face and background, straight lines may be distorted because the two regions follow different projections. We preserve straight lines by encouraging the output mesh to scale rather than twist by adding a line-bending term. The line-bending term penalizes the shearing of the grid, and therefore preserves the edge structure on the background
Method - Mesh Boundary Extension
Similar to gradient-domain image processing, a simple boundary condition by forcing
vi = pi on the mesh boundary would do the job. However,
it creates strong distortions when faces are close to image boundary.
This distributes the distortion to padded vertices, and reduces artifacts near the boundary of the output image.
Method - Similarity Constraint
Constrain the transformation around each facial area to be a similarity transformation that preserves scale
Extension of the Method to Other Object Categories
Thank you!
Mirror, Mirror On the Wall: Detecting Mirrors and Correcting Point Clouds
Sachit Mahajan
sachitma
345
345
345
345
Motivation
Matterport3D: Learning from RGB-D Data in Indoor Environments
346
346
346
346
Motivation
Taken from Phone (camera+Lidar) and ‘3d Scanner App’
347
347
347
347
Setup
Logitech c920
Velodyne
VLP-16
AprilTAG
IMU
348
348
348
348
Methodology
349
349
349
349
Calibrate Extrinsics and Locate Mirror/Entrances
350
350
350
350
Classify mirrors and Remove Points
Two different methods
351
351
351
351
Classify mirrors and Remove Points
Mask Obtained From Region Growing
Mask Obtained From MirrorNet
352
352
352
352
Results
353
353
353
353
Future Work
354
354
354
354
Image Deblurring Using Inertial Measurement Sensors
Kevin O’Brien
kobrien
Problem Motivation/Description
Blurry images stink! She gets it --------------------->
Deconvolution is difficult
-Optimization with no closed-form solution
-Different optimization strategies perform well for some types of scenes, poorly for others
-Generally want to enforce sparse gradients in the image to preserve sharp edges
Blind Deconvolution
Simultaneously solve for the sharp image I and the blur kernel K
-This is an even harder optimization problem, solving for two things at once
-Advances in blind deconvolution come from introducing additional constraints
-sharp edge prediction, color statistics
MAIN IDEA
When we’re dealing with camera shake, use the camera’s motion to help inform the blind deconvolution problem
Blur as a Homography
For small camera movements during exposure, blurring can be parametrized as small planar warps
Sensor Data Processing
Accelerometer readings give linear acceleration (m/s^2)
Gyroscope readings give rotational velocity (rad/s)
Procedure:
Implementation Status
ADMM deblurring optimization with known blur kernel
Implementation Status
Deblurring using ground truth camera positions
Implementation Status
Use raw sensor values next
Thank you!
Light field 3D reconstruction
Wenxuan Ou (Owen)
Abstract
Algorithm
Reference: Kim et al
Algorithm
Reference: Lin et al
Results
Original image
EPI
Disparity
Confidence
Results
Original image
My depth map
AFI depth map
Discussion and conclusion
Thank you !
3D Scanning with Structured Light
George Ralph
Hello to everyone passing through!
Good luck on your presentations ♥
Background
Reducing Global Illumination Effects
Pattern Binarization
Two-Image Binary Codes
One-Image XOR Gray Codes
Other Filtering Techniques
Median Filtering of Correspondences
Point Cloud Filtering
Examples
Examples
Examples
Thank You!
Feel free to ask questions.
Adaptive SPAD Imaging with Depth Priors
Po Ryan
Background
SPAD Sensor
[Gupta et al. “Photon-Flooded Single-Photon 3D Cameras” CVPR 2019]
Depth Estimate
Establish Depth Prior
Adjust Gate and Active Time of SPAD
Method
Stereo Pairs
Depth Estimate
Adaptive SPAD
Conventional SPAD
Stereo Depth Prior
[Xia et al. “Generating and Exploiting Probabilistic Monocular Depth Estimates” CVPR 2020]
CNN
Single RGB Image
Depth Estimate
Uncertainty
Adaptive SPAD
Conventional SPAD
NN-Based Probabilistic Single RGB Depth
Application to Async. Acquisition
[Gupta et al. “Asynchronous Single-Photon 3D Imaging”]
Integration with NN-Based Pipelines
[Siddiqui et al. “An Extensible Multi-Sensor Fusion Framework for 3D Imaging” CVPR 2020]
SPAD Only Adaptation?
[Pediredla et al. “Signal Processing Based Pile-up Compensation for Gated Single-Photon Avalanche Diodes”]
Thanks!
Computational Photography Project
CMU-15663
Tejas Zodage , Harsh Sharma
Handheld Photometric Stereo
Motivation
To develop a simple device for dense 3D reconstruction of static scene.
Problem definition
A set of images taken by camera with a fixed Point light source attached to it.
Add photometric constraints to multi-view stereo to get a dense depth map.
391
Problem definition
Camera
Point light source rigidly attached to the camera
Input: m images
.
.
.
Output: depth image
algorithm
Procedure
393
Initial estimates
Depth planes
Optimization
Model the problem into a graph-cut optimization problem and solve for the constraints jointly
395
Can’t we solve it like HW5?
Near-light photometric stereo | Conventional photometric stereo |
Directional light source | Point light source |
Orthographic camera | Perspective camera |
No!
Data generation
Blender + Mitsuba
397
Results
Plane sweep stereo - depth map
398
Challenges
Future work
References
Image Morphing
Jiazheng Sun
Sample effect
Step 1 - set up
Step 2 - build control points
Step 3 - perturb, interpolate & blend
Curves & Blending
Citation: Cem, Yuksel. 2020. A class of C2 interpolating splies. ACM Transactions on Graphics, 39, 5, 2020
Step 4 - texture mapping
Additional functionalities
Precompute trajectory demo
3D Scene Flow Using Lightfield Images
Kevin Wang
Lightfield Image Structure
Relating Ray Flow to Optical Flow
Lucas-Kanade
Horn-Schunck
Example Results
A Computational Approach for Obstruction- Free Photography
Wesley Wang
(wesleyw)
Overview:
Overview:
Overview:
Implementation:
Implementation:
Results:
Questions?
Confocal Stereo
Hiroshi Wu
Confocal Stereo
In assignment 4...
The whole process (as detailed in paper)
Although the core idea is the same, the whole process is more involved
Example: f=50, a=2.8
2. Relative exitance estimation
3. Image alignment
Smaller FOV
3. Image alignment
4. AFI
5. AFI model fitting
6. Result from piano dataset
Image smoothing via L0 Gradient Minimization
Ye Wu
Objective
Algorithm
Pipeline
Start
Gradient y
Gradient x
Mix 1
Pipeline
Mix 1 with threshold
Gradient x new
Gradient y new
Mix 2
end
Results
k=2.0
k=3.0
l=0.1
l=0.01
l=0.001
l=0.0001
l=0.00001
Results
k=2.0
k=4.0
k=6.0
k=8.0
k=10.0
l=0.1
l=0.03
l=0.00001
Comparison
L0 minimization
3.05s
Bilateral filtering
64.13s
Application
Image abstraction
Application
Clip-art compression artifact removal
Application
Combination of BLF and L0 method
Origin
L0
BLF
BLF+L0
Main References
Thank you!
Q&A
Fast Reflection Removal using Hierarchical Bilateral Grids
Zhichao Yin
(zhichaoy@)
Reflection removal
451
Photo taken through glass
Ground-truth background/transmissive layer
High-res reflection removal
452
High-res challenges – runtime & memory
453
CPU runtime & memory vs. input resolution on real-world test set (208 images)
High-res challenges – degraded quality
454
Input
Niklaus etal.1
Ground-truth
Result upsampled from prediction with 512-res input
High-res challenges – degraded quality
455
Input
Niklaus etal.1
Ground-truth
Result upsampled from prediction with 1024-res input
High-res challenges – degraded quality
456
Input
Niklaus etal.1
Ground-truth
Result upsampled from prediction with 2048-res input
HDRNet1 works well for tone mapping
457
Input
Output
But HDRNet1 cannot work for dereflection
458
Addition of reflection parts makes the transformation highly non-linear in color space.
Input
Output
Our method – overview
459
High-res�Comparison
Input, Resolution: 2048 x 1536
GT, Resolution: 2048 x 1536
High-res�Comparison
Input, Resolution: 2048 x 1536
Niklaus etal.
High-res�Comparison
Input, Resolution: 2048 x 1536
Ours
High-res�Comparison
Input, Resolution: 2048 x 1536
Zhang etal.
High-res�Comparison
Input, Resolution: 2048 x 1536
Ours
High-res�Comparison
Input, Resolution: 2048 x 1536
High-res�Comparison
GT, Resolution: 2048 x 1536
High-res�Comparison
Niklaus etal.
High-res�Comparison
Ours
High-res�Comparison
Zhang etal.
High-res�Comparison
Ours
Quantitative Evaluation
471
LPIPS ↓
SSIM ↑
PSNR ↑
Runtime Comparison
472
CPU
GPU
Relative Improvement
Lensless Imaging with DiffuserCam
Emily Zheng
Exploring Lensless Imaging
DiffuserCam Setup + Calibration
Setup
Calibration
Example PSF and Image Capture
Reconstruction with Deconvolution
Results
Results
Remaining course logistics
480
- Gradescope submission site available.
- Gradescope submission site available.
- Tuesday 12/22, 1 – 2 pm, Grace Kuo (lensless cameras, diffuserCam, AR/VR displays).
https://docs.google.com/spreadsheets/d/1CVg7nUbI701pvZFPX3BR0uzKB76Y6PEl3tXmq1UF4AU/edit#gid=0
Class evaluation*s* – please take them!
481