This document summarizes Senthilkumar Gopal's computational photography portfolio from Georgia Tech. It includes summaries of 11 assignments completed as part of the program. The assignments covered topics like camera obscura, HDR imaging, panoramas, feature detection, and video textures. Through the assignments, Senthilkumar learned techniques like pyramid blending, gradient filters, and building image processing pipelines. He also demonstrated going above and beyond the assignments by experimenting with different techniques.
2. Assignment # 1: A Photograph is a Photograph
PURPOSE
The purpose of this assignment is to showcase one of our photographs and identify methods to
improve the image using computational photography techniques.
LEARNINGS
This assignment served two purposes. Being the first course in the program, this helped me
understand the nuances of assignment submission, evaluation process and the logistics in a gentle
manner. This also served as a great introduction to the course and the amount of research being done
in this field in contemporary times.
I cited the techniques of “pixel-wise flow field motion” and “edge flow” methods to identify the motion of
the video in each layer and create layer masks to mark the reflection/occlusion parts of the image, in
this paper which can be used to improve this image by removing the occluding fence grids.
3. Assignment # 1: A Photograph is a Photograph
Oakland, CA, USA
August 7, 2016 - 9:56 AM PST
Canon 550D - 1/800, f/6.3, ISO 1600
“The unblinking eye (Lilford Crane)”
4. PURPOSE
The purpose of this assignment is to get the development environment setup using vagrant and
understand simple arithmetic operations on images. This assignment also gives a great introduction
into Numpy library and its usage for image manipulations.
LEARNINGS
This assignment was my first attempt in using Python and the exercises were simple enough to
complete without any major impediments. Some of the learnings from this assignment are:
● understanding the importance of vectorization
● usage of operator broadcasting rather than iteration
● built in functions similar to fliplr such as flipud, rot90 and many more for image manipulation
● significance of datatypes and their implications for computational manipulations
● basic principles of how blending would work
Assignment # 2: Image I/O
5. Input Output
Observation: This vignette in the image was converted
into black pixels and the boat/sail were converted to
white pixels. However, an interesting observation is the
presence of black pixels in the water, even though the
original image does not give the impression that these
are dark grey pixels.
Blended Output image
Input Images
Convert to Black and White Image Average of Two Images
Assignment # 2: Image I/O
6. Assignment # 3: Epsilon Photography
PURPOSE
The purpose of this assignment is to generate a novel image or artifact which combines images of the
same scene or object with a variation of just one photographic parameter in the process.
SETUP
Inspired by this video, I choose to create a stop motion photography artifact where my epsilon is to
vary the location of the elements in the composition, slightly between images to create the novel
imagery of “stop-motion photography”.
The images were of a real rose flower with location of each of its petals changing slightly across
different pictures. I took a total of 32 images out of which the last 8 were submitted for the assignment
and used to build the final GIF images.
7. Assignment # 3: Epsilon Photography
Final GIF: http://imgur.com/a/psRQr
Full Stop Motion Video: https://www.youtube.com/watch?v=afJHcxMQHG0
8. Assignment # 3: Epsilon Photography
LEARNINGS
● Being my first attempt at Epsilon photography, I was very proud of the result.
● Testing the setup end to end using sample pictures and generating a test GIF, gave me the
confidence in the plan before implementing with the rose petals.
● The setup with external flash and tripod worked well, eliminating any need for heavy post
processing for alignment.
● I realized that the idea of motion is to illustrate more fluidity which can only be achieved with
more granular changes. For great output, I would have made really smaller changes (almost
1/3rd of what is performed in these images)
9. Assignment # 3: Epsilon Photography
ABOVE & BEYOND
For my extra credit, i tried to perform a different type of Epsilon photography by changing the focal
plane of the images and then combining them to create an all focused image using the technique
“Focus stacking”
Input Images with varying focal planes Output Image with all planes in focus
10. Assignment # 4: Gradients and Edges
PURPOSE
● The purpose of this assignment is to provide the implementation of different functions to
investigate various filters and create a rudimentary edge detection pipeline.
● The second part is to investigate the effects of different filters such box filter, Gaussian filter and
some of the commonly used gradient filters (Sobel filter).
● The last part is to use the functions created earlier and the Sobel filter to form an edge detection
pipeline and compare results with the Canny edge detector available in the OpenCV library.
11. Assignment # 4: Gradients and Edges
LEARNINGS
● Variations of the copyMakeBorder and np.pad can create effects such as vignetting, mirroring etc.,
● Different filter types - Box, Blur, Emboss, Edge etc., and its significance
● Importance of cross-Correlation and convolution
● How sampling and downsizing of images work
● How to use intensity changes and their direction to identify an edge
● How edge detection can be used for feature matching
● Doing this exercise has taught me to appreciate how a simple filter configuration, Sobel, box filter
etc., can perform varying computations to generate novel images
● Applications for feature detection, machine learning, artificial intelligence
● Learnt the significance of eliminating noise and why sharpness of an image matters for computer
vision and computational photography
12. Assignment # 4: Gradients and Edges
Gradient X Normalized Image Gradient Y Normalized Image copyMakeBorder() np.pad(img, 200, 'reflect')
Input image Gaussian filter Box filter
13. Assignment # 4: Gradients and Edges
SIMPLE EDGE DETECTION PIPELINE
● Combine the x and y direction magnitudes using np.hypot which provides the discrete magnitude
● A rudimentary implementation of edge detection from the gradients is to use the simple
convertToBlackAndWhite operation on the gradient magnitude
● However, a threshold of 128 gave very few details and missing critical ones while a threshold of 30
included lot of noise and the intermediary value of 90 was still not complete with details.
● Though this method was fast and provides some reliable results, it has some disadvantages.
FINAL EDGE DETECTION PIPELINE
● We need to run the image through a Gaussian filter to remove high intensity spikes and noise
● Compute the Sobel gradients on the x and y directions
● Combine them to identify the discrete magnitude at each pixel
● Perform non-maximum suppression by identifying gradient direction of the Sobel gradient magnitude
● Now all pixels are suppressed other than the ones part of the local maxima(edge).
● This also helps in “thinning” the edges which was not possible with the simple
convertToBlackAndWhite operation.
14. Assignment # 4: Gradients and Edges
#1 Gaussian filter appliedInput Image #2 Sobel filter - X direction
#3 Sobel filter - Y direction
#4 Sobel gradient magnitude Thinning of edges / local Maxima
Final image after
thresholding
changes
LO=0.1m HI=0.2m
15. Assignment # 4: Gradients and Edges
LO=0.1m HI=0.2m
LO=0.05m HI=0.1m
LO=0.05m HI=0.1m
FEW OTHER RESULTS WITH THE EDGE DETECTION PIPELINE
16. PURPOSE
● This was the most “physically active” assignment. Until then I had used pictures from the past,
but this time I had to “create” one.
● This assignment required considerable amount of time for setup and preparation of the location.
● As my apartment was on the second floor, it is one of the best lit buildings and covering all the
light sources in the room was the biggest task.
● Moving boxes, black plastic sheeting, Gorilla tape and a sharp knife were used to aid the setup.
● The image captured is of my apartment complex as seen from my bedroom window, which was
covered and used as the camera obscura.
LEARNINGS / OUTCOME
● This assignment reproduces one of the oldest image capturing techniques used from 400 BC.
● Multiple images were stitched using Photoshop to create the final output image.
● During “BULB” mode, AF (auto focus) should be turned off and long exposures drain the battery
quickly.
Assignment # 5: Camera Obscura
17. Blocked window Tools used
Camera & Pinhole location Screens
Assignment # 5: Camera Obscura
SCENE OUTSIDE THE PINHOLESETUP
19. Assignment # 6: Blending
PURPOSE
● This assignment explores the steps necessary to build a Gaussian and Laplacian pyramids and
use them to blend two images using a pre-generated mask.
● This exercise introduces the concept of pyramid building using REDUCE and EXPAND algorithm
and multiple other uses of the Gaussian and Laplacian pyramids.
● Multiple functions which aid the pyramid building and the blending of the images were created as
part of the assignment
● These functions are then used to demonstrate the capability of blending using these pyramids.
LEARNINGS / OUTCOME
● The information retained in the final blended image was a great learning exercise and gives a
peek into how this could be useful for progressive transmission of image data.
● The blended output image was extremely close to the output from Photoshop.
● The resulting image demonstrated the clear usage and purpose of the Laplacian pyramid and the
minute details in the image which were retained in the final output
23. Assignment # 6: Blending
● For my A&B, i started investigating the other types of blending algorithms and capabilities
available in OpenCV which can perform the blending with better smoothing and better tonal
retention
● Common ones used are linear blending using weighted or arithmetic operations to incorporate an
opaque image into another image.
● However, one of the more recent and highly used algorithm is the Poisson distribution based
blending which has better retention of tonal details.
Image 2Image 1
Blended Image with Poisson
24. Assignment # 7: Feature Detection
PURPOSE
● This assignment builds the function blocks to perform feature matching between different images
of the same object under different conditions of scale, lighting, rotation and among other objects.
● The second part of the assignment deals with experimenting with various parameters of the ORB
algorithm used for feature matching and its results.
LEARNINGS / OUTCOME
● This assignment opened up avenues of how feature matching works and various factors
influencing good and bad results.
● Choosing a good template image is critical to ensure good results.
● Image quality and closeness in scale and intensity are important factors for feature detection.
● Pre-enhancing the image definitely helps in feature detection.
● A single pipeline cannot be fit to match all types of images.
● There are multiple optimizations possible at each step of the pipeline for feature matching. A
better descriptor or a faster detector etc., can be utilized based on the needs of the pipeline.
27. Assignment # 7: Feature Detection
POOR IMAGES SIFT MATCHES
GOOD IMAGES SIFT MATCHES
● For my A&B, I tried to investigate the other detectors and descriptors which are available and their
results against my images.
● Tested Brute-Force Matching with SIFT Descriptors and FLANN Matchers.
● The results suggest that ORB is a very good and extremely fast alternative to the commercially
available feature matchers available.
28. PURPOSE
● This assignment creates the function blocks for building a pipeline which can stiches multiple
images from left to right creating a panorama image.
● The functions help find features, calculate the homography, warp images from left to right, align
and then blend the images to create the final artifact.
LEARNINGS / OUTCOME
● Feature matching and alignment based on RANSAC worked perfectly without any false positives.
● Blending seemed to be a process which becomes increasingly complex with need for better
quality. However, basic blending algorithms such as linear weighted, center weighted does give a
great result.
● The panorama pipeline is very consistent with no specialization required for different image sets.
● Even while stitching five images, the results were consistent and high quality.
● It is possible to achieve commercial software quality output with a simple pipeline.
Assignment # 8: Panoramas
29. Assignment # 8: Panoramas
Test image result using center
weighting image blending
My image result using center weighting image
blending
30. Assignment # 8: Panoramas
Center weighted blending
Linear weighted blending
Gaussian pyramid blending
For my A&B, I experimented with different blending functions - linear weighted, center weighted and
gaussian pyramid blending out of which linear and center weighted does give a great result.
31. Assignment # 9: HDR
PURPOSE
● This assignment develops the code to combine images of varying exposures to build a composite
image which captures a wider dynamic range of irradiance.
● The building block functions computes the response curve and radiance map which then is used
to form the final image by using pixels from all the different exposure images.
LEARNINGS / OUTCOME
● Choosing a scene for HDR needs preparation and proper bracketing techniques to capture the
wide range of the spectrum.
● As there was no tone mapping, the results were dull and not exhibiting the bright intensity
exhibited by the original image. Adding tone mapping via one of the algorithms such as Durand,
Drago or Reinhard helps the image get displayed better.
● Adding a tone map to better visually display the image, different weighting algorithms and
vectorization would be enhancements possible for the pipeline.
34. Assignment # 9: HDR
For my A&B, i worked with my own set of images and attempted to match the tones of the image
using linear and Reinhard tone mapping
My own HDR images
35. Assignment # 9: HDR
Reinhard Tone MappingLinear Tone MappingHDR without tone mapping
36. Assignment #10: Pictures of Space
PURPOSE
● The purpose of this assignment is to build a point cloud and panorama using available software
to learn the usage and capabilities of the basics learnt earlier.
● A point cloud is a set of data points in a three-dimensional coordinate system and are used to
represent three dimensional objects or capture a location in three dimensions.
● The point cloud and panorama was captured in Central Park, Santa Clara, CA which is a public
park with a great water fountain filled with ducks and a trail around the fountain.
LEARNINGS / OUTCOME
● This assignment helped learn the different projection systems for creating a panorama and
helped identify my attempt as Cylindrical panorama with a planar reprojection
● The point clouds also related to the Microsoft Photosynth product help learn the usage of multiple
images to form a novel image and capability to provide crowd sourced image clouds.
37. Used 44 images for my panorama to get a 360o
FOV and stitched using Adobe Photoshop
The full version of this image is available at KUULA and Google Drive
Assignment #10: Pictures of Space
38. Assignment #10: Pictures of Space
Images + Mesh for 300 images available here. Created using OpenSFM
● The images to form the point cloud was captured in Central Park, Santa Clara, CA. I wanted to
create the point cloud as a "Walk" and attempt to capture various image along the path.
● To provide a wider view of the scene, I took three overlapping picture for every step which is
evident from the camera angles in the point cloud.
40. Assignment #10: Pictures of Space
● For my A&B work, i attempted different spatial projections as earlier ones were only horizontal.
● So i took images of the top and bottom sections of the view (vertical projections) and stitched them all
together using Photoshop to a FOV of 170o
● The image is available at Google Drive and at Kuulu
41. PURPOSE
● The purpose of this assignment is to identify a part of a video which can generate an infinite loop
video with minimal or no visible break.
● As part of this assignment, we build a pipeline to compute the similarity matrix between frames of
a video and use them to identify and create a video texture.
● The assignment focuses on finding the biggest loop possible from the similarity matrix and
generating the infinite looping video.
LEARNINGS / OUTCOME
● Being the first assignment using a video instead of images, this assignment helped understand
the structure and how multiple frames compose a video.
● The building blocks of the assignment provided insight into how the pipeline would function to
create a video texture.
Assignment #11: Video Textures
42. Assignment #11: Video Textures
The assignment provides a set of candle images to demonstrate the video texture.
Link to the candle video texture gif: https://imgur.com/a/SG2OT
43. Assignment #11: Video Textures
I used a video clip that i recorded at home to create a video texture.
Link to the water drops video texture gif: https://imgur.com/a/SG2OT
44. Assignment #11: Video Textures
For my A&B, I experimented with few other video clips available from the internet
Source
https://www.youtube.com/watch?v=5lCK4jkqQxA
Link to video texture gif
http://imgur.com/a/OauHQ
Source
https://www.youtube.com/watch?v=WE-nFpUScEU
Link to video texture gif
http://imgur.com/a/V2BgY
45. FINAL PROJECT
GOAL
● The final project is based on this paper (Whiteboard scanning and image enhancement) which
discusses the method of extracting usable documentation out of whiteboard images which is a
useful artifact for many groups.
● The scope of this project is to take a picture of whiteboard in any perspective and produce a
perspective corrected, trimmed and enhanced image which can make it so much easier to save
and process the whiteboard images further.
● As part of the project, I also included the capability to take in multiple images forming a
panorama of the whiteboard and produce a corrected output.
● I also ported the code to Android and create an app to perform the similar activity as described.
51. FINAL PROJECT - Pipeline
Extraction pipeline and intermediate results
52. FINAL PROJECT - Results
Video of the app usage - Using a Camera
Video of the app usage - Using an existing image
● This was a great learning experience about uses of feature
detection and customization to address specific image types,
generate novel images and find useful data.
● This project helped me learn about different contrast and image
data enhancement techniques.
53. ● Georgia Tech for creating this great platform for learning - OMSCS
● Prof. Irfan Essa for the lectures and such an enriched course
● The TAs who are always available to discuss and help our questions and thoughts
on both Piazza and Slack.
● Stackoverflow which helped with all the questions i had regarding Numpy operations,
OpenCV errors etc.,
Thanks!