{"218":0,"2429":0,"2430":0,"2432":0,"2433":0,"2434":0,"2435":0}
Site Home
Site Home
Drexel University Libraries
Drexel University
Contact Us
å
iDEA: DREXEL LIBRARIES E-REPOSITORY AND ARCHIVES
iDEA: DREXEL LIBRARIES E-REPOSITORY AND ARCHIVES
Main sections
Main menu
Home
Search
Collections
Names
Subjects
Titles
About
You are here
Home
/
Islandora Repository
/
Theses, Dissertations, and Projects
/
Method for Attaining Object-Side Light Fields and Metric Depth Maps from a Commercial Plenoptic Camera
Method for Attaining Object-Side Light Fields and Metric Depth Maps from a Commercial Plenoptic Camera
Details
Title
Method for Attaining Object-Side Light Fields and Metric Depth Maps from a Commercial Plenoptic Camera
Author(s)
Aenchbacher, Weston
Advisor(s)
Kurzweg, Timothy P.
Keywords
Optics
;
Physics--Computer simulation
;
Electrical engineering
;
Image processing--Computer simulation
;
Three-dimensional imaging
;
Distances--Measurement
;
Photography--Digital techniques
Date
2018-02
Publisher
Drexel University
Thesis
Ph.D., Electrical Engineering -- Drexel University, 2018
Abstract
Determining object distances and positions is a common computer vision task. This has obvious use in tasks such as autonomous navigation, robotic object retrieval, and surface profiling. Traditional methods for obtaining scene depth information may require active illumination, multiple cameras, or multiple exposures as hardware parameters are incremented. A plenoptic camera can be used to obtain scene depths and has the advantages of being passive and only requiring a single exposure with a single device. Traditional photographic cameras integrate all light arriving at a single detector element or film location, discarding information about the incidence angle of any individual contribution. In a plenoptic camera, an array of microlenses is placed in front of the image sensor, such that light arriving at a given microlens from different directions will arrive at different detector elements. With this, a signal can be attributed to all rays arriving at the microlens array, for which the assemblage is referred to as a "light field." Common light field operations include the creation multiple digital images from one exposure to simulate the effect of changing the camera's focal length ("digital refocus") and the determination of relative distances to scene objects. Existing light field methods are concerned with the light field within the camera (the "image-side light field"), as this is sufficient for most tasks. In contrast, this thesis describes a method for characterization of the light field external to a plenoptic camera (the "object-side light field") by performing experiments that determine the mapping from object-space rays to image-side light field coordinates. Use of an object-side light field allows for the determination of distances to scene objects in absolute terms, rather than the relative depths commonly obtained with image-side light fields. The object-side light field is obtained by observing the appearance of an occluding edge in the image-side light field as it is translated across the plane of the camera's front lens. Perspective images are created for simulated pinhole apertures at the front lens plane and used to view a patterned target, completing the characterization. In this thesis, descriptions of the characterized object-side light field are provided, offering an intuitive understanding of the structure of the light collected by the plenoptic camera. Demonstrations of common light field operations are provided to confirm that there is no loss of functionality through use of an object-side light field rather than an image-side light field. Simulations confirm that the distance to objects near to the camera can be determined to within 1~mm, with accuracy decreasing as the distance to an object increases.
In Collections
Theses, Dissertations, and Projects
/islandora/object/idea%3A8175/datastream/OBJ/view
Search iDEA
All formats
Search by:
Keyword
Name
Subject
Title
Advanced Search
My Account
Login