Content-Aware Manipulations for Image and Video Collections
Author
Dale, Kevin
Metadata
Show full item recordCitation
Dale, Kevin. 2012. Content-Aware Manipulations for Image and Video Collections. Doctoral dissertation, Harvard University.Abstract
Digital photography and videography have become ubiquitous. With ever-cheaper andmore capable cameras, found in dedicated devices and, increasingly, in multipurpose
smartphones and tablets, it is now easier than ever for the casual user to generate their
own dense stream of personal multimedia data. With popular photo and video sharing
sites, like Flickr, Facebook, and YouTube, users can share their images and video with
the world, making for a vast amount of multimedia data stored at home and on the web.
This flood of data presents many challenges, particularly for the non-professional, to
manage their data, for example, to apply simple photographic adjustments, or to find
interesting shots worth keeping among megabytes of data from even a short weekend
trip. At the same time, researchers have a unique opportunity to exploit the vast
amount of publicly available multimedia data to make graphics tasks easier for the
individual. This dissertation presents work that seeks to address some of these challenges, and,
where possible, exploit existing data sets to do so. First, we discuss a general approach
that finds images similar to a given input from among a collection of photographs, from
which various task-specific properties are transferred to the input. We demonstrate this
basic approach in two distinct settings–image restoration and CG image enhancement.
Next, we focus on collections of video. We first present a method for efficiently
browsing and summarizing collections of related videos. Our approach is based on a
simple pairwise video alignment that identifies a relevant sequence of video clips that
best matches an input video. Finally, we discuss our work on replacing facial
performances in video that requires no special hardware and can be used to retarget
existing footage to synthesize new performances.
Terms of Use
This article is made available under the terms and conditions applicable to Other Posted Material, as set forth at http://nrs.harvard.edu/urn-3:HUL.InstRepos:dash.current.terms-of-use#LAACitable link to this page
http://nrs.harvard.edu/urn-3:HUL.InstRepos:9414577
Collections
- FAS Theses and Dissertations [5848]
Contact administrator regarding this item (to report mistakes or request changes)