aerial surveillance, real-time, dense correspondence algorithm, graphics processing units
Video from aerial surveillance can provide a rich source of data for many applications and can be enhanced for display and analysis through such methods as mosaic construction, super-resolution, and mover detection. All of these methods require accurate frame-to-frame registration, which for live use must be performed in real time. In many situations, scene parallax may make alignment using global transformations impossible or error-prone, limiting the performance of subsequent processing and applications. For these cases, dense (per-pixel) correspondence is required, but this can be computationally prohibitive. This paper presents a hierarchical dense correspondence algorithm designed for implementation on graphics processing units (GPUs). Since the method does not rely on epipolar geometry, it is also suitable for use when there are uncorrected nonlinear lens distortions. A framework for using this dense correspondence to implement local mosaicking, super-resolution enhancement, and mover detection is also presented and demonstrated using examples of each of these types of enhancement and different types of video sources.
Original Publication Citation
S. Cluff, B. Morse, M. Duchaineau, and J. Cohen, "GPU-accelerated hierarchical dense correspondence for real-time aerial video processing," in IEEE Workshop on Motion and Video Computing (WMVC), December 29.
BYU ScholarsArchive Citation
Cluff, Stephen; Morse, Bryan S.; Cohen, Jonathan D.; and Duchaineau, Mark, "GPU-Accelerated Hierarchical Dense Correspondence for Real-Time Aerial Video Processing" (2009). All Faculty Publications. 114.
Physical and Mathematical Sciences
© 2009 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.
Copyright Use Information