Abstract
One common shortcoming of modern computer vision is the inability of most models to generalize to new classes—one/few shot image recognition. We propose a new problem formulation for this task and present a network architecture and training methodology to solve this task. Further, we provide insights into how careful focus on how not just the data, but the way data presented to the model can have significant impact on performance. Using these method, we achieve high accuracy in few-shot image recognition tasks.
Degree
MS
College and Department
Physical and Mathematical Sciences; Computer Science
Rights
https://lib.byu.edu/about/copyright/
BYU ScholarsArchive Citation
Hurlburt, Daniel, "The "What"-"Where" Network: A Tool for One-Shot Image Recognition and Localization" (2021). Theses and Dissertations. 9366.
https://scholarsarchive.byu.edu/etd/9366
Date Submitted
2021-01-06
Document Type
Thesis
Handle
http://hdl.lib.byu.edu/1877/etd12003
Keywords
computer vision, semantic segmentation, few-shot learning, one-shot learning, embedding
Language
english