|Title||One Shot Learning via Compositions of Meaningful Patches|
|Publication Type||Conference Paper|
|Year of Publication||2015|
|Authors||Wong, A, Yuille, A|
|Conference Name||International Conference on Computer Vision (ICCV)|
The task of discriminating one object from another is al- most trivial for a human being. However, this task is compu- tationally taxing for most modern machine learning meth- ods; whereas, we perform this task at ease given very few examples for learning. It has been proposed that the quick grasp of concept may come from the shared knowledge be- tween the new example and examples previously learned. We believe that the key to one-shot learning is the sharing of common parts as each part holds immense amounts of in- formation on how a visual concept is constructed. We pro- pose an unsupervised method for learning a compact dictio- nary of image patches representing meaningful components of an objects. Using those patches as features, we build a compositional model that outperforms a number of popu- lar algorithms on a one-shot learning task. We demonstrate the effectiveness of this approach on hand-written digits and show that this model generalizes to multiple datasets.
- CBMM Funded