Abstract
To assist humans with their daily tasks, mobile robots are expected to navigate complex and dynamic environments, presenting unpredictable combinations of known and unknown objects. Most state-of-the-art object recognition methods are unsuitable for this scenario because they require that: (i) all target object classes are known beforehand, and (ii) a vast number of training examples is provided for each class. This evidence calls for novel methods to handle unknown object classes, for which fewer images are initially available (few-shot recognition). One way of tackling the problem is learning how to match novel objects to their most similar supporting example. Here, we compare different (shallow and deep) approaches to few-shot image matching on a novel data set, consisting of 2D views of common object types drawn from a combination of ShapeNet and Google. First, we assess if the similarity of objects learned from a combination of ShapeNet and Google can scale up to new object classes, i.e., categories unseen at training time. Furthermore, we show how normalising the learned embeddings can impact the generalisation abilities of the tested methods, in the context of two novel configurations: (i) where the weights of a Convolutional two-branch Network are imprinted and (ii) where the embeddings of a Convolutional Siamese Network are L2-normalised.
Original language | English |
---|---|
Article number | 380 |
Pages (from-to) | 1-14 |
Number of pages | 14 |
Journal | Electronics (Switzerland) |
Volume | 9 |
Issue number | 3 |
Early online date | 25 Feb 2020 |
DOIs | |
Publication status | Published - Mar 2020 |
Funding
Funding: This work has been partially supported by a European Union’s Horizon 2020 grant—Sciroc, No780086.
Funders | Funder number |
---|---|
Open University | |
Horizon 2020 Framework Programme | |
Horizon 2020 | 780086 |
Keywords
- Few-shot object recognition
- Image matching
- Robotics