<div class="csl-bib-body">
<div class="csl-entry">Thalhammer, S., Jean-Baptiste Weibel, Markus Vincze, & Rodriguez-Garcia, J. (2023). Self-supervised Vision Transformers for 3D Pose Estimation of Novel Objects. <i>Image and Vision Computing</i>, <i>139</i>, Article 104816. https://doi.org/10.1016/j.imavis.2023.104816</div>
</div>
-
dc.identifier.issn
0262-8856
-
dc.identifier.uri
http://hdl.handle.net/20.500.12708/190773
-
dc.description.abstract
Object pose estimation is important for object manipulation and scene understanding. In order to improve the
general applicability of pose estimators, recent research focuses on providing estimates for novel objects, that is,
objects unseen during training. Such works use deep template matching strategies to retrieve the closest template connected to a query image, which implicitly provides object class and pose. Despite the recent success and improvements of Vision Transformers over CNNs for many vision tasks, the state of the art uses CNN-based
approaches for novel object pose estimation. This work evaluates and demonstrates the differences between
self-supervised CNNs and Vision Transformers for deep template matching. In detail, both types of approaches
are trained using contrastive learning to match training images against rendered templates of isolated objects. At
test time such templates are matched against query images of known and novel objects under challenging settings, such as clutter, occlusion and object symmetries, using masked cosine similarity. The presented results not only demonstrate that Vision Transformers improve matching accuracy over CNNs but also that for some cases pre-trained Vision Transformers do not need fine-tuning to achieve the improvement. Furthermore, we highlight the differences in optimization and network architecture when comparing these two types of networks for deep template matching.
en
dc.language.iso
en
-
dc.publisher
ELSEVIER
-
dc.relation.ispartof
Image and Vision Computing
-
dc.subject
Object recognition
en
dc.title
Self-supervised Vision Transformers for 3D Pose Estimation of Novel Objects