Submission: FoundPose+FeatRef+Megapose/TUD-L/FoundPose+FeatRef+Megapose

Download submission
Submission name FoundPose+FeatRef+Megapose
Submission time (UTC) Jan. 29, 2024, 9:44 a.m.
User epi
Task 6D localization of unseen objects
Dataset TUD-L
Training model type Default
Training image type None
Description
Evaluation scores
AR:0.633
AR_MSPD:0.786
AR_MSSD:0.601
AR_VSD:0.513
average_time_per_image:0.943

Method: FoundPose+FeatRef+Megapose

User epi
Publication
Implementation
Training image modalities RGB
Test image modalities RGB
Description

The presented results were achieved by FoundPose with the featuremetric refinement and additional MegaPose [D] refinement (row 8 of Table 1 in [A]).

In this submission, FoundPose uses default CNOS-FastSAM [B] segmentations provided for BOP'23. For pose estimation, the method uses features from layer 18 of DINOv2 (ViT-L) with registers [C].

Note that FoundPose doesn't do any task-specific training -- it only uses frozen FastSAM (via CNOS) and frozen DINOv2. The only component that is used in this submission and trained in a task-specific manner is the MegaPose refiner (we used weights from the official MegaPose repository and didn't train them further).

[A] Anonymous: FoundPose: Unseen Object Pose Estimation with Foundation Features.

[B] Nguyen et al.: CNOS: A Strong Baseline for CAD-based Novel Object Segmentation, ICCVW 2023.

[C] Darcet et al.: Vision transformers need registers, arXiv 2023.

[D] Labbé et al.: MegaPose: 6D Pose Estimation of Novel Objects via Render & Compare, CoRL 2022.

Computer specifications Tesla P100 16GB