footnote
Method Towards CVPR 2021 Image Matching Challenge
Abstract
This report describes Megvii-3D team’s approach towards SimLocMatch Challenge @ CVPR 2021 Image Matching Workshop. It includes the submissions:
- •
- •
- •
- •
- •
- •
- •
The methodology we took is similar to our solution towards CVPR 2021 Image Matching Challenge, hence we only report the additional strategies and tricks besides it.
Method
For the simloc matching Challenge, we tried DISK[1] 8k keypoints with its SuperGlue[2] matcher, as well as the Superpoint[3] 2k keypoints together with DISK 6k keypoints as a combination, where each feature was matched by its corresponding SugerGlue matching. We noticed that the latter outperformed the former by a obvious margin, so we stuck with the second approach.
We compared the single-scale result with the multiple-scale one and found that the first one has better performance in the number of inliers and matching success ratio, whereas the last one worked better in terms of the number of non-matches. This founding agreed with our intuition, as the more the number of matching is, the more the number of inliers potentially, which leads to a later superior matching success ratio. Also, mixing multiple-scale matching outcomes together would result in a boost of the number of matches for non-matching images.
Since the simloc dataset contains two indoor scenes, office and restaurant, which fits the indoor SuperGlue weights better. Due to the limitation of time, we did not have enough resources to train a DISK-SuperGlue matcher for the indoor scenes, thus we only applied the indoor weights of SuperPoint-SuperGlue provided by the original author.
In order to further reduce the number of matches in non-matching images, we decided to set a threshold regarding the number of matches between each pair, as the ’discarding threshold’. We would indicate that 8 is a theoretical number as if the number of matches is less than 8, the pose cannot be solved. While other thresholds are practical numbers. We carried out an analysis of the distribution of the number of matches for each pair and conducted some experiments.
The outlier rejection was implemented by DegenSAC[4] without further tuning.
References
- [1] M. Tyszkiewicz, P. Fua, and E. Trulls, “Disk: Learning local features with policy gradient,” Advances in Neural Information Processing Systems, vol. 33, 2020.
- [2] P.-E. Sarlin, D. DeTone, T. Malisiewicz, and A. Rabinovich, “SuperGlue: Learning feature matching with graph neural networks,” in CVPR, 2020. [Online]. Available: https://arxiv.org/abs/1911.11763
- [3] D. DeTone, T. Malisiewicz, and A. Rabinovich, “Superpoint: Self-supervised interest point detection and description,” CoRR, vol. abs/1712.07629, 2017. [Online]. Available: http://arxiv.org/abs/1712.07629
- [4] O. Chum, T. Werner, and J. Matas, “Two-view geometry estimation unaffected by a dominant plane,” in CVPR, 2005.
Appendix: Details about each Submission
methods | image size(sp disk) | scannet | discard nums | degensac th | scale | disk nms | sp nms | disk max keypoints | sp max keypoints | disk match score | sp match score | degensac iter | inlieres | match success rate | matches(non-matches) |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
sss-sd_100k_1 | 1600/1600 | N | 0 | 1.1 | Y/Y | 4 | 4 | 6000 | 2048 | 0.7 | 0.2 | 100k | 248.66 | 51.02% | 47.72 |
sss-sd_100k_6 | 1600/1600 | Y | 0 | 1.1 | Y/Y | 4 | 4 | 6000 | 2048 | 0.7 | 0.2 | 1000k | 241.23 | 50.12% | 36.99 |
sss-sd_100k_8 | 1600/1600 | N | 0 | 1.1 | Y/Y | 4 | 4 | 6000 | 4096 | 0.7 | 0.2 | 1000k | 319.02 | 52.19% | 47.23 |
aaa-1000k_no_ms | 1600/1600 | N | 8 | 1.1 | N/N | 4 | 4 | 6000 | 4096 | 0.7 | 0.2 | 1000k | 314.43 | 51.04% | 30.01 |
aaa-1000k_no_ms2 | 1600/1600 | N | 50 | 1.1 | N/N | 4 | 4 | 6000 | 4096 | 0.7 | 0.2 | 1000k | 312.75 | 44.63% | 23.15 |
aaa-1000k_80_no_ms111 | 1600/1600 | N | 8 | 0.8 | N/N | 4 | 4 | 6000 | 4096 | 0.7 | 0.2 | 100k | 275.39 | 51.40% | 26.30 |
aaa-1000k_50_no_ms111 | 1600/1600 | N | 8 | 0.5 | N/N | 4 | 4 | 6000 | 4096 | 0.7 | 0.2 | 100k | 214.70 | 51.56% | 21.80 |