FM-Bench: Feature Matching Benchmark (BMVC 2019)

Abstract

Matching two images while estimating their relative geometry is a key step in many computer vision applications. For decades, a well-established pipeline, consisting of SIFT, RANSAC, and 8-point algorithm, has been used for this task. Recently, many new approaches were proposed and shown to outperform previous alternatives on standard benchmarks, including the learned features, correspondence pruning algorithms, and robust estimators. However, whether it is beneficial to incorporate them into the classic pipeline is less-investigated. To this end, we are interested in i) evaluating the performance of these recent algorithms in the context of image matching and epipolar geometry estimation, and ii) leveraging them to design more practical registration systems. The experiments are conducted in four large-scale datasets using strictly defined evaluation metrics, and the promising results provide insight into which algorithms suit which scenarios. According to this, we propose three high-quality matching systems and a Coarse-to-Fine RANSAC estimator. They show remarkable performances and have potentials to a large part of computer vision tasks. To facilitate future research, the full evaluation pipeline and the proposed methods are made publicly available.

Publication

Jia-Wang Bian, Yu-Huan Wu, Ji Zhao, Yun Liu, Le Zhang, Ming-Ming Cheng, Ian Reid, An Evaluation of Feature Matchers for Fundamental Matrix Estimation, British Machine Vision Conference (BMVC), 2019 [PDF] [Code] [Chinese Translation]

@inproceedings{bian2019bench,
  title={An Evaluation of Feature Matchers for Fundamental Matrix Estimation},
  author={Bian, Jia-Wang and Wu, Yu-Huan and Zhao, Ji and Liu, Yun and Zhang, Le and Cheng, Ming-Ming and Reid, Ian},
  booktitle= {British Machine Vision Conference (BMVC)},
  year={2019}
}

Contribution

  1. We present an evaluation protocol for local features, robust stimators, and especially correspondence pruning algorithms such as [7, 26, 47] which have not been carefully investigated.
  2. We evaluate algorithms on four large-scale datasets using strictly defined metrics. The results provide insights into which datasets are particularly challenging and which algorithms suit which scenarios.
  3. We propose three high-quality and efficient matching systems, which perform on par with the powerful CODE [23] system but are several orders of magnitude faster.
  4. Interestingly, we observe that the recent GC-RANSAC [6] (also USAC [34]) does not show consistently high performance on geometry estimation but permits effective outlier pruning. We hence propose to first use it for outlier removal, and then apply LMedS based estimator [36] for model fitting. The resulting approach, termed Coarse-to-Fine RANSAC, shows significant superiority over other alternatives.

Datasets

Metrics

  • %Recall refers to the success ratio of fundamental matrix estimation.
  • %Inlier(-m) refers to the inlier rate after (before) RANSAC-like estimators.
  • #Corrs(-m) refers to the number of correspondences after (before) RANSAC.

Results

Updated Results

  1. By adjusting parameters, GC-RSC shows better results than those reported in the original paper. The new results are shown below:
TUM: 66.40    KITTI: 90.40    T&T: 86.40    CPC:60.00

9 thoughts on “FM-Bench: Feature Matching Benchmark (BMVC 2019)”

  1. 边老师您好,我最近做了一个特征点匹配的算法,想用您的评价标准-An Evaluation of Feature Matchers for Fundamental Matrix Estimation,做一个评估。但是您在github留的数据集下载链接:https://1drv.ms/u/s!AiV6XqkxJHE2g3bdq2yQkr2ET4w5?e=oWEjC5显示无法链接。最后想问下老师Image-Matching-Benchmark与An Evaluation of Feature Matchers for Fundamental Matrix Estimation是同一个评价标准吗?谢谢老师!!

    1. 我更新了一下链接,应该没有问题。Onedrive国内可以访问。这个指标和之前benchmak原理非常接近,实现上略有不同。之前是评估pose,现在是fundamental matrix。其实pose也是从epipolar geometry来的,所以原理一样。

      1. 不好意思啊,边老师,我试了下我们实验室几台电脑还是都连不上啊。然后上了谷歌也连不上。那我先用POSE做评估吧。这个数据集的百度云链接能成功下载对应数据集,谢谢边老师。

  2. 请教边老师,刚学习这个方向的知识,看了您的文章和代码,受益匪浅,但是还有几个问题没理解明白
    1. 评估CF-RSC时,是替换baseline中对应的算法得到的结果么?那如果组合RootSIFT-PCA or HardNet + CF-RSC,那岂不是可能是效果最好的?

    2. Table5中的CF-RSC的结果是没有是用Updated Result中提到的调参么? 如果用了Updated Result的优化参数,Table5的CF-RSC岂不是会更好?

    3. GMS论文中用的orb,提取的点相比sift多,点的数量影响到GMS的精度,对于ORB+GMS在上述评测的算法中效果如何?

    1. To 1 & 2. CF-RSC用的是sift,为了和ransac, lmeds那些保持一致。只是为了公平地对比体现算法优势,并没有专门为了最好的结果。但其实也是有上限的。因为比如sift+gms已经足够精确了,那你用cf-rsc或者lmeds结果都差不多。因为cf-rsc只是在lmeds之前再过滤了一些错误匹配。你可以根据这篇文章的经验去调最高性能的系统。
      To 3. orb有个问题就是特征分布不均匀。在小图片上还好,大图片上缩在一团,所以不如sift-gms效果好。我记得有试过在cpc上大概36左右。

  3. 请教边老师,关于下载数据集的问题。
    目前在https://1drv.ms/f/s!AiV6XqkxJHE2g3ZC4zYYR05eEY_m上下载不了,因为连接不上。看了上面的提问和回答,是否可以提供百度云链接的地址。谢谢!

Leave a Reply

Your email address will not be published. Required fields are marked *