Journal of Graphics ›› 2023, Vol. 44 ›› Issue (4): 739-746.DOI: 10.11996/JG.j.2095-302X.2023040739
Previous Articles Next Articles
GUO Yin-hong(), WANG Li-chun(
), LI Shuang
Received:
2022-11-28
Accepted:
2023-04-06
Online:
2023-08-31
Published:
2023-08-16
Contact:
Wang Li-chun (1975-), professo, Ph.D. Her main research interests cover computer vision and human-computer interaction, etc. E-mail:About author:
GUO Yin-hong (1997-), master student. His main research interest covers computer vision. E-mail:gyh20200216@163.com
Supported by:
CLC Number:
GUO Yin-hong, WANG Li-chun, LI Shuang. Image feature matching based on repeatability and specificity constraints[J]. Journal of Graphics, 2023, 44(4): 739-746.
Add to citation manager EndNote|Ris|BibTeX
URL: http://www.txxb.com.cn/EN/10.11996/JG.j.2095-302X.2023040739
类别 | 方法 | 单应性估计AUC | ||
---|---|---|---|---|
@3px | @5px | @10px | ||
有检测器 | D2Net+NN | 23.2 | 35.9 | 53.6 |
R2D2+NN | 50.6 | 63.9 | 76.8 | |
DISK+NN | 52.3 | 64.9 | 78.9 | |
SP+ SuperGlue | 53.9 | 68.3 | 81.7 | |
无检测器 | DRC-Net | 50.6 | 56.2 | 68.3 |
LoFTR | 65.9 | 75.6 | 84.6 | |
Ours | 66.8 | 76.9 | 86.1 |
Table 1 Homography estimation on HPatches
类别 | 方法 | 单应性估计AUC | ||
---|---|---|---|---|
@3px | @5px | @10px | ||
有检测器 | D2Net+NN | 23.2 | 35.9 | 53.6 |
R2D2+NN | 50.6 | 63.9 | 76.8 | |
DISK+NN | 52.3 | 64.9 | 78.9 | |
SP+ SuperGlue | 53.9 | 68.3 | 81.7 | |
无检测器 | DRC-Net | 50.6 | 56.2 | 68.3 |
LoFTR | 65.9 | 75.6 | 84.6 | |
Ours | 66.8 | 76.9 | 86.1 |
类别 | 方法 | 位姿估计AUC | ||
---|---|---|---|---|
@5° | @10° | @20° | ||
有检测器 | SP[ | 16.16 | 33.81 | 51.84 |
无检测器 | DRC-Net | 7.69 | 17.93 | 30.49 |
LoFTR | 22.06 | 40.80 | 57.96 | |
Ours | 22.87 | 41.75 | 59.10 |
Table 2 Relative pose estimation on indoor dataset ScanNet
类别 | 方法 | 位姿估计AUC | ||
---|---|---|---|---|
@5° | @10° | @20° | ||
有检测器 | SP[ | 16.16 | 33.81 | 51.84 |
无检测器 | DRC-Net | 7.69 | 17.93 | 30.49 |
LoFTR | 22.06 | 40.80 | 57.96 | |
Ours | 22.87 | 41.75 | 59.10 |
类别 | 方法 | 位姿估计AUC | ||
---|---|---|---|---|
@5° | @10° | @20° | ||
有检测器 | SP+Superglue | 42.18 | 61.16 | 75.96 |
无检测器 | DRC-Net | 27.01 | 42.96 | 58.31 |
LoFTR | 52.81 | 69.19 | 81.18 | |
Ours | 53.63 | 70.20 | 83.56 |
Table 3 Relative pose estimation on outdoor dataset MegaDepth
类别 | 方法 | 位姿估计AUC | ||
---|---|---|---|---|
@5° | @10° | @20° | ||
有检测器 | SP+Superglue | 42.18 | 61.16 | 75.96 |
无检测器 | DRC-Net | 27.01 | 42.96 | 58.31 |
LoFTR | 52.81 | 69.19 | 81.18 | |
Ours | 53.63 | 70.20 | 83.56 |
方法 | 模糊核 | 位姿估计AUC | ||
---|---|---|---|---|
@5° | @10° | @20° | ||
LoFTR | 5×5 | 40.63 | 56.70 | 70.53 |
Ours | 5×5 | 44.60 | 63.50 | 76.52 |
LoFTR | 12×12 | 32.37 | 47.32 | 61.68 |
Ours | 12×12 | 41.10 | 59.5 | 70.63 |
LoFTR | 24×24 | 18.86 | 31.86 | 47.18 |
Ours | 24×24 | 32.68 | 45.20 | 57.24 |
Table 4 Comparison of pose estimation using images with different blurriness
方法 | 模糊核 | 位姿估计AUC | ||
---|---|---|---|---|
@5° | @10° | @20° | ||
LoFTR | 5×5 | 40.63 | 56.70 | 70.53 |
Ours | 5×5 | 44.60 | 63.50 | 76.52 |
LoFTR | 12×12 | 32.37 | 47.32 | 61.68 |
Ours | 12×12 | 41.10 | 59.5 | 70.63 |
LoFTR | 24×24 | 18.86 | 31.86 | 47.18 |
Ours | 24×24 | 32.68 | 45.20 | 57.24 |
数据集 | 图像 重建 | 重复性和 特异性约束 | 位姿估计AUC | ||
---|---|---|---|---|---|
@5° | @10° | @20° | |||
MegaDepth | √ | √ | 53.63 | 70.20 | 83.56 |
√ | - | 52.88 | 69.30 | 81.18 | |
- | √ | 53.58 | 70.16 | 83.47 | |
- | - | 52.81 | 69.19 | 81.18 | |
MegaDepth-B (模糊核12×12) | √ | √ | 41.10 | 59.5 | 70.63 |
√ | - | 40.56 | 58.4 | 69.03 | |
- | √ | 33.96 | 49.12 | 63.98 | |
- | - | 32.37 | 47.32 | 61.68 |
Table 5 Ablation Experiment
数据集 | 图像 重建 | 重复性和 特异性约束 | 位姿估计AUC | ||
---|---|---|---|---|---|
@5° | @10° | @20° | |||
MegaDepth | √ | √ | 53.63 | 70.20 | 83.56 |
√ | - | 52.88 | 69.30 | 81.18 | |
- | √ | 53.58 | 70.16 | 83.47 | |
- | - | 52.81 | 69.19 | 81.18 | |
MegaDepth-B (模糊核12×12) | √ | √ | 41.10 | 59.5 | 70.63 |
√ | - | 40.56 | 58.4 | 69.03 | |
- | √ | 33.96 | 49.12 | 63.98 | |
- | - | 32.37 | 47.32 | 61.68 |
Fig. 4 Visualization results on MegaDepth dataset ((a) Clear images in MegaDepth; (b) Moderate blurring (blurring kernel 12×12); (c) Poor image quality and high blurring (blurring kernel 24×24))
[1] | 吴凡, 宗艳桃, 汤霞清. 视觉SLAM的研究现状与展望[J]. 计算机应用研究, 2020, 37(8): 2248-2254. |
WU F, ZONG Y T, TANG X Q. Research status and prospect of vision SLAM[J]. Application Research of Computers, 2020, 37(8): 2248-2254 (in Chinese). | |
[2] | SUN J M, SHEN Z H, WANG Y A, et al. LoFTR: detector-free local feature matching with transformers[C]// 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2021: 8922-8931. |
[3] | KATHAROPOULOS A, VYAS A, PAPPAS N, et al. Transformers are RNNs: fast autoregressive transformers with linear attention[EB/OL]. [2022-05-11]. https://arxiv.org/abs/2006.16236. |
[4] |
LOWE D G. Distinctive image features from scale-invariant keypoints[J]. International Journal of Computer Vision, 2004, 60(2): 91-110.
DOI URL |
[5] | RUBLEE E, RABAUD V, KONOLIGE K, et al. ORB: an efficient alternative to SIFT or SURF[C]// 2011 International Conference on Computer Vision. New York: IEEE Press, 2011: 2564-2571. |
[6] | ROSTEN E, DRUMMOND T. Machine learning for high-speed corner detection[C]// The 9th European Conference on Computer Vision - Volume Part I. New York: ACM, 2006: 430-443. |
[7] | CALONDER M, LEPETIT V, STRECHA C, et al. Brief: binary robust independent elementary features[C]// European Conference on Computer Vision. Heidelberg: Springer, 2010: 778-792. |
[8] |
MUR-ARTAL R, TARDÓS J D. ORB-SLAM2: an open-source SLAM system for monocular, stereo, and RGB-D cameras[J]. IEEE Transactions on Robotics, 2017, 33(5): 1255-1262.
DOI URL |
[9] | YI K M, TRULLS E, LEPETIT V, et al. LIFT: Learned Invariant Feature Transform[C]// European Conference on Computer Vision. Cham: Springer International Publishing, 2016: 467-483. |
[10] | ETONE D, MALISIEWICZ T, RABINOVICH A. Toward geometric deep SLAM[EB/OL]. [2022-05-16]. https://arxiv.org/abs/1707.07410. |
[11] | DETONE D, MALISIEWICZ T, RABINOVICH A. SuperPoint: self-supervised interest point detection and description[C]// 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. New York: IEEE Press, 2018: 337:1-337:12. |
[12] | SARLIN P E, DETONE D, MALISIEWICZ T, et al. Superglue: learning feature matching with graph neural networks[C]// 2020 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2020: 4938-4947. |
[13] | 白铂, 刘玉婷, 马驰骋, 等. 图神经网络[J]. 中国科学: 数学, 2020, 3: 367-384. |
BAI B, LIU Y T, MA C C, et al. Graph neural network[J]. Science in China: Mathematics, 2020, 3: 367-384 (in Chinese). | |
[14] | ROCCO I, ARANDJELOVIĆ R, SIVIC J. Efficient Neighbourhood Consensus Networks via Submanifold Sparse Convolutions[C]// European Conference on Computer Vision. Cham: Springer International Publishing, 2020: 605-621. |
[15] | YANG G, RAMANAN D. Volumetric correspondence networks for optical flow[C]// Advances in Neural Information Processing Systems. Cambridge: MIT Press, 2019: 794-805. |
[16] | LIN T Y, DOLLÁR P, GIRSHICK R, et al. Feature pyramid networks for object detection[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2017: 2117-2125. |
[17] | HE K M, ZHANG X Y, REN S Q, et al. Deep residual learning for image recognition[C]// 2016 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2016: 770-778. |
[18] | ROCCO I, CIMPOI M, ARANDJELOVIĆ R, et al. Neighbourhood consensus networks[EB/OL]. [2022-05-16]. https://arxiv.org/abs/1810.10510. |
[19] | TYSZKIEWICZ M J, FUA P, TRULLS E. DISK: learning local features with policy gradient[C]// The 34th International Conference on Neural Information Processing Systems. New York: ACM, 2020: 14254-14265. |
[20] | BALNTAS V, LENC K, VEDALDI A, et al. HPatches: a benchmark and evaluation of handcrafted and learned local descriptors[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2017: 5173-5182. |
[21] | LI Z Q, SNAVELY N. MegaDepth: learning single-view depth prediction from Internet photos[C]// 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2018: 2041-2050. |
[22] | REVAUD J, WEINZAEPFEL P, DE SOUZA C, et al. R2D2: repeatable and reliable detector and descriptor[EB/OL]. [2022-05-16]. https://arxiv.org/abs/1906.06195. |
[23] | DUSMANU M, ROCCO I, PAJDLA T, et al. D2-net: a trainable cnn for joint detection and description of local features[EB/OL]. [2022-05-16]. https://arxiv.org/abs1905.03561. |
[24] | TYSZKIEWICZ M J, FUA P, TRULLS E. DISK: learning local features with policy gradient[C]// The 34th International Conference on Neural Information Processing Systems. New York: ACM, 2020: 14254-14265. |
[25] | LI X H, HAN K, LI S D, et al. Dual-resolution correspondence networks[C]// The 34th International Conference on Neural Information Processing Systems. New York: ACM, 2020: 17346-17357. |
[26] | SCHÖNBERGER J L, FRAHM J M. Structure-from-motion revisited[C]// 2016 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2016: 4104-4113. |
[1] |
YANG Chen-cheng, DONG Xiu-cheng, HOU Bing, ZHANG Dang-cheng, XIANG Xian-ming, FENG Qi-ming.
Reference based transformer texture migrates depth imagessuper resolution reconstruction
[J]. Journal of Graphics, 2023, 44(5): 861-867.
|
[2] |
DANG Hong-she, XU Huai-biao, ZHANG Xuan-de.
Deep learning stereo matching algorithm fusing structural information
[J]. Journal of Graphics, 2023, 44(5): 899-906.
|
[3] |
ZHAI Yong-jie, GUO Cong-bin, WANG Qian-ming, ZHAO Kuan, BAI Yun-shan, ZHANG Ji.
Multi-fitting detection method for transmission lines based onimplicit spatial knowledge fusion
[J]. Journal of Graphics, 2023, 44(5): 918-927.
|
[4] |
YANG Hong-ju, GAO Min, ZHANG Chang-you, BO Wen, WU Wen-jia, CAO Fu-yuan.
A local optimization generation model for image inpainting
[J]. Journal of Graphics, 2023, 44(5): 955-965.
|
[5] | BI Chun-yan, LIU Yue. A survey of video human action recognition based on deep learning [J]. Journal of Graphics, 2023, 44(4): 625-639. |
[6] | CAO Yi-qin, ZHOU Yi-wei, XU Lu. A real-time metallic surface defect detection algorithm based on E-YOLOX [J]. Journal of Graphics, 2023, 44(4): 677-690. |
[7] | SHAO Jun-qi, QIAN Wen-hua, XU Qi-hao. Landscape image generation based on conditional residual generative adversarial network [J]. Journal of Graphics, 2023, 44(4): 710-717. |
[8] | YU Wei-qun, LIU Jia-tao, ZHANG Ya-ping. Monocular depth estimation based on Laplacian pyramid with attention fusion [J]. Journal of Graphics, 2023, 44(4): 728-738. |
[9] | MAO Ai-kun, LIU Xin-ming, CHEN Wen-zhuang, SONG Shao-lou. Improved substation instrument target detection method for YOLOv5 algorithm [J]. Journal of Graphics, 2023, 44(3): 448-455. |
[10] | WANG Jia-jing, WANG Chen, ZHU Yuan-yuan, WANG Xiao-mei. Graph element detection matching based on Republic of China banknotes [J]. Journal of Graphics, 2023, 44(3): 492-501. |
[11] | YANG Liu, WU Xiao-qun. 3D shape completion via deep learning: a method survey [J]. Journal of Graphics, 2023, 44(2): 201-215. |
[12] | ZENG Wu, ZHU Heng-liang, XING Shu-li, LIN Jiang-hong, MAO Guo-jun. Saliency detection-guided for image data augmentation [J]. Journal of Graphics, 2023, 44(2): 260-270. |
[13] | LUO Qi-ming, WU Hao, XIA Xin, YUAN Guo-wu. Prediction of damaged areas in Yunnan murals using Dual Dense U-Net [J]. Journal of Graphics, 2023, 44(2): 304-312. |
[14] | LI Hong-an, ZHENG Qiao-xue, TAO Ruo-lin, ZHANG Min, LI Zhan-li, KANG Bao-sheng. Review of image super-resolutionbased on deep learning [J]. Journal of Graphics, 2023, 44(1): 1-15. |
[15] | SHAO Ying-jie, YIN Hui, XIE Ying, HUANG Hua. A sketch-guided facial image completion network via selective recurrent inference [J]. Journal of Graphics, 2023, 44(1): 67-76. |
Viewed | ||||||
Full text |
|
|||||
Abstract |
|
|||||