Journal of Graphics ›› 2023, Vol. 44 ›› Issue (5): 966-977.DOI: 10.11996/JG.j.2095-302X.2023050966
• Image Processing and Computer Vision • Previous Articles Next Articles
CHEN Peng1(), JIANG Hao2, XIANG Wei1(
)
Received:
2023-02-07
Accepted:
2023-06-15
Online:
2023-10-31
Published:
2023-10-31
Contact:
XIANG Wei (1991-), lecturer, Ph.D. His main research interests cover intelligent design, etc. E-mail:About author:
CHEN Peng (1996-), master students. His main research interests cover artificial intelligence and digital image processing. E-mail:chen_peng2023@163.com
CLC Number:
CHEN Peng, JIANG Hao, XIANG Wei. Stereoscopic image generation considering human perception[J]. Journal of Graphics, 2023, 44(5): 966-977.
Add to citation manager EndNote|Ris|BibTeX
URL: http://www.txxb.com.cn/EN/10.11996/JG.j.2095-302X.2023050966
特点 | 样例 | |
---|---|---|
高级 特征 | 识别率:由 模型识别分 割块的类型 和精度 | ![]() |
显著性图: 显示注意力 分布的灰度图 (亮区域表示 浓度高) | ![]() | |
低级 特征 | 颜色:分割 块与其他分 割块之间的 颜色对比的 总和 | ![]() |
布局:线段 中心与图像 中心之间的 距离 | ![]() | |
深度:深度 图中分割块 的平均深度 | ![]() |
Table 1 The features used in the stage of attention calculation
特点 | 样例 | |
---|---|---|
高级 特征 | 识别率:由 模型识别分 割块的类型 和精度 | ![]() |
显著性图: 显示注意力 分布的灰度图 (亮区域表示 浓度高) | ![]() | |
低级 特征 | 颜色:分割 块与其他分 割块之间的 颜色对比的 总和 | ![]() |
布局:线段 中心与图像 中心之间的 距离 | ![]() | |
深度:深度 图中分割块 的平均深度 | ![]() |
Fig. 4 Depth optimization ((a) Example of a depth conflict; (b) Depth distribution of two giraffes in the example (not real data, but provided for explanation); (c) Examples of two operations)
Fig. 5 Visualization of scores ((a) Original color image;(b) Scores based on high-level features; (c) Scores based on low-level features; (d) Scores based on depth information; (e) Attention scores combining all three scores)
Fig. 7 Comparison of depth maps ((a) Original color image; (b) Shows the estimated depth maps generated by[31]; (c) The enhanced depth maps optimized[35]; (d) The manual depth maps; (e) The depth maps optimized by DOCHP)
模型 | P(M) | P(M|data) | BF10 |
---|---|---|---|
Null model (包含被试、图像) | 0.333 | 5.932×10-11 | 1.000 |
生成方法 | 0.333 | 0.510 | 8.598×109 |
生成方法、图像 | 0.333 | 0.490 | 8.260×109 |
Table 2 Model-based analysis
模型 | P(M) | P(M|data) | BF10 |
---|---|---|---|
Null model (包含被试、图像) | 0.333 | 5.932×10-11 | 1.000 |
生成方法 | 0.333 | 0.510 | 8.598×109 |
生成方法、图像 | 0.333 | 0.490 | 8.260×109 |
变量 | 水平 | 均值 | 方差 | 95%置信 度-低 | 95%置信 度-高 |
---|---|---|---|---|---|
生成 方法 | 本文方法 | 0.404 | 0.102 | 0.198 | 0.609 |
手工 | 0.484 | 0.104 | 0.276 | 0.692 | |
估计 | -0.610 | 0.105 | -0.820 | -0.403 | |
传统优化 | -0.278 | 0.101 | -0.484 | -0.080 | |
图 | -0.328 | 0.166 | -0.661 | -0.002 | |
0.240 | 0.162 | -0.080 | 0.564 | ||
-0.307 | 0.163 | -0.634 | 0.010 | ||
0.239 | 0.163 | -0.089 | 0.563 | ||
-0.205 | 0.161 | -0.530 | 0.117 | ||
0.513 | 0.169 | 0.181 | 0.852 | ||
0.314 | 0.165 | -0.014 | 0.638 | ||
-0.329 | 0.164 | -0.670 | -0.014 | ||
-0.108 | 0.162 | -0.433 | 0.211 | ||
-0.028 | 0.161 | -0.352 | 0.283 |
Table 3 Model summary
变量 | 水平 | 均值 | 方差 | 95%置信 度-低 | 95%置信 度-高 |
---|---|---|---|---|---|
生成 方法 | 本文方法 | 0.404 | 0.102 | 0.198 | 0.609 |
手工 | 0.484 | 0.104 | 0.276 | 0.692 | |
估计 | -0.610 | 0.105 | -0.820 | -0.403 | |
传统优化 | -0.278 | 0.101 | -0.484 | -0.080 | |
图 | -0.328 | 0.166 | -0.661 | -0.002 | |
0.240 | 0.162 | -0.080 | 0.564 | ||
-0.307 | 0.163 | -0.634 | 0.010 | ||
0.239 | 0.163 | -0.089 | 0.563 | ||
-0.205 | 0.161 | -0.530 | 0.117 | ||
0.513 | 0.169 | 0.181 | 0.852 | ||
0.314 | 0.165 | -0.014 | 0.638 | ||
-0.329 | 0.164 | -0.670 | -0.014 | ||
-0.108 | 0.162 | -0.433 | 0.211 | ||
-0.028 | 0.161 | -0.352 | 0.283 |
[1] | 李艳莉, 徐若锋. 基于统计特性的DIBR图像的无参考质量评价[J]. 激光与光电子学进展, 2022, 59(8): 228-236. |
LI Y L, XU R F. No-reference image quality assessment of DIBR-synthesized images based on statistical characteristics[J]. Laser & Optoelectronics Progress, 2022, 59(8): 228-236. (in Chinese) | |
[2] | PATIL S, CHARLES P. Review on 2D-to-3D image and video conversion methods[C]// 2015 International Conference on Computing Communication Control and Automation. New York: IEEE Press, 2015: 728-732. |
[3] | 沙浩, 刘越. 基于深度学习的图像本征属性预测方法综述[J]. 图学学报, 2021, 42(3): 385-397. |
SHA H, LIU Y. Review on deep learning based prediction of image intrinsic properties[J]. Journal of Graphics, 2021, 42(3): 385-397. (in Chinese) | |
[4] | LI Z Q, SNAVELY N. MegaDepth: learning single-view depth prediction from Internet photos[C]// 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2018: 2041-2050. |
[5] | DE SILVA D V S X, FERNANDO W A C, WORRALL S T, et al. Just noticeable difference in depth model for stereoscopic 3D displays[C]// 2010 IEEE International Conference on Multimedia and Expo. New York: IEEE Press, 2010: 1219-1224. |
[6] |
TERZIĆ K, HANSARD M. Methods for reducing visual discomfort in stereoscopic 3D: a review[J]. Signal Processing: Image Communication, 2016, 47: 402-416.
DOI URL |
[7] | 牟琦, 张寒, 何志强, 等. 基于深度估计和特征融合的尺度自适应目标跟踪算法[J]. 图学学报, 2021, 42(4): 563-571. |
MU Q, ZHANG H, HE Z Q, et al. Scale adaptive target tracking algorithm based on depth estimation and feature fusion[J]. Journal of Graphics, 2021, 42(4): 563-571. (in Chinese) | |
[8] | JI P, WANG L H, LI D X, et al. An automatic 2D to 3D conversion algorithm using multi-depth cues[C]// 2012 International Conference on Audio, Language and Image Processing. New York: IEEE Press, 2012: 546-550. |
[9] | KONRAD J, WANG M, ISHWAR P. 2D-to-3D image conversion by learning depth from examples[C]// 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops. New York: IEEE Press, 2012: 16-22. |
[10] | 李林鑫. 2D视频到双目3D视频转换关键技术研究[D]. 南京: 东南大学, 2021. |
LI L X. Research on key technologies of 2D video to binocular 3D video conversion[D]. Nanjing: Southeast University, 2021. (in Chinese) | |
[11] | 黄军, 王聪, 刘越, 等. 单目深度估计技术进展综述[J]. 中国图象图形学报, 2019, 24(12): 2081-2097. |
HUANG J, WANG C, LIU Y, et al. The progress of monocular depth estimation technology[J]. Journal of Image and Graphics, 2019, 24(12): 2081-2097. (in Chinese) | |
[12] | LADICKÝ L, SHI J B, POLLEFEYS M. Pulling things out of perspective[C]// 2014 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2014: 89-96. |
[13] | SHI J P, TAO X, XU L, et al. Break Ames room illusion: depth from general single images[J]. ACM Transactions on Graphics, 2015, 34(6): 225: 1-225:11. |
[14] | ZHOU W, SALZMANN M, HE X M, et al. Indoor scene structure analysis for single image depth estimation[C]// 2015 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2015: 614-622. |
[15] |
KARSCH K, LIU C, KANG S B. Depth transfer: depth extraction from video using non-parametric sampling[J]. IEEE transactions on pattern analysis and machine intelligence, 2014, 36(11): 2144-2158.
DOI PMID |
[16] | KIM S, PARK K, SOHN K, et al. Unified depth prediction and intrinsic image decomposition from a single image via joint convolutional neural fields[C]// European Conference on Computer Vision. Cham: Springer, 2016: 143-159. |
[17] | 张晓娜. 基于深度学习的单目图像深度估计[D]. 石家庄: 河北师范大学, 2020. |
ZHANG X N. Deep learning based monocular depth estimation[D]. Shijiazhuang: Hebei Normal University, 2020. (in Chinese) | |
[18] | GEIGER A, LENZ P, URTASUN R. Are we ready for autonomous driving? The KITTI vision benchmark suite[C]// 2012 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2012: 3354-3361. |
[19] | SAXENA A, SUN M, NG A Y. Learning 3-D scene structure from a single still image[C]// 2007 IEEE 11th International Conference on Computer Vision. New York: IEEE Press, 2007: 1-8. |
[20] |
LIN L X, HUANG G H, CHEN Y J, et al. Efficient and high-quality monocular depth estimation via gated multi-scale network[J]. IEEE Access, 2020, 8: 7709-7718.
DOI URL |
[21] |
FU J W, LIANG J, WANG Z Y. Monocular depth estimation based on multi-scale graph convolution networks[J]. IEEE Access, 2019, 8: 997-1009.
DOI URL |
[22] | KUZNIETSOV Y, STÜCKLER J, LEIBE B. Semi-supervised deep learning for monocular depth map prediction[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2017: 2215-2223. |
[23] | GOLDMAN M, HASSNER T, AVIDAN S. Learn stereo, infer mono: Siamese networks for self-supervised, monocular, depth estimation[C]// 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops. New York: IEEE Press, 2020: 2886-2895. |
[24] | CHEN W F, FU Z, YANG D W, et al. Single-image depth perception in the wild[C]// The 30th International Conference on Neural Information Processing Systems. New York: ACM, 2016: 730-738. |
[25] |
JUNG Y J, SOHN H, LEE S I, et al. Visual comfort improvement in stereoscopic 3D displays using perceptually plausible assessment metric of visual comfort[J]. IEEE Transactions on Consumer Electronics, 2014, 60(1): 1-9.
DOI URL |
[26] |
OH C, HAM B, CHOI S, et al. Visual fatigue relaxation for stereoscopic video via nonlinear disparity remapping[J]. IEEE Transactions on Broadcasting, 2015, 61(2): 142-153.
DOI URL |
[27] |
LEI J J, ZHANG C C, FANG Y M, et al. Depth sensation enhancement for multiple virtual view rendering[J]. IEEE Transactions on Multimedia, 2015, 17(4): 457-469.
DOI URL |
[28] |
ISLAM M B, WONG L-K, LOW K-L, et al. Aesthetics-driven stereoscopic 3-D image recomposition with depth adaptation[J]. IEEE Transactions on Multimedia, 2018, 20(11): 2964-2979.
DOI URL |
[29] |
SHAO F, LIN W C, LIN W S, et al. QoE-guided warping for stereoscopic image retargeting[J]. IEEE Transactions on Image Processing: a Publication of the IEEE Signal Processing Society, 2017, 26(10): 4790-4805.
DOI URL |
[30] |
CHUN M M. Contextual cueing of visual attention[J]. Trends in Cognitive Sciences, 2000, 4(5): 170-178.
PMID |
[31] | HOU Q B, CHENG M M, HU X W, et al. Deeply supervised salient object detection with short connections[C]// 2017 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2017: 5300-5309. |
[32] | HE K M, GKIOXARI G, DOLLÁR P, et al. Mask R-CNN[C]// 2017 IEEE International Conference on Computer Vision. New York: IEEE Press, 2017: 2980-2988. |
[33] | YAN Q, XU L, SHI J P, et al. Hierarchical saliency detection[C]// 2013 IEEE Conference on Computer Vision and Pattern Recognition. New York: IEEE Press, 2013: 1155-1162. |
[34] |
HIBBARD P B, HAINES A E, HORNSEY R L. Magnitude, precision, and realism of depth perception in stereoscopic vision[J]. Cognitive Research: Principles and Implications, 2017, 2: 25: 1-25:11.
DOI URL |
[35] |
JUNG S W, KO S J. Depth sensation enhancement using the just noticeable depth difference[J]. IEEE Transactions on Image Processing, 2012, 21(8): 3624-3637.
DOI URL |
[36] | KAY M, NELSON G L, HEKLER E B. Researcher-centered design of statistics: why Bayesian statistics better fit the culture and incentives of HCI[C]// 2016 CHI Conference on Human Factors in Computing Systems. New York: ACM, 2016: 4521-4532. |
[37] |
MOREY R D, ROMEIJN J W, ROUDER J N. The philosophy of Bayes factors and the quantification of statistical evidence[J]. Journal of Mathematical Psychology, 2016, 72: 6-18.
DOI URL |
[38] |
WAGENMAKERS E J, MARSMAN M, JAMIL T, et al. Bayesian inference for psychology. Part I: Theoretical advantages and practical ramifications[J]. Psychonomic Bulletin & Review, 2018, 25(1): 35-57.
DOI URL |
[39] |
SCHÖNBRODT F D, WAGENMAKERS E J. Bayes factor design analysis: planning for compelling evidence[J]. Psychonomic Bulletin & Review, 2018, 25(1): 128-142.
DOI URL |
[40] |
ROUDER J N, MOREY R D, VERHAGEN J, et al. Bayesian analysis of factorial designs[J]. Psychological Methods, 2017, 22(2): 304-321.
DOI PMID |
[1] | YAN Shan-wu, XIAO Hong-bing, WANG Yu, SUN Mei. Video anomaly detection combining pedestrian spatiotemporal information [J]. Journal of Graphics, 2023, 44(1): 95-103. |
[2] | GU Yu, ZHAO Jun. Research on image detection algorithm of freight train brake shoe bolt and brake shoe fault [J]. Journal of Graphics, 2023, 44(1): 88-94. |
[3] | ZHANG Chen-yang, CAO Yan-hua, YANG Xiao-zhong. Multi-focus image fusion method based on fractional wavelet combined with guided filtering [J]. Journal of Graphics, 2023, 44(1): 77-87. |
[4] | SHAO Ying-jie, YIN Hui, XIE Ying, HUANG Hua. A sketch-guided facial image completion network via selective recurrent inference [J]. Journal of Graphics, 2023, 44(1): 67-76. |
[5] | PAN Sen-lei, QIAN Wen-hua, CAO Jin-de, XU Dan. Learning attention for Dongba paintings emotion classification [J]. Journal of Graphics, 2023, 44(1): 59-66. |
[6] | SHAN Fang-mei, WANG Meng-wen, LI Min. Multi-scale convolutional neural network incorporating attention mechanism for intestinal polyp segmentation [J]. Journal of Graphics, 2023, 44(1): 50-58. |
[7] | ZHANG Qian, WANG Xia-li, WANG Wei-hao, WU Li-zhan, LI Chao. Cell counting method based on multi-scale feature fusion [J]. Journal of Graphics, 2023, 44(1): 41-49. |
[8] | SHAO Wen-bin, LIU Yu-jie, SUN Xiao-rui, LI Zong-min. Cross modality person re-identification based on residual enhanced attention [J]. Journal of Graphics, 2023, 44(1): 33-40. |
[9] | PI Jun, LIU Yu-heng, LI Jiu-hao. Research on lightweight forest fire detection algorithm based on YOLOv5s [J]. Journal of Graphics, 2023, 44(1): 26-32. |
[10] | PEI Hui-ning, SHAO Xing-chen, TAN Zhao-yun, HUANG Xue-qin, BAI Zhong-hang. Prediction model of cultural image based on DE-GWO and SVR [J]. Journal of Graphics, 2023, 44(1): 184-193. |
[11] | FAN Zhen, LIU Xiao-jing, LI Xiao-bo, CUI Ya-chao. A homography estimation method robust to illumination and occlusion [J]. Journal of Graphics, 2023, 44(1): 166-176. |
[12] | LI Xiao-bo, LI Yang-gui, GUO Ning, FAN Zhen. Mask detection algorithm based on YOLOv5 integrating attention mechanism [J]. Journal of Graphics, 2023, 44(1): 16-25. |
[13] | LIU Zhen-ye, CHEN Ren-jie, LIU Li-gang. Edge length based 3D shape interpolation [J]. Journal of Graphics, 2023, 44(1): 158-165. |
[14] | WANG Jia-dong, CAO Juan, CHEN Zhong-gui. Feature-preserving skeleton extraction algorithm for point clouds [J]. Journal of Graphics, 2023, 44(1): 146-157. |
[15] | WANG Yu-ping, ZENG Yi, LI Sheng-hui, ZHANG Lei. A Transformer-based 3D human pose estimation method [J]. Journal of Graphics, 2023, 44(1): 139-145. |
Viewed | ||||||
Full text |
|
|||||
Abstract |
|
|||||