Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2024, Vol. 45 ›› Issue (4): 834-844.DOI: 10.11996/JG.j.2095-302X.2024040834

• Computer Graphics and Virtual Reality • Previous Articles     Next Articles

A text-driven 3D scene editing method based on key views

ZHANG Ji1,2,3(), CUI Wenshuai1, ZHANG Ronghua1,3(), WANG Wenbin1, LI Yaqi1   

  1. 1. Department of Computer, North China Electric Power University, Baoding Hebei 071003, China
    2. Hebei Key Laboratory of Knowledge Computing for Energy & Power, Baoding Hebei 071003, China
    3. Engineering Research Center of Intelligent Computing for Complex Energy Systems, Ministry of Education, Baoding Hebei 071003, China
  • Received:2024-02-10 Accepted:2024-05-15 Online:2024-08-31 Published:2024-09-03
  • Contact: ZHANG Ronghua
  • About author:First author contact:

    ZHANG Ji (1972-), associate professor, Ph.D. His main research interests cover intelligent information processing, deep learning, image processing, etc. E-mail:72zhangji@163.com

  • Supported by:
    Hebei Provincial Science and Technology Program Funding(22310302D)

Abstract:

The zero-shot image editing method based on denoising diffusion model has made remarkable achievements, and its application to 3D scene editing enables zero-shot text-driven 3D scene editing. However, its 3D editing results are easily affected by the 3D continuity of the diffusion model and over-editing, leading to erroneous editing results. To address these problems, a new text-driven 3D editing method was proposed, which started from the dataset and proposed key view-based data iteration and pixel-based abnormal data masking module. The key view data could guide the editing of a 3D area to minimize the effect of 3D inconsistent data, while the data masking module could filter out anomalies in the 2D input data. Using this method, vivid photo-quality text-driven 3D scene editing effects could be realized. Experiments have demonstrated that compared to some current advanced text-driven 3D scene editing methods, the erroneous editing in the 3D scenes could be greatly reduced, resulting in more vivid and realistic 3D editing effects. In addition, the editing results generated by the method in this paper were more diversified and more efficient.

Key words: diffusion model, text-driven, 3D scene editing, key views, data mask

CLC Number: