Recent advancements in 3D scene editing have been propelled by the rapid development of generative models. Existing methods typically utilize generative models to perform text-guided editing on 3D representations, such as 3D Gaussian Splatting (3DGS). However, these methods are often limited to texture modifications and fail when addressing geometric changes, such as editing a character's head to turn around. Moreover, such methods lack accurate control over the spatial position of editing results, as language struggles to precisely describe the extent of edits. To overcome these limitations, we introduce DYG, an effective 3D drag-based editing method for 3D Gaussian Splatting. It enables users to conveniently specify the desired editing region and the desired dragging direction through the input of 3D masks and pairs of control points, thereby enabling precise control over the extent of editing. DYG integrates the strengths of the implicit triplane representation to establish the geometric scaffold of the editing results, effectively overcoming suboptimal editing outcomes caused by the sparsity of 3DGS in the desired editing regions. Additionally, we incorporate a drag-based Latent Diffusion Model into our method through the proposed Drag-SDS loss function, enabling flexible, multi-view consistent, and fine-grained editing. Extensive experiments demonstrate that DYG conducts effective drag-based editing guided by control point prompts, surpassing other baselines in terms of editing effect and quality, both qualitatively and quantitatively.
近期,3D 场景编辑的进展得益于生成模型的快速发展。现有方法通常利用生成模型对 3D 表示(如 3D Gaussian Splatting,3DGS)进行文本引导编辑。然而,这些方法通常仅限于纹理修改,在处理几何变化时(例如编辑角色的头部使其转向)会失败。此外,这些方法缺乏对编辑结果空间位置的精确控制,因为语言难以准确描述编辑的程度。为克服这些限制,我们提出了 DYG,一种有效的基于拖拽的 3D 编辑方法,适用于 3D Gaussian Splatting。通过输入 3D 掩模和控制点对,DYG 使用户能够方便地指定期望的编辑区域和拖拽方向,从而实现对编辑范围的精确控制。DYG 将隐式三平面表示的优势与编辑结果的几何支架结合起来,有效克服了由于 3DGS 在所需编辑区域的稀疏性所造成的次优编辑效果。此外,我们通过提出的 Drag-SDS 损失函数将基于拖拽的潜在扩散模型集成到方法中,实现灵活的、多视角一致的细粒度编辑。大量实验表明,DYG 通过控制点提示引导的拖拽编辑在编辑效果和质量上均超越了其他基线方法,表现出显著的定性和定量优势。