Skip to content

Latest commit

 

History

History
5 lines (3 loc) · 2.15 KB

2406.09394.md

File metadata and controls

5 lines (3 loc) · 2.15 KB

WonderWorld: Interactive 3D Scene Generation from a Single Image

We present WonderWorld, a novel framework for interactive 3D scene extrapolation that enables users to explore and shape virtual environments based on a single input image and user-specified text. While significant improvements have been made to the visual quality of scene generation, existing methods are run offline, taking tens of minutes to hours to generate a scene. By leveraging Fast Gaussian Surfels and a guided diffusion-based depth estimation method, WonderWorld generates geometrically consistent extrapolation while significantly reducing computational time. Our framework generates connected and diverse 3D scenes in less than 10 seconds on a single A6000 GPU, enabling real-time user interaction and exploration. We demonstrate the potential of WonderWorld for applications in virtual reality, gaming, and creative design, where users can quickly generate and navigate immersive, potentially infinite virtual worlds from a single image. Our approach represents a significant advancement in interactive 3D scene generation, opening up new possibilities for user-driven content creation and exploration in virtual environments.

我们推出了WonderWorld,这是一个用于交互式3D场景外推的新型框架,使用户能够基于单张输入图像和用户指定的文本探索和塑造虚拟环境。尽管在场景生成的视觉质量上取得了显著提升,但现有方法运行在离线状态,生成一个场景需要花费数十分钟到数小时。通过利用快速高斯表面(Fast Gaussian Surfels)和基于引导扩散的深度估计方法,WonderWorld在大幅减少计算时间的同时,生成几何上一致的场景外推。我们的框架在单个A6000 GPU上不到10秒内生成连通且多样的3D场景,使得用户能够实时互动和探索。我们展示了WonderWorld在虚拟现实、游戏和创意设计等应用中的潜力,用户可以快速生成并导航沉浸式、潜在无限的虚拟世界。我们的方法在交互式3D场景生成方面代表了一个重大进步,为用户驱动的内容创建和虚拟环境中的探索开辟了新的可能性。