Meta 3D Gen

Raphael Bensadoun, Tom Monnier, Yanir Kleiman,Filippos Kokkinos,Yawar Siddiqui, Mahendra Kariya, Omri Harosh,Roman Shapovalov,Benjamin Graham, Emilien Garreau,Animesh Karnewar,Ang Cao, Idan Azuri,Iurii Makarov, Eric-Tuan Le, Antoine Toisoul,David Novotny,Oran Gafni,Natalia Neverova,Andrea Vedaldi

arxiv(2024)

引用 0|浏览6
暂无评分
摘要
We introduce Meta 3D Gen (3DGen), a new state-of-the-art, fast pipeline for text-to-3D asset generation. 3DGen offers 3D asset creation with high prompt fidelity and high-quality 3D shapes and textures in under a minute. It supports physically-based rendering (PBR), necessary for 3D asset relighting in real-world applications. Additionally, 3DGen supports generative retexturing of previously generated (or artist-created) 3D shapes using additional textual inputs provided by the user. 3DGen integrates key technical components, Meta 3D AssetGen and Meta 3D TextureGen, that we developed for text-to-3D and text-to-texture generation, respectively. By combining their strengths, 3DGen represents 3D objects simultaneously in three ways: in view space, in volumetric space, and in UV (or texture) space. The integration of these two techniques achieves a win rate of 68 We compare 3DGen to numerous industry baselines, and show that it outperforms them in terms of prompt fidelity and visual quality for complex textual prompts, while being significantly faster.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要