Chrome Extension
WeChat Mini Program
Use on ChatGLM

TexSliders: Diffusion-Based Texture Editing in CLIP Space

SIGGRAPH '24 ACM SIGGRAPH 2024 Conference Papers(2024)

Cited 0|Views25
No score
Abstract
Generative models have enabled intuitive image creation and manipulationusing natural language. In particular, diffusion models have recently shownremarkable results for natural image editing. In this work, we propose to applydiffusion techniques to edit textures, a specific class of images that are anessential part of 3D content creation pipelines. We analyze existing editingmethods and show that they are not directly applicable to textures, since theircommon underlying approach, manipulating attention maps, is unsuitable for thetexture domain. To address this, we propose a novel approach that insteadmanipulates CLIP image embeddings to condition the diffusion generation. Wedefine editing directions using simple text prompts (e.g., "aged wood" to "newwood") and map these to CLIP image embedding space using a texture prior, witha sampling-based approach that gives us identity-preserving directions in CLIPspace. To further improve identity preservation, we project these directions toa CLIP subspace that minimizes identity variations resulting from entangledtexture attributes. Our editing pipeline facilitates the creation of arbitrarysliders using natural language prompts only, with no ground-truth annotateddata necessary.
More
Translated text
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined