A new approach for personalizing text-to-image diffusion models to user needs has been proposed, enabling the synthesis of novel images of a specific subject in different contexts. This method involves fine-tuning a pretrained model with a few input images of the subject and using a unique identifier to embed the subject in the output domain. As a result, fully-novel photorealistic images of the subject can be synthesized in diverse scenes, poses, views, and lighting conditions that are not present in the reference images.
A new approach for personalizing text-to-image diffusion models to user needs has been proposed, enabling the synthesis of novel images of a specific subject in different contexts. This method involves fine-tuning a pretrained model with a few input images of the subject and using a unique identifier to embed the subject in the output domain. As a result, fully-novel photorealistic images of the subject can be synthesized in diverse scenes, poses, views, and lighting conditions that are not present in the reference images.
b703a192d93c2abd814b16e2c55e17127d627fbd
2023-03-22T23:18:13+00:00