Dream-in-Style: Text-to-3D Generation using Stylized Score Distillation



Trinity College Dublin

Abstract

We present a method to generate 3D objects in styles. Our method takes a text prompt and a style reference image as input and reconstructs a neural radiance field to synthesize a 3D model with the content aligning with the text prompt and the style following the reference image. To simultaneously generate the 3D object and perform style transfer in one go, we propose a stylized score distillation loss to guide a text-to-3D optimization process to output visually plausible geometry and appearance. Our stylized score distillation is based on a combination of an original pretrained text-to-image model and its modified sibling with the key and value features of self-attention layers manipulated to inject styles from the reference image. Comparisons with state-of-the-art methods demonstrated the strong visual performance of our method, further supported by the quantitative results from our user study.

Dream-in-Style stylizes text-to-3D generation.

BibTeX

@inproceedings{kompanowski2024dreaminstyle,
    author    = {Hubert Kompanowski and Binh-Son Hua},
    title     = {Dream-in-Style: Text-to-3D Generation using Stylized Score Distillation},
    booktitle = {International Conference on 3D Vision},
    year      = {2025},
}

Acknowledgement

This work was conducted with the financial support of the Research Ireland Centre for Research Training in Digitally Enhanced Reality (d-real) under Grant No. 18/CRT/6224. For the purpose of Open Access, the author has applied a CC BY public copyright licence to any Author Accepted Manuscript version arising from this submission.

This project is supported by Research Ireland under the Research Ireland Frontiers for the Future Programme, award number 22/FFP-P/11522.


D-Real Logo.
SFI Logo.

Website template modified from NeRFies.