skip to main content
10.1145/3551626.3564961acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
short-paper

Zero-Shot Font Style Transfer with a Differentiable Renderer

Authors Info & Claims
Published:13 December 2022Publication History

ABSTRACT

Recently, a large-scale language-image multi-modal model, CLIP, has been used to realize language-based image translation in a zero-shot manner without training. In this study, we attempted to generate language-based decorative fonts for font images using CLIP. By the existing image style transfer methods using CLIP, stylized font images are usually only surrounded by decorations, and the characters themselves do not change significantly. On the other hand, in this study, we use CLIP and vector graphics image representation using a differentiable renderer to achieve a style transfer of text images that matches the input text. The experimental results show that the proposed method transfers the style of font images to match the given texts. In addition to text images, we confirmed that the proposed method was also able to transform the style of simple logo patterns based on the given texts.

References

  1. Gantugs Atarsaikhan, Brian Kenji Iwana, and Seiichi Uchida. 2018. Contained Neural Style Transfer for Decorated Logo Generation. In 13th IAPR International Workshop on Document Analysis Systems (2018).Google ScholarGoogle Scholar
  2. Kevin Frans, LB Soros, and Olaf Witkowski. 2021. CLIPDraw: Exploring text-to-drawing synthesis through language-image encoders. arXiv preprint arXiv:2106.14843 (2021).Google ScholarGoogle Scholar
  3. Rinon Gal, Or Patashnik, Haggai Maron, Gal Chechik, and Daniel Cohen-Or. 2021. StyleGAN-NADA: Clip-guided domain adaptation of image generators. arXiv preprint arXiv:2108.00946 (2021).Google ScholarGoogle Scholar
  4. Leon A Gatys, Alexander S Ecker, and Matthias Bethge. 2016. Image style transfer using convolutional neural networks. Proc. of IEEE Computer Vision and Pattern Recognition (2016), 2414--2423.Google ScholarGoogle ScholarCross RefCross Ref
  5. Kwon Gihyun and Ye Jong Chul. 2022. CLIPStyler: Image style transfer with a single text condition. Proc. of IEEE Computer Vision and Pattern Recognition (2022), 18062--18071.Google ScholarGoogle Scholar
  6. Tero Karras, Samuli Laine, , and Timo Aila. 2019. A style-based generator architecture for generative adversarial networks. Proc. of IEEE Computer Vision and Pattern Recognition (2019), 4401--4410.Google ScholarGoogle ScholarCross RefCross Ref
  7. Tzu-Mao Li, Michal Lukáč, Michaël Gharbi, and Jonathan Ragan-Kelley. 2020. Differentiable Vector Graphics Rasterization for Editing and Learning. ACM Trans. Graph. (Proc. SIGGRAPH Asia) 39, 6 (2020), 193:1--193:15.Google ScholarGoogle Scholar
  8. Or Patashnik, Zongze Wu, Eli Shechtman, Daniel Cohen-Or, and Dani Lischinski. 2021. Styleclip: Text-driven manipulation of stylegan imagery. arXiv preprint arXiv:2103.17249 (2021).Google ScholarGoogle Scholar
  9. Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Jack Clark Pamela Mishkin, Gretchen Krueger, and Ilya Sutskever. 2021. Learning transferable visual models from natural language supervision. arXiv preprint arXiv:2103.00020 (2021).Google ScholarGoogle Scholar
  10. Peter Schaldenbrand, Zhixuan Liu, and Jean Oh. 2021. StyleCLIPDraw: Coupling Content and Style in Text-to-Drawing Translation. arXiv preprint arXiv:2202.12362 (2021).Google ScholarGoogle Scholar
  11. Wenjing Wang, Jiaying Liu, Shuai Yang, and Zongming Guo. 2019. Typography with decor: Intelligent text style transfer. Proc. of IEEE Computer Vision and Pattern Recognition (2019), 5889--5897.Google ScholarGoogle ScholarCross RefCross Ref
  12. Shuai Yang, Zhangyang Wang, Zhaowen Wang, Ning Xu, Jiaying Liu, and Zongming Guo. 2019. Controllable artistic text style transfer via shape-matching GAN. Proc. of IEEE International Conference on Computer Vision (2019), 4442--4451.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Zero-Shot Font Style Transfer with a Differentiable Renderer

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      MMAsia '22: Proceedings of the 4th ACM International Conference on Multimedia in Asia
      December 2022
      296 pages
      ISBN:9781450394789
      DOI:10.1145/3551626

      Copyright © 2022 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 13 December 2022

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • short-paper

      Acceptance Rates

      Overall Acceptance Rate59of204submissions,29%

      Upcoming Conference

      MM '24
      MM '24: The 32nd ACM International Conference on Multimedia
      October 28 - November 1, 2024
      Melbourne , VIC , Australia
    • Article Metrics

      • Downloads (Last 12 months)107
      • Downloads (Last 6 weeks)6

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader