Dynamic Typography: Bringing Text to Life via Video Diffusion Prior (2404.11614v3)
Abstract: Text animation serves as an expressive medium, transforming static communication into dynamic experiences by infusing words with motion to evoke emotions, emphasize meanings, and construct compelling narratives. Crafting animations that are semantically aware poses significant challenges, demanding expertise in graphic design and animation. We present an automated text animation scheme, termed "Dynamic Typography", which combines two challenging tasks. It deforms letters to convey semantic meaning and infuses them with vibrant movements based on user prompts. Our technique harnesses vector graphics representations and an end-to-end optimization-based framework. This framework employs neural displacement fields to convert letters into base shapes and applies per-frame motion, encouraging coherence with the intended textual concept. Shape preservation techniques and perceptual loss regularization are employed to maintain legibility and structural integrity throughout the animation process. We demonstrate the generalizability of our approach across various text-to-video models and highlight the superiority of our end-to-end methodology over baseline methods, which might comprise separate tasks. Through quantitative and qualitative evaluations, we demonstrate the effectiveness of our framework in generating coherent text animations that faithfully interpret user prompts while maintaining readability. Our code is available at: https://animate-your-word.github.io/demo/.
- Adobe Systems Inc. 1990. Adobe Type 1 Font Format. Addison Wesley Publishing Company.
- Multi-Content GAN for Few-Shot Font Style Transfer. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. https://doi.org/10.1109/cvpr.2018.00789
- C Barber and Hannu Huhdanpaa. 1995. Qhull. The Geometry Center, University of Minnesota.
- Stable video diffusion: Scaling latent video diffusion models to large datasets. arXiv preprint arXiv:2311.15127 (2023).
- Bay-Wei Chang and David Ungar. 1993. Animation: from cartoons to the user interface. In Proceedings of the 6th Annual ACM Symposium on User Interface Software and Technology (Atlanta, Georgia, USA) (UIST ’93). Association for Computing Machinery, New York, NY, USA, 45–55. https://doi.org/10.1145/168642.168647
- Videocrafter1: Open diffusion models for high-quality video generation. arXiv preprint arXiv:2310.19512 (2023).
- Gen-2 contributors. 2023a. Gen-2. https://research.runwayml.com/gen2
- PikaLabs contributors. 2023b. Pikalabs. https://www.pika.art/
- Werner Lemberg David Turner. 2009. FreeType library. Retrieved Mar 19, 2024 from https://freetype.org/
- Boris Delaunay et al. 1934. Sur la sphere vide. Izv. Akad. Nauk SSSR, Otdelenie Matematicheskii i Estestvennyka Nauk 7, 793-800 (1934), 1–2.
- Scalable vector graphics (SVG) 1.0 specification. iuniverse Bloomington.
- SketchPatch. ACM Transactions on Graphics (Dec 2020), 1–14. https://doi.org/10.1145/3414685.3417816
- James D Foley. 1996. Computer graphics: principles and practice. Vol. 12110. Addison-Wesley Professional.
- Kinetic typography. In CHI ’97 extended abstracts on Human factors in computing systems looking to the future - CHI ’97. https://doi.org/10.1145/1120212.1120387
- The kinedit system. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. https://doi.org/10.1145/642611.642677
- Breathing Life Into Sketches Using Text-to-Video Priors. (2023). arXiv:2311.13608 [cs.CV]
- AnimateDiff: Animate Your Personalized Text-to-Image Diffusion Models without Specific Tuning. In The Twelfth International Conference on Learning Representations. https://openreview.net/forum?id=Fx2SbBgcte
- Latent Video Diffusion Models for High-Fidelity Video Generation with Arbitrary Lengths. (Nov 2022).
- Kai Hormann and Günther Greiner. 2000. MIPS: An efficient global parametrization method. Curve and Surface Design: Saint-Malo 1999 (2000), 153–162.
- Word-As-Image for Semantic Typography. ACM Trans. Graph. 42, 4, Article 151 (jul 2023), 11 pages. https://doi.org/10.1145/3592123
- SCFont: Structure-Guided Chinese Font Generation via Deep Stacked Networks. Proceedings of the AAAI Conference on Artificial Intelligence (Sep 2019), 4015–4022. https://doi.org/10.1609/aaai.v33i01.33014015
- Noise-free Score Distillation. In The Twelfth International Conference on Learning Representations. https://openreview.net/forum?id=dlIMcmlAdk
- Using kinetic typography to convey emotion in text-based interpersonal communication. In Proceedings of the 6th Conference on Designing Interactive Systems (University Park, PA, USA) (DIS ’06). Association for Computing Machinery, New York, NY, USA, 41–49. https://doi.org/10.1145/1142405.1142414
- The kinetic typography engine. In Proceedings of the 15th annual ACM symposium on User interface software and technology. https://doi.org/10.1145/571985.571997
- The kinetic typography engine: an extensible system for animating expressive text. In Proceedings of the 15th Annual ACM Symposium on User Interface Software and Technology (Paris, France) (UIST ’02). Association for Computing Machinery, New York, NY, USA, 81–90. https://doi.org/10.1145/571985.571997
- Differentiable vector graphics rasterization for editing and learning. ACM Transactions on Graphics (Dec 2020), 1–15. https://doi.org/10.1145/3414685.3417871
- Evalcrafter: Benchmarking and evaluating large video generation models. arXiv preprint arXiv:2310.11440 (2023).
- A learned representation for scalable vector graphics. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 7930–7939.
- VideoFusion: Decomposed Diffusion Models for High-Quality Video Generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
- X-CLIP: End-to-End Multi-grained Contrastive Learning for Video-Text Retrieval. In Proceedings of the 30th ACM International Conference on Multimedia (¡conf-loc¿, ¡city¿Lisboa¡/city¿, ¡country¿Portugal¡/country¿, ¡/conf-loc¿) (MM ’22). Association for Computing Machinery, New York, NY, USA, 638–647. https://doi.org/10.1145/3503161.3547910
- Intelligent typography: Artistic text style transfer for complex texture and structure. IEEE Transactions on Multimedia (2022).
- DynTypo: Example-Based Dynamic Text Effects Transfer. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/cvpr.2019.00602
- Nerf: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 65, 1 (2021), 99–106.
- Mitsuru Minakuchi and Yutaka Kidawara. 2008. Kinetic typography for ambient displays. In Proceedings of the 2nd international conference on Ubiquitous information management and communication. https://doi.org/10.1145/1352793.1352805
- Mitsuru Minakuchi and Katsumi Tanaka. 2005. Automatic kinetic typography composer. In Proceedings of the 2005 ACM SIGCHI International Conference on Advances in computer entertainment technology. https://doi.org/10.1145/1178477.1178512
- Conditional Image-to-Video Generation with Latent Flow Diffusion Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 18444–18455.
- Codef: Content deformation fields for temporally consistent video processing. arXiv preprint arXiv:2308.07926 (2023).
- Nerfies: Deformable neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision. 5865–5874.
- Laurence Penny. 1996. A History of TrueType. Retrieved Mar 19, 2024 from https://www.truetype-typography.com
- DreamFusion: Text-to-3D using 2D Diffusion. In The Eleventh International Conference on Learning Representations. https://openreview.net/forum?id=FjNys5c7VyY
- Learning transferable visual models from natural language supervision. In International conference on machine learning. PMLR, 8748–8763.
- High-Resolution Image Synthesis with Latent Diffusion Models. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/cvpr52688.2022.01042
- Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling. arXiv preprint arXiv:2401.15977 (2024).
- First Order Motion Model for Image Animation. Neural Information Processing Systems,Neural Information Processing Systems (Jan 2019).
- DS-Fusion: Artistic Typography via Discriminated and Stylized Diffusion. (Mar 2023).
- Zachary Teed and Jia Deng. 2020. Raft: Recurrent all-pairs field transforms for optical flow. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part II 16. Springer, 402–419.
- Modelscope text-to-video technical report. arXiv preprint arXiv:2308.06571 (2023).
- Typography With Decor: Intelligent Text Style Transfer. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.org/10.1109/cvpr.2019.00604
- Videocomposer: Compositional video synthesis with motion controllability. Advances in Neural Information Processing Systems 36 (2024).
- Yizhi Wang and Zhouhui Lian. 2021. DeepVecFont: Synthesizing High-quality Vector Fonts via Dual-modality Learning. ACM Transactions on Graphics 40, 6 (2021), 15 pages. https://doi.org/10.1145/3478513.3480488
- Wakey-Wakey: Animate Text by Mimicking Characters in a GIF. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology. https://doi.org/10.1145/3586183.3606813
- Dynamicrafter: Animating open-domain images with video diffusion priors. arXiv preprint arXiv:2310.12190 (2023).
- Awesome Typography: Statistics-Based Text Effects Transfer. Cornell University - arXiv,Cornell University - arXiv (Nov 2016).
- Shape-Matching GAN++: Scale Controllable Dynamic Artistic Text Style Transfer. IEEE Transactions on Pattern Analysis and Machine Intelligence (Jan 2021), 1–1. https://doi.org/10.1109/tpami.2021.3055211
- The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference on computer vision and pattern recognition. 586–595.
- MagicVideo: Efficient Video Generation With Latent Diffusion Models. (Nov 2022).
- Zichen Liu (34 papers)
- Yihao Meng (2 papers)
- Hao Ouyang (45 papers)
- Yue Yu (343 papers)
- Bolin Zhao (2 papers)
- Daniel Cohen-Or (172 papers)
- Huamin Qu (141 papers)