Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multimodal Style Transfer via Graph Cuts (1904.04443v6)

Published 9 Apr 2019 in cs.CV

Abstract: An assumption widely used in recent neural style transfer methods is that image styles can be described by global statics of deep features like Gram or covariance matrices. Alternative approaches have represented styles by decomposing them into local pixel or neural patches. Despite the recent progress, most existing methods treat the semantic patterns of style image uniformly, resulting unpleasing results on complex styles. In this paper, we introduce a more flexible and general universal style transfer technique: multimodal style transfer (MST). MST explicitly considers the matching of semantic patterns in content and style images. Specifically, the style image features are clustered into sub-style components, which are matched with local content features under a graph cut formulation. A reconstruction network is trained to transfer each sub-style and render the final stylized result. We also generalize MST to improve some existing methods. Extensive experiments demonstrate the superior effectiveness, robustness, and flexibility of MST.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Yulun Zhang (167 papers)
  2. Chen Fang (157 papers)
  3. Yilin Wang (156 papers)
  4. Zhaowen Wang (55 papers)
  5. Zhe Lin (163 papers)
  6. Yun Fu (131 papers)
  7. Jimei Yang (58 papers)
Citations (78)

Summary

We haven't generated a summary for this paper yet.