A Unified Arbitrary Style Transfer Framework via Adaptive Contrastive Learning.

Autor: YUXIN ZHANG, FAN TANG, WEIMING DONG, HAIBIN HUANG, CHONGYANG MA, TONG-YEE LEE, CHANGSHENG XU
Předmět:
Zdroj: ACM Transactions on Graphics; Oct2023, Vol. 42 Issue 5, p1-16, 16p
Abstrakt: This work presents Unified Contrastive Arbitrary Style Transfer (UCAST), a novel style representation learning and transfer framework, that can fit in most existing arbitrary image style transfer models, such as CNN-based, ViT-based, and flow-based methods. As the key component in image style transfer tasks, a suitable style representation is essential to achieve satisfactory results. Existing approaches based on deep neural networks typically use second-order statistics to generate the output. However, these hand-crafted features computed from a single image cannot leverage style information sufficiently, which leads to artifacts such as local distortions and style inconsistency. To address these issues, we learn style representation directly from a large number of images based on contrastive learning by considering the relationships between specific styles and the holistic style distribution. Specifically, we present an adaptive contrastive learning scheme for style transfer by introducing an input-dependent temperature. Our framework consists of three key components: a parallel contrastive learning scheme for style representation and transfer, a domain enhancement (DE) module for effective learning of style distribution, and a generative network for style transfer. Qualitative and quantitative evaluations show the results of our approach are superior to those obtained via state-of-the-art methods. [ABSTRACT FROM AUTHOR]
Databáze: Complementary Index