Paper Digest: Recent Papers on Style Transfer
Paper Digest Team extracted all recent Style Transfer related papers on our radar, and generated highlight sentences for them. The results are then sorted by relevance & date. In addition to this ‘static’ page, we also provide a real-time version of this article, which has more coverage and is updated in real time to include the most recent updates on this topic.
Based in New York, Paper Digest is dedicated to helping people generate contents & reason over unstructured data. Different from black-box approaches, we build deep models on semantics, which allows results to be produced with explainations. Such models power this website, and are behind our services including “search engine”, “summarization”, “question answering”, and “literature review”.
If you do not want to miss interesting academic papers, you are welcome to sign up our daily paper digest service to get updates on new papers published in your area every day. You are also welcome to follow us on Twitter and Linkedin to get updated with new conference digests.
Paper Digest Team
New York City, New York, 10017
team@paperdigest.org
TABLE 1: Paper Digest: Recent Papers on Style Transfer
Paper | Author(s) | Source | Date | |
---|---|---|---|---|
1 | SAFL-Net: Semantic-Agnostic Feature Learning Network with Auxiliary Plugins for Image Manipulation Detection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we propose SAFL-Net, which constrains a feature extractor to learn semantic-agnostic features by designing specific modules with corresponding auxiliary tasks. |
Zhihao Sun; Haoran Jiang; Danding Wang; Xirong Li; Juan Cao; | iccv | 2023-09-27 |
2 | Synthetic Latent Fingerprint Generation Using Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a simple and effective approach using style transfer and image blending to synthesize realistic latent fingerprints. |
Amol S. Joshi; Ali Dabouei; Nasser Nasrabadi; Jeremy Dawson; | arxiv-cs.CV | 2023-09-27 |
3 | StyleInV: A Temporal Style Modulated Inversion Network for Unconditional Video Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce a novel motion generator design that uses a learning-based inversion network for GAN. |
Yuhan Wang; Liming Jiang; Chen Change Loy; | iccv | 2023-09-27 |
4 | StyleDiffusion: Controllable Disentangled Style Transfer Via Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a new C-S disentangled framework for style transfer without using previous assumptions. |
Zhizhong Wang; Lei Zhao; Wei Xing; | iccv | 2023-09-27 |
5 | Scenimefy: Learning to Craft Anime Scene Via Semi-Supervised Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Despite promising attempts, previous efforts are still incompetent in achieving satisfactory results with consistent semantic preservation, evident stylization, and fine details. In this study, we propose Scenimefy, a novel semi-supervised image-to-image translation framework that addresses these challenges. |
Yuxin Jiang; Liming Jiang; Shuai Yang; Chen Change Loy; | iccv | 2023-09-27 |
6 | Bidirectionally Deformable Motion Modulation For Video-based Human Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Considering the difficulties in transferring highly structural patterns on the garments and discontinuous poses, existing methods often generate unsatisfactory results such as distorted textures and flickering artifacts. To address these issues, we propose a novel Deformable Motion Modulation (DMM) that utilizes geometric kernel offset with adaptive weight modulation to simultaneously perform feature alignment and style transfer. |
WING-YIN YU et. al. | iccv | 2023-09-27 |
7 | Cross-modal Latent Space Alignment for Image to Avatar Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a novel method for automatic vectorized avatar generation from a single portrait image. |
MANUEL LADRON DE GUEVARA et. al. | iccv | 2023-09-27 |
8 | Zero-Shot Contrastive Loss for Text-Guided Diffusion Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Existing methods require computationally expensive fine-tuning of diffusion models or additional neural network. To address this, here we propose a zero-shot contrastive loss for diffusion models that doesn’t require additional fine-tuning or auxiliary networks. |
Serin Yang; Hyunmin Hwang; Jong Chul Ye; | iccv | 2023-09-27 |
9 | StylerDALLE: Language-Guided Style Transfer Using A Vector-Quantized Tokenizer of A Large-Scale Generative Model Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, these abstract semantics can be captured by models like DALL-E or CLIP, which have been trained using huge datasets of images and textual documents. In this paper, we propose StylerDALLE, a style transfer method that exploits both of these models and uses natural language to describe abstract art styles. |
Zipeng Xu; Enver Sangineto; Nicu Sebe; | iccv | 2023-09-27 |
10 | WaveIPT: Joint Attention and Flow Alignment in The Wavelet Domain for Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To leverage the advantages of both attention and flow simultaneously, we propose Wavelet-aware Image-based Pose Transfer (WaveIPT) to fuse the attention and flow in the wavelet domain. |
Liyuan Ma; Tingwei Gao; Haitian Jiang; Haibin Shen; Kejie Huang; | iccv | 2023-09-27 |
11 | AesPA-Net: Aesthetic Pattern-Aware Style Transfer Networks Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce a novel metric, namely pattern repeatability, that quantifies the repetition of patterns in the style image. |
KIBEOM HONG et. al. | iccv | 2023-09-27 |
12 | Not All Steps Are Created Equal: Selective Diffusion Distillation for Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a novel framework, Selective Diffusion Distillation (SDD), that ensures both the fidelity and editability of images. |
Luozhou Wang; Shuai Yang; Shu Liu; Ying-cong Chen; | iccv | 2023-09-27 |
13 | Collecting The Puzzle Pieces: Disentangled Self-Driven Human Pose Transfer By Permuting Textures Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose Pose Transfer by Permuting Textures, a self-driven human pose transfer approach that disentangles pose from texture at the patch-level. |
Nannan Li; Kevin J Shih; Bryan A. Plummer; | iccv | 2023-09-27 |
14 | Frequency-aware GAN for Adversarial Manipulation Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we design an Adversarial Manipulation Generation (AMG) task to explore the vulnerability of image manipulation detectors. |
Peifei Zhu; Genki Osada; Hirokatsu Kataoka; Tsubasa Takahashi; | iccv | 2023-09-27 |
15 | In-Style: Bridging Text and Uncurated Videos with Style Transfer for Text-Video Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To this end, we propose an approach, In-Style, that learns the style of the text queries and transfers it to uncurated web videos. |
Nina Shvetsova; Anna Kukleva; Bernt Schiele; Hilde Kuehne; | iccv | 2023-09-27 |
16 | Towards Generic Image Manipulation Detection with Weakly-Supervised Self-Consistency Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Despite the success of recent learning-based approaches for image manipulation detection, they typically require expensive pixel-level annotations to train, while exhibiting degraded performance when testing on images that are differently manipulated compared with training images. To address these limitations, we propose weakly-supervised image manipulation detection, such that only binary image-level labels (authentic or tampered with) are required for training purpose. |
Yuanhao Zhai; Tianyu Luan; David Doermann; Junsong Yuan; | iccv | 2023-09-27 |
17 | Locally Stylized Neural Radiance Fields Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a stylization framework for NeRF based on local style transfer. |
Hong-Wing Pang; Binh-Son Hua; Sai-Kit Yeung; | iccv | 2023-09-27 |
18 | UMFuse: Unified Multi View Fusion for Human Editing Applications Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we explore the utilization of multiple views to minimize the issue of missing information and generate an accurate representation of the underlying human model. |
RISHABH JAIN et. al. | iccv | 2023-09-27 |
19 | General Image-to-Image Translation with One-Shot Image Guidance Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Current methods are inadequate in meeting this demand as they lack the ability to preserve content or translate visual concepts effectively. Inspired by this, we propose a novel framework named visual concept translator (VCT) with the ability to preserve content in the source image and translate the visual concepts guided by a single reference image. |
Bin Cheng; Zuhao Liu; Yunbo Peng; Yue Lin; | iccv | 2023-09-27 |
20 | Two Birds, One Stone: A Unified Framework for Joint Learning of Image and Video Style Transfers Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In order to achieve satisfying image and video style transfers, two different models are inevitably required with separate training processes on image and video domains, respectively. In this paper, we show that this can be precluded by introducing UniST, a Unified Style Transfer framework for both images and videos. |
Bohai Gu; Heng Fan; Libo Zhang; | iccv | 2023-09-27 |
21 | Incorporating Ensemble and Transfer Learning For An End-To-End Auto-Colorized Image Detection Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a novel approach that combines the advantages of transfer and ensemble learning approaches to help reduce training time and resource requirements while proposing a model to classify natural color and computer-colorized images. |
Ahmed Samir Ragab; Shereen Aly Taie; Howida Youssry Abdelnaby; | arxiv-cs.CV | 2023-09-25 |
22 | MM-NeRF: Multimodal-Guided 3D Multi-Style Transfer of Neural Radiance Field Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we reveal that the same objects in 3D scenes show various states (color tone, details, etc.) from different views after stylization since previous methods optimized by single-view image-based style loss functions, leading NeRF to tend to smooth texture details, further resulting in low-quality rendering. |
Zijiang Yang; Zhongwei Qiu; Chang Xu; Dongmei Fu; | arxiv-cs.CV | 2023-09-24 |
23 | MOSAIC: Multi-Object Segmented Arbitrary Stylization Using CLIP Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: On the other hand, diffusion style transfer methods also suffer from the same issue because the regional stylization control over the stylized output is ineffective. To address this problem, We propose a new method Multi-Object Segmented Arbitrary Stylization Using CLIP (MOSAIC), that can apply styles to different objects in the image based on the context extracted from the input prompt. |
PRAJWAL GANUGULA et. al. | arxiv-cs.CV | 2023-09-24 |
24 | Portrait Stylization: Artistic Style Transfer with Auxiliary Networks for Human Face Stylization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes the use of embeddings from an auxiliary pre-trained face recognition model to encourage the algorithm to propagate human face features from the content image to the final stylized result. |
Thiago Ambiel; | arxiv-cs.CV | 2023-09-23 |
25 | Masked Discriminators for Content-Consistent Unpaired Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we show that masking the inputs of a global discriminator for both domains with a content-based mask is sufficient to reduce content inconsistencies significantly. |
Bonifaz Stuhr; Jürgen Brauer; Bernhard Schick; Jordi Gonzàlez; | arxiv-cs.CV | 2023-09-22 |
26 | TextCLIP: Text-Guided Face Image Generation And Manipulation Without Adversarial Training Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose TextCLIP, a unified framework for text-guided image generation and manipulation without adversarial training. |
Xiaozhou You; Jian Zhang; | arxiv-cs.CV | 2023-09-21 |
27 | Retinex-guided Channel-grouping Based Patch Swap for Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Since the finite features harvested from one single aesthetic style image are inadequate to represent the rich textures of the content natural image, existing techniques treat the full-channel style feature patches as simple signal tensors and create new style feature patches via signal-level fusion, which ignore the implicit diversities existed in style features and thus fail for generating better stylised results. In this paper, we propose a Retinex theory guided, channel-grouping based patch swap technique to solve the above challenges. |
Chang Liu; Yi Niu; Mingming Ma; Fu Li; Guangming Shi; | arxiv-cs.CV | 2023-09-19 |
28 | Speech-to-Speech Translation with Discrete-Unit-Based Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose an S2ST framework with an acoustic language model based on discrete units from a self-supervised model and a neural codec for style transfer. |
YONGQI WANG et. al. | arxiv-cs.SD | 2023-09-14 |
29 | Nucleus-aware Self-supervised Pretraining Using Unpaired Image-to-image Translation for Histopathology Images Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a novel nucleus-aware self-supervised pretraining framework for histopathology images. |
ZHIYUN SONG et. al. | arxiv-cs.CV | 2023-09-13 |
30 | DreamStyler: Paint By Style Inversion with Text-to-Image Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To this end, we introduce DreamStyler, a novel framework designed for artistic image synthesis, proficient in both text-to-image synthesis and style transfer. |
NAMHYUK AHN et. al. | arxiv-cs.CV | 2023-09-13 |
31 | TSSAT: Two-Stage Statistics-Aware Transformation for Artistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Consequently, the stylization results either fail to capture abundant and diversified local style patterns, or contain undesired semantic information of the style image and deviate from the global style distribution. To address this issue, we imitate the drawing process of humans and propose a Two-Stage Statistics-Aware Transformation (TSSAT) module, which first builds the global style foundation by aligning the global statistics of content and style features and then further enriches local style details by swapping the local statistics (instead of local features) in a patch-wise manner, significantly improving the stylization effects. |
Haibo Chen; Lei Zhao; Jun Li; Jian Yang; | arxiv-cs.CV | 2023-09-12 |
32 | PAI-Diffusion: Constructing and Serving A Family of Open Chinese Diffusion Models for Text-to-image Synthesis on The Cloud Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: While existing diffusion models have shown promise in generating images from textual descriptions, they often neglect domain-specific contexts and lack robustness in handling the Chinese language. This paper introduces PAI-Diffusion, a comprehensive framework that addresses these limitations. |
CHENGYU WANG et. al. | arxiv-cs.CL | 2023-09-11 |
33 | MoEController: Instruction-based Arbitrary Image Manipulation with Mixture-of-Expert Controllers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a method with a mixture-of-expert (MOE) controllers to align the text-guided capacity of diffusion models with different kinds of human instructions, enabling our model to handle various open-domain image manipulation tasks with natural language instructions. |
Sijia Li; Chen Chen; Haonan Lu; | arxiv-cs.CV | 2023-09-08 |
34 | Stroke-based Neural Painting and Stylization with Dynamically Predicted Painting Region Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To solve the problem, we propose Compositional Neural Painter, a novel stroke-based rendering framework which dynamically predicts the next painting region based on the current canvas, instead of dividing the image plane uniformly into painting regions. |
TENG HU et. al. | arxiv-cs.CV | 2023-09-07 |
35 | StyleAdapter: A Single-Pass LoRA-Free Model for Stylized Image Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a LoRA-free method for stylized image generation that takes a text prompt and style reference images as inputs and produces an output image in a single pass. |
ZHOUXIA WANG et. al. | arxiv-cs.CV | 2023-09-04 |
36 | Impact of Image Context for Single Deep Learning Face Morphing Attack Detection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This study investigates the impact of the alignment settings of input images on deep learning face morphing detection performance. |
Joana Pimenta; Iurii Medvedev; Nuno Gonçalves; | arxiv-cs.CV | 2023-09-01 |
37 | Shape-Consistent One-Shot Unsupervised Domain Adaptation for Rail Surface Defect Segmentation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Deep neural networks have greatly improved the performance of rail surface defect segmentation when the test samples have the same distribution as the training samples. However, … |
SHUAI MA et. al. | IEEE Transactions on Industrial Informatics | 2023-09-01 |
38 | Semantic Image Synthesis Via Class-Adaptive Cross-Attention Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In response, we designed a novel architecture where cross-attention layers are used in place of de-normalization ones for conditioning the image generation. |
Tomaso Fontanini; Claudio Ferrari; Giuseppe Lisanti; Massimo Bertozzi; Andrea Prati; | arxiv-cs.CV | 2023-08-30 |
39 | ParaGuide: Guided Diffusion Paraphrasers for Plug-and-Play Textual Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In contrast, we introduce a novel diffusion-based framework for general-purpose style transfer that can be flexibly adapted to arbitrary target styles at inference time. |
Zachary Horvitz; Ajay Patel; Chris Callison-Burch; Zhou Yu; Kathleen McKeown; | arxiv-cs.CL | 2023-08-29 |
40 | WSAM: Visual Explanations from Style Augmentation As Adversarial Attacker and Their Influence in Image Classification Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: With our augmentation strategy, all models not only present incredible robustness against image stylizing but also outperform all previous methods and surpass the state-of-the-art performance for the STL-10 dataset. |
Felipe Moreno-Vera; Edgar Medina; Jorge Poco; | arxiv-cs.CV | 2023-08-28 |
41 | MagicEdit: High-Fidelity and Temporally Coherent Video Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this report, we present MagicEdit, a surprisingly simple yet effective solution to the text-guided video editing task. |
Jun Hao Liew; Hanshu Yan; Jianfeng Zhang; Zhongcong Xu; Jiashi Feng; | arxiv-cs.CV | 2023-08-28 |
42 | Pixel-Aware Stable Diffusion for Realistic Image Super-resolution and Personalized Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a pixel-aware stable diffusion (PASD) network to achieve robust Real-ISR as well as personalized stylization. |
Tao Yang; Peiran Ren; Xuansong Xie; Lei Zhang; | arxiv-cs.CV | 2023-08-28 |
43 | ARF-Plus: Controlling Perceptual Factors in Artistic Radiance Fields for 3D Scene Stylization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present ARF-Plus, a 3D neural style transfer framework offering manageable control over perceptual factors, to systematically explore the perceptual controllability in 3D scene stylization. |
Wenzhao Li; Tianhao Wu; Fangcheng Zhong; Cengiz Oztireli; | arxiv-cs.CV | 2023-08-23 |
44 | Controlling Neural Style Transfer with Deep Reinforcement Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose the first deep Reinforcement Learning (RL) based architecture that splits one-step style transfer into a step-wise process for the NST task. |
CHENGMING FENG et. al. | ijcai | 2023-08-23 |
45 | TeSTNeRF: Text-Driven 3D Style Transfer Via Cross-Modal Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Simply combining image/video style transfer methods and novel view synthesis methods results in flickering when changing viewpoints, while existing 3D style transfer methods learn styles from images instead of texts. To address this problem, we for the first time design an efficient text-driven model for 3D style transfer, named TeSTNeRF, to stylize the scene using texts via cross-modal learning: we leverage an advanced text encoder to embed the texts in order to control 3D style transfer and align the input text and output stylized images in latent space. |
JIAFU CHEN et. al. | ijcai | 2023-08-23 |
46 | A Large-Scale Film Style Dataset for Learning Multi-frequency Driven Film Enhancement Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Inspired by the features of FilmSet images, we propose a novel framework called FilmNet based on Laplacian Pyramid for stylizing images across frequency bands and achieving film style outcomes. |
Zinuo Li; Xuhang Chen; Shuqiang Wang; Chi-Man Pun; | ijcai | 2023-08-23 |
47 | Improving The Transferability of Adversarial Examples with Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Hence, we propose a novel attack method named Style Transfer Method (STM) that utilizes a proposed arbitrary style transfer network to transform the images into different domains. |
ZHIJIN GE et. al. | arxiv-cs.CV | 2023-08-21 |
48 | MRI Field-transfer Reconstruction with Limited Data: Regularization By Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a regularization by neural style transfer (RNST) method to further leverage the priors from the neural transfer and denoising engine. |
GUOYAO SHEN et. al. | arxiv-cs.CV | 2023-08-21 |
49 | Color Prompting for Data-Free Continual Unsupervised Domain Adaptive Person Re-Identification Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a Color Prompting (CoP) method for data-free continual unsupervised domain adaptive person Re-ID. |
JIANYANG GU et. al. | arxiv-cs.CV | 2023-08-21 |
50 | A White-Box False Positive Adversarial Attack Method on Contrastive Loss-Based Offline Handwritten Signature Verification Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we tackle the challenge of white-box false positive adversarial attacks on contrastive loss-based offline handwritten signature verification models. |
Zhongliang Guo; Yifei Qian; Ognjen Arandjelović; Lei Fang; | arxiv-cs.CV | 2023-08-17 |
51 | Diff-CAPTCHA: An Image-based CAPTCHA with Security Enhanced By Denoising Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, an image-click CAPTCHA scheme called Diff-CAPTCHA is proposed based on denoising diffusion models. |
Ran Jiang; Sanfeng Zhang; Linfeng Liu; Yanbing Peng; | arxiv-cs.CR | 2023-08-16 |
52 | CoDeF: Content Deformation Fields for Temporally Consistent Video Processing Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present the content deformation field CoDeF as a new type of video representation, which consists of a canonical content field aggregating the static contents in the entire video and a temporal deformation field recording the transformations from the canonical image (i.e., rendered from the canonical content field) to each individual frame along the time axis.Given a target video, these two fields are jointly optimized to reconstruct it through a carefully tailored rendering pipeline.We advisedly introduce some regularizations into the optimization process, urging the canonical content field to inherit semantics (e.g., the object shape) from the video.With such a design, CoDeF naturally supports lifting image algorithms for video processing, in the sense that one can apply an image algorithm to the canonical image and effortlessly propagate the outcomes to the entire video with the aid of the temporal deformation field.We experimentally show that CoDeF is able to lift image-to-image translation to video-to-video translation and lift keypoint detection to keypoint tracking without any training.More importantly, thanks to our lifting strategy that deploys the algorithms on only one image, we achieve superior cross-frame consistency in processed videos compared to existing video-to-video translation approaches, and even manage to track non-rigid objects like water and smog.Project page can be found at https://qiuyu96.github.io/CoDeF/. |
HAO OUYANG et. al. | arxiv-cs.CV | 2023-08-15 |
53 | Hierarchy Flow For High-Fidelity Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose Hierarchy Flow, a novel flow-based model to achieve better content preservation during translation. |
Weichen Fan; Jinghuan Chen; Ziwei Liu; | arxiv-cs.CV | 2023-08-13 |
54 | Zero-shot Text-driven Physically Interpretable Face Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes a novel and physically interpretable method for face editing based on arbitrary text prompts. |
YAPENG MENG et. al. | arxiv-cs.CV | 2023-08-11 |
55 | BATINet: Background-Aware Text to Image Synthesis and Manipulation Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we analyzed a novel Background-Aware Text2Image (BAT2I) task in which the generated content matches the input background. |
Ryugo Morita; Zhiqiang Zhang; Jinjia Zhou; | arxiv-cs.CV | 2023-08-10 |
56 | A Forensic Methodology for Detecting Image Manipulations Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this study, image file and mobile forensic artifacts analysis were conducted for detecting image manipulation. |
Jiwon Lee; Seungjae Jeon; Yunji Park; Jaehyun Chung; Doowon Jeong; | arxiv-cs.MM | 2023-08-09 |
57 | VAST: Vivify Your Talking Avatar Via Zero-Shot Expressive Facial Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes an unsupervised variational style transfer model (VAST) to vivify the neutral photo-realistic avatars. |
LIYANG CHEN et. al. | arxiv-cs.CV | 2023-08-09 |
58 | A Comparative Study of Image-to-Image Translation Using GANs for Synthetic Child Race Data Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work proposes the utilization of image-to-image transformation to synthesize data of different races and thus adjust the ethnicity of children’s face data. |
Wang Yao; Muhammad Ali Farooq; Joseph Lemley; Peter Corcoran; | arxiv-cs.CV | 2023-08-08 |
59 | DiffSynth: Latent In-Iteration Deflickering for Realistic Video Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose DiffSynth, a novel approach that aims to convert image synthesis pipelines to video synthesis pipelines. |
ZHONGJIE DUAN et. al. | arxiv-cs.CV | 2023-08-07 |
60 | Photorealistic and Identity-Preserving Image-Based Emotion Manipulation with Latent Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we investigate the emotion manipulation capabilities of diffusion models with in-the-wild images, a rather unexplored application area relative to the vast and rapidly growing literature for image-to-image translation tasks. |
Ioannis Pikoulis; Panagiotis P. Filntisis; Petros Maragos; | arxiv-cs.CV | 2023-08-06 |
61 | FAST: Font-Agnostic Scene Text Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, most of the existing STE methods show inferior editing performance because of (1) complex image backgrounds, (2) various font styles, and (3) varying word lengths within the text. To address such inferior editing performance issues, in this paper, we propose a novel font-agnostic scene text editing framework, named FAST, for simultaneously generating text in arbitrary styles and locations while preserving a natural and realistic appearance through combined mask generation and style transfer. |
ALLOY DAS et. al. | arxiv-cs.CV | 2023-08-05 |
62 | MSSRNet: Manipulating Sequential Style Representation for Unsupervised Text Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In fact, each token of a text contains different style intensity and makes different contribution to the overall style. Our proposed method addresses this issue by assigning individual style vector to each token in a text, allowing for fine-grained control and manipulation of the style strength. |
Yazheng Yang; Zhou Zhao; Qi Liu; | kdd | 2023-08-04 |
63 | ADS-Cap: A Framework for Accurate and Diverse Stylized Captioning with Unpaired Stylistic Corpora Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a novel framework to generate Accurate and Diverse Stylized Captions (ADS-Cap). |
KANZHI CHENG et. al. | arxiv-cs.CV | 2023-08-02 |
64 | UCDFormer: Unsupervised Change Detection Using A Transformer-driven Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To this end, we propose a change detection with domain shift setting for remote sensing images. |
Qingsong Xu; Yilei Shi; Jianhua Guo; Chaojun Ouyang; Xiao Xiang Zhu; | arxiv-cs.CV | 2023-08-02 |
65 | ImageBrush: Learning Visual In-Context Instructions for Exemplar-Based Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel manipulation methodology, dubbed ImageBrush, that learns visual instructions for more accurate image editing. |
YASHENG SUN et. al. | arxiv-cs.CV | 2023-08-01 |
66 | Controlling Geometric Abstraction and Texture for Artistic Images Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present a novel method for the interactive control of geometric abstraction and texture in artistic images. |
MARTIN BÜSSEMEYER et. al. | arxiv-cs.CV | 2023-07-31 |
67 | InfoStyler: Disentanglement Information Bottleneck for Artistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Although effective, ignoring the clear disentanglement of the content features and the style features from the first beginning, they have difficulty in balancing between content preservation and style transferring. To tackle this problem, we propose a novel information disentanglement method, named InfoStyler, to capture the minimal sufficient information for both content and style representations from the pre-trained encoding network. |
Yueming Lyu; Yue Jiang; Bo Peng; Jing Dong; | arxiv-cs.CV | 2023-07-30 |
68 | StylePrompter: All Styles Need Is Attention Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: GAN inversion aims at inverting given images into corresponding latent codes for Generative Adversarial Networks (GANs), especially StyleGAN where exists a disentangled latent … |
Chenyi Zhuang; Pan Gao; Aljosa Smolic; | arxiv-cs.CV | 2023-07-30 |
69 | CLIP-PAE: Projection-Augmentation Embedding to Extract Relevant Features for A Disentangled, Interpretable and Controllable Text-Guided Face Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Disentanglement, interpretability, and controllability are also hard to guarantee for manipulation. To alleviate these problems, we propose to define corpus subspaces spanned by relevant prompts to capture specific image characteristics. |
Chenliang Zhou; Fangcheng Zhong; Cengiz Öztireli; | siggraph | 2023-07-26 |
70 | Drag Your GAN: Interactive Point-based Manipulation on The Generative Image Manifold IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we study a powerful yet much less explored way of controlling GANs, that is, to "drag" any points of the image to precisely reach target points in a user-interactive manner, as shown in Fig.1. |
XINGANG PAN et. al. | siggraph | 2023-07-26 |
71 | Zero-shot Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we introduce pix2pix-zero, an image-to-image translation method that can preserve the original image’s content without manual prompting. |
GAURAV PARMAR et. al. | siggraph | 2023-07-26 |
72 | Face-PAST: Facial Pose Awareness and Style Transfer Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a facial pose awareness and style transfer (Face-PAST) network that preserves facial details and structures while generating high-quality stylized images. |
Sunder Ali Khowaja; Ghulam Mujtaba; Jiseok Yoon; Ik Hyun Lee; | arxiv-cs.CV | 2023-07-18 |
73 | On The Fly Neural Style Smoothing for Risk-Averse Domain Generalization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To enable risk-averse predictions from a DG classifier, we propose a novel inference procedure, Test-Time Neural Style Smoothing (TT-NSS), that uses a style-smoothed version of the DG classifier for prediction at test time. |
Akshay Mehra; Yunbei Zhang; Bhavya Kailkhura; Jihun Hamm; | arxiv-cs.CV | 2023-07-17 |
74 | Dense Multitask Learning to Reconfigure Comics Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we develop a MultiTask Learning (MTL) model to achieve dense predictions for comics panels to, in turn, facilitate the transfer of comics from one publication channel to another by assisting authors in the task of reconfiguring their narratives. |
Deblina Bhattacharjee; Sabine Süsstrunk; Mathieu Salzmann; | arxiv-cs.CV | 2023-07-16 |
75 | Bidirectionally Deformable Motion Modulation For Video-based Human Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Considering the difficulties in transferring highly structural patterns on the garments and discontinuous poses, existing methods often generate unsatisfactory results such as distorted textures and flickering artifacts. To address these issues, we propose a novel Deformable Motion Modulation (DMM) that utilizes geometric kernel offset with adaptive weight modulation to simultaneously perform feature alignment and style transfer. |
WING-YIN YU et. al. | arxiv-cs.CV | 2023-07-15 |
76 | Sem-CS: Semantic CLIPStyler for Text-Based Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, the ground semantics of objects in the style transfer output is lost due to style spill-over on salient and background objects (content mismatch) or over-stylization. To solve this, we propose Semantic CLIPStyler (Sem-CS), that performs semantic style transfer. |
Chanda Grover Kamra; Indra Deep Mastan; Debayan Gupta; | arxiv-cs.CV | 2023-07-12 |
77 | DIFF-NST: Diffusion Interleaving For DeFormable Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: With the recent introduction of diffusion models, such as Stable Diffusion, we can access far more powerful image generation techniques, enabling new possibilities. In our work, we propose using this new class of models to perform style transfer while enabling deformable style transfer, an elusive capability in previous models. |
DAN RUTA et. al. | arxiv-cs.CV | 2023-07-09 |
78 | Text Style Transfer Back-Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: For natural inputs, BT brings only slight improvements and sometimes even adverse effects. To address this issue, we propose Text Style Transfer Back Translation (TST BT), which uses a style transfer to modify the source side of BT data. |
DAIMENG WEI et. al. | acl | 2023-07-08 |
79 | StoryTrans: Non-Parallel Story Author-Style Transfer with Discourse Representations and Content Enhancing Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we formulate the task of non-parallel story author-style transfer, which requires transferring an input story into a specified author style while maintaining source semantics. |
Xuekai Zhu; Jian Guan; Minlie Huang; Juan Liu; | acl | 2023-07-08 |
80 | PEIT: Bridging The Modality Gap with Pre-trained Models for End-to-End Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose PEIT, an end-to-end image translation framework that bridges the modality gap with pre-trained models. |
Shaolin Zhu; Shangjie Li; Yikun Lei; Deyi Xiong; | acl | 2023-07-08 |
81 | StyleStegan: Leak-free Style Transfer Based on Feature Steganography Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In modern social networks, existing style transfer methods suffer from a serious content leakage issue, which hampers the ability to achieve serial and reversible stylization, thereby hindering the further propagation of stylized images in social networks. To address this problem, we propose a leak-free style transfer method based on feature steganography. |
Xiujian Liang; Bingshan Liu; Qichao Ying; Zhenxing Qian; Xinpeng Zhang; | arxiv-cs.CV | 2023-07-01 |
82 | ReDi: Efficient Learning-Free Diffusion Inference Via Trajectory Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To accelerate the inference, we propose ReDi, a simple yet learning-free Retrieval-based Diffusion sampling framework. |
Kexun Zhang; Xianjun Yang; William Yang Wang; Lei Li; | icml | 2023-06-27 |
83 | SinDDM: A Single Image Denoising Diffusion Model IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Here, we introduce a framework for training a DDM on a single image. |
Vladimir Kulikov; Shahar Yadin; Matan Kleiner; Tomer Michaeli; | icml | 2023-06-27 |
84 | Progressive Energy-Based Cooperative Learning for Multi-Domain Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Since the style generator is represented as an domain-specific distribution of style codes, the translator can provide a one-to-many transformation (i.e., diversified generation) between source domain and target domain. To train our framework, we propose a likelihood-based multi-domain cooperative learning algorithm to jointly train the multi-domain descriptor and the diversified image generator (including translator, style encoder, and style generator modules) via multi-domain MCMC teaching, in which the descriptor guides the diversified image generator to shift its probability density toward the data distribution, while the diversified image generator uses its randomly translated images to initialize the descriptor’s Langevin dynamics process for efficient sampling. |
Weinan Song; Yaxuan Zhu; Lei He; Yingnian Wu; Jianwen Xie; | arxiv-cs.CV | 2023-06-26 |
85 | SHUNIT: Style Harmonization for Unpaired Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel solution for unpaired image-to-image (I2I) translation. |
Seokbeom Song; Suhyeon Lee; Hongje Seong; Kyoungwon Min; Euntai Kim; | aaai | 2023-06-26 |
86 | Preserving Structural Consistency in Arbitrary Artist and Artwork Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: These methods not only homogenize the artist-style of different artworks of the same artist but also lack generalization for the unseen artists. To solve these challenges, we propose a double-style transferring module (DSTM). |
JINGYU WU et. al. | aaai | 2023-06-26 |
87 | MicroAST: Towards Super-fast Ultra-Resolution Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Despite the recent rapid progress, existing AST methods are either incapable or too slow to run at ultra-resolutions (e.g., 4K) with limited resources, which heavily hinders their further applications. In this paper, we tackle this dilemma by learning a straightforward and lightweight model, dubbed MicroAST. |
ZHIZHONG WANG et. al. | aaai | 2023-06-26 |
88 | Practical Disruption of Image Translation Deepfake Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work we propose Leaking Transferable Perturbations (LTP), an algorithm that significantly reduces the number of queries needed to disrupt an image translation network by dynamically re-purposing previous disruptions into new query efficient disruptions. |
Nataniel Ruiz; Sarah Adel Bargal; Cihang Xie; Stan Sclaroff; | aaai | 2023-06-26 |
89 | AdaCM: Adaptive ColorMLP for Real-Time Universal Photo-Realistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose the Adaptive ColorMLP (AdaCM), an effective and efficient framework for universal photo-realistic style transfer. |
TIANWEI LIN et. al. | aaai | 2023-06-26 |
90 | CLIPVG: Text-Guided Image Manipulation Using Differentiable Vector Graphics Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce CLIPVG, a text-guided image manipulation framework using differentiable vector graphics, which is also the first CLIP-based general image manipulation framework that does not require any additional generative models. |
YIREN SONG et. al. | aaai | 2023-06-26 |
91 | Frequency Domain Disentanglement for Arbitrary Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Therefore, these methods always suffer from low-quality results because of the sub-optimal disentanglement. To address such a challenge, this paper proposes the frequency mixer (FreMixer) module that disentangles and re-entangles the frequency spectrum of content and style components in the frequency domain. |
DONGYANG LI et. al. | aaai | 2023-06-26 |
92 | Target-Free Text-Guided Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We tackle the problem of target-free text-guided image manipulation, which requires one to modify the input reference image based on the given text instruction, while no ground truth target image is observed during training. To address this challenging task, we propose a Cyclic-Manipulation GAN (cManiGAN) in this paper, which is able to realize where and how to edit the image regions of interest. |
Wan-Cyuan Fan; Cheng-Fu Yang; Chiao-An Yang; Yu-Chiang Frank Wang; | aaai | 2023-06-26 |
93 | CFFT-GAN: Cross-Domain Feature Fusion Transformer for Exemplar-Based Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a more general learning approach by considering two domain features as a whole and learning both inter-domain correspondence and intra-domain potential information interactions. |
TIANXIANG MA et. al. | aaai | 2023-06-26 |
94 | User-Controllable Arbitrary Style Transfer Via Entropy Regularization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel solution ensuring both efficiency and diversity for generating multiple user-controllable AST results by systematically modulating AST behavior at run-time. |
JIAXIN CHENG et. al. | aaai | 2023-06-26 |
95 | PP-GAN : Style Transfer from Korean Portraits to ID Photos Using Landmark Extractor with GAN Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Owing to its distinct characteristics from the hair in ID photos, transferring the Gat is challenging. To address this issue, this study proposes a deep learning network that can perform style transfer, including the Gat, while preserving the identity of the face. |
Jongwook Si; Sungyoung Kim; | arxiv-cs.CV | 2023-06-23 |
96 | What to Learn: Features, Image Transformations, or Both? Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose to combine an image transformation network and a feature-learning network to improve long-term localization performance. |
Yuxuan Chen; Binbin Xu; Frederike Dümbgen; Timothy D. Barfoot; | arxiv-cs.RO | 2023-06-22 |
97 | Motion Capture Dataset for Practical Use of AI-based Motion Editing and Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we proposed a new style-diverse dataset for the domain of motion style transfer. |
Makito Kobayashi; Chen-Chieh Liao; Keito Inoue; Sentaro Yojima; Masafumi Takahashi; | arxiv-cs.CV | 2023-06-15 |
98 | ArtFusion: Arbitrary Style Transfer Using Dual Conditional Latent Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a new approach, ArtFusion, which provides a flexible balance between content and style. |
Dar-Yen Chen; | arxiv-cs.CV | 2023-06-15 |
99 | GBSD: Generative Bokeh with Stage Diffusion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present GBSD, the first generative text-to-image model that synthesizes photorealistic images with a bokeh style. |
Jieren Deng; Xin Zhou; Hao Tian; Zhihong Pan; Derek Aguiar; | arxiv-cs.CV | 2023-06-14 |
100 | GP-UNIT: Generative Prior for Versatile Unsupervised Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce a novel versatile framework, Generative Prior-guided UNsupervised Image-to-image Translation (GP-UNIT), that improves the quality, applicability and controllability of the existing translation models. |
Shuai Yang; Liming Jiang; Ziwei Liu; Chen Change Loy; | arxiv-cs.CV | 2023-06-07 |
101 | Improving Diffusion-based Image Translation Using Asymmetric Gradient Guidance Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Yet, these methods often require computationally intense fine-tuning of diffusion models or additional neural networks. To address these challenges, here we present an approach that guides the reverse process of diffusion sampling by applying asymmetric gradient guidance. |
Gihyun Kwon; Jong Chul Ye; | arxiv-cs.CV | 2023-06-07 |
102 | Interpretable Style Transfer for Text-to-Speech with ControlVAE and Diffusion Bridge Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a new TTS system that can perform style transfer with interpretability and high fidelity. |
WENHAO GUAN et. al. | arxiv-cs.SD | 2023-06-07 |
103 | Instruct-Video2Avatar: Video-to-Avatar Generation with Instructions Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a method for synthesizing edited photo-realistic digital avatars with text instructions. |
Shaoxu Li; | arxiv-cs.CV | 2023-06-05 |
104 | Identifying The Style By A Qualified Reader on A Short Fragment of Generated Poetry Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: I used 3 character-based LSTM-models to work with style reproducing assessment. |
Boris Orekhov; | arxiv-cs.CL | 2023-06-05 |
105 | Intelligent Grimm — Open-ended Visual Storytelling Via Latent Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we focus on the task of generating a series of coherent image sequence based on a given storyline, denoted as open-ended visual storytelling. |
Chang Liu; Haoning Wu; Yujie Zhong; Xiaoyun Zhang; Weidi Xie; | arxiv-cs.CV | 2023-06-01 |
106 | Image Reference-guided Fashion Design with Structure-aware Transfer By Diffusion Models Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Image-based fashion design with AI techniques has attracted increasing attention in recent years. We focus on a new fashion design task, where we aim to transfer a reference … |
Shidong Cao; Wenhao Chai; Shengyu Hao; Gaoang Wang; | 2023 IEEE/CVF Conference on Computer Vision and Pattern … | 2023-06-01 |
107 | Zero-shot Pose Transfer for Unrigged Stylized 3D Characters Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a zero-shot approach that requires only the widely available deformed non-stylized avatars in training, and deforms stylized characters of significantly different shapes at inference. |
JIASHUN WANG et. al. | arxiv-cs.CV | 2023-05-31 |
108 | Context-Preserving Two-Stage Video Domain Translation for Portrait Stylization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To address the issue, we propose a novel two-stage video translation framework with an objective function which enforces a model to generate a temporally coherent stylized video while preserving context in the source video. |
DOYEON KIM et. al. | arxiv-cs.CV | 2023-05-30 |
109 | SAVE: Spectral-Shift-Aware Adaptation of Image Diffusion Models for Text-guided Video Editing Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Although the latter is computationally less expensive, it still takes a significant amount of time for per-video adaption. To address this issue, we propose SAVE, a novel spectral-shift-aware adaptation framework, in which we fine-tune the spectral shift of the parameter space instead of the parameters themselves. |
Nazmul Karim; Umar Khalid; Mohsen Joneidi; Chen Chen; Nazanin Rahnavard; | arxiv-cs.CV | 2023-05-29 |
110 | Conditional Score Guidance for Text-Driven Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a novel algorithm for text-driven image-to-image translation based on a pretrained text-to-image diffusion model. |
Hyunsoo Lee; Minsoo Kang; Bohyung Han; | arxiv-cs.CV | 2023-05-29 |
111 | StyleS2ST: Zero-shot Style Transfer for Direct Speech-to-speech Translation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Direct speech-to-speech translation (S2ST) has gradually become popular as it has many advantages compared with cascade S2ST. However, current research mainly focuses on the … |
KUN SONG et. al. | arxiv-cs.SD | 2023-05-28 |
112 | CLIP3Dstyler: Language Guided 3D Arbitrary Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel language-guided 3D arbitrary neural style transfer method (CLIP3Dstyler). |
MING GAO et. al. | arxiv-cs.CV | 2023-05-25 |
113 | Balancing Effect of Training Dataset Distribution of Multiple Styles for Multi-Style Text Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Through quantitative analysis, we explore the impact of multiple style distributions in training data on style-transferred output. |
Debarati Das; David Ma; Dongyeop Kang; | arxiv-cs.CL | 2023-05-24 |
114 | Unpaired Image-to-Image Translation Via Neural Schrödinger Bridge Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose the Unpaired Neural Schr\odinger Bridge (UNSB), which combines SB with adversarial training and regularization to learn a SB between unpaired data. |
Beomsu Kim; Gihyun Kwon; Kwanyoung Kim; Jong Chul Ye; | arxiv-cs.CV | 2023-05-24 |
115 | SAMScore: A Semantic Structural Similarity Metric for Image Translation Evaluation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Traditional image-level similarity metrics are of limited use, since the semantics of an image are high-level, and not strongly governed by pixel-wise faithfulness to an original image. Towards filling this gap, we introduce SAMScore, a generic semantic structural similarity metric for evaluating the faithfulness of image translation models. |
YUNXIANG LI et. al. | arxiv-cs.CV | 2023-05-24 |
116 | Dual-Side Feature Fusion 3D Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a Dual-Side Feature Fusion Pose Transfer Network to improve the pose transfer accuracy of the lightweight method. |
Jue Liu; Feipeng Da; | arxiv-cs.CV | 2023-05-24 |
117 | SAR-to-Optical Image Translation Via Thermodynamics-inspired Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Motivated by the analogy between pixels during the S2O image translation and molecules in a heat field, Thermodynamics-inspired Network for SAR-to-Optical Image Translation (S2O-TDN) is proposed in this paper. |
MINGJIN ZHANG et. al. | arxiv-cs.CV | 2023-05-23 |
118 | Variational Bayesian Framework for Advanced Image Generation with Domain-Related Variables Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, it remains challenging for existing methods to address advanced conditional generative problems without annotations, which can enable multiple applications like image-to-image translation and image editing. We present a unified Bayesian framework for such problems, which introduces an inference stage on latent variables within the learning process. |
Yuxiao Li; Santiago Mazuelas; Yuan Shen; | arxiv-cs.CV | 2023-05-23 |
119 | Control-A-Video: Controllable Text-to-Video Generation with Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a controllable text-to-video (T2V) diffusion model, named Video-ControlNet, that generates videos conditioned on a sequence of control signals, such as edge or depth maps. |
WEIFENG CHEN et. al. | arxiv-cs.CV | 2023-05-23 |
120 | InstructVid2Vid: Controllable Video Editing with Natural Language Instructions Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present an end-to-end diffusion-based method for editing videos with human language instructions, namely $\textbf{InstructVid2Vid}$. |
Bosheng Qin; Juncheng Li; Siliang Tang; Tat-Seng Chua; Yueting Zhuang; | arxiv-cs.CV | 2023-05-20 |
121 | Brain Captioning: Decoding Human Brain Activity Into Images and Text Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Recent breakthroughs in functional magnetic resonance imaging (fMRI) have enabled scientists to extract visual information from human brain activity patterns. In this study, we present an innovative method for decoding brain activity into meaningful images and captions, with a specific focus on brain captioning due to its enhanced flexibility as compared to brain decoding into images. |
Matteo Ferrante; Furkan Ozcelik; Tommaso Boccato; Rufin VanRullen; Nicola Toschi; | arxiv-cs.CV | 2023-05-19 |
122 | Drag Your GAN: Interactive Point-based Manipulation on The Generative Image Manifold IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we study a powerful yet much less explored way of controlling GANs, that is, to drag any points of the image to precisely reach target points in a user-interactive manner, as shown in Fig.1. |
XINGANG PAN et. al. | arxiv-cs.CV | 2023-05-18 |
123 | Domain Adaptive Sim-to-Real Segmentation of Oropharyngeal Organs Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a domain adaptive Sim-to-Real framework called IoU-Ranking Blend-ArtFlow (IRB-AF) for image segmentation of oropharyngeal organs. |
Guankun Wang; Tian-Ao Ren; Jiewen Lai; Long Bai; Hongliang Ren; | arxiv-cs.AI | 2023-05-18 |
124 | Masked and Adaptive Transformer for Exemplar Based Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present a novel framework for exemplar based image translation. |
CHANG JIANG et. al. | cvpr | 2023-05-17 |
125 | Inversion-Based Style Transfer With Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Specifically, we perceive style as a learnable textual description of a painting.We propose an inversion-based style transfer method (InST), which can efficiently and accurately learn the key information of an image, thus capturing and transferring the artistic style of a painting. |
YUXIN ZHANG et. al. | cvpr | 2023-05-17 |
126 | Plug-and-Play Diffusion Features for Text-Driven Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a new framework that takes text-to-image synthesis to the realm of image-to-image translation — given a guidance image and a target text prompt as input, our method harnesses the power of a pre-trained text-to-image diffusion model to generate a new image that complies with the target text, while preserving the semantic layout of the guidance image. |
Narek Tumanyan; Michal Geyer; Shai Bagon; Tali Dekel; | cvpr | 2023-05-17 |
127 | Tunable Convolutions With Parametric Multi-Loss Optimization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose to optimize a parametric tunable convolutional layer, which includes a number of different kernels, using a parametric multi-loss, which includes an equal number of objectives. |
Matteo Maggioni; Thomas Tanay; Francesca Babiloni; Steven McDonagh; Aleš Leonardis; | cvpr | 2023-05-17 |
128 | Unpaired Image-to-Image Translation With Shortest Path Regularization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we start from a different perspective and consider the paths connecting the two domains. |
Shaoan Xie; Yanwu Xu; Mingming Gong; Kun Zhang; | cvpr | 2023-05-17 |
129 | StyleRF: Zero-Shot 3D Style Transfer of Neural Radiance Fields Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose StyleRF (Style Radiance Fields), an innovative 3D style transfer technique that resolves the three-way dilemma by performing style transformation within the feature space of a radiance field. |
KUNHAO LIU et. al. | cvpr | 2023-05-17 |
130 | Neural Preset for Color Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a Neural Preset technique to address the limitations of existing color style transfer methods, including visual artifacts, vast memory requirement, and slow style switching speed. |
Zhanghan Ke; Yuhao Liu; Lei Zhu; Nanxuan Zhao; Rynson W.H. Lau; | cvpr | 2023-05-17 |
131 | Transforming Radiance Field With Lipschitz Network for Photorealistic 3D Scene Stylization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Simply coupling NeRF with photorealistic style transfer (PST) will result in cross-view inconsistency and degradation of stylized view syntheses. Through a thorough analysis, we demonstrate that this non-trivial task can be simplified in a new light: When transforming the appearance representation of a pre-trained NeRF with Lipschitz mapping, the consistency and photorealism across source views will be seamlessly encoded into the syntheses. |
ZICHENG ZHANG et. al. | cvpr | 2023-05-17 |
132 | Learning Dynamic Style Kernels for Artistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To further enhance the flexibility of our style transfer method, we propose a Style Alignment Encoding (SAE) module complemented with a Content-based Gating Modulation (CGM) module for learning the dynamic style kernels in focusing regions. |
Wenju Xu; Chengjiang Long; Yongwei Nie; | cvpr | 2023-05-17 |
133 | Master: Meta Style Transformer for Controllable Zero-Shot and Few-Shot Artistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we devise a novel Transformer model termed as Master specifically for style transfer. |
HAO TANG et. al. | cvpr | 2023-05-17 |
134 | EDICT: Exact Diffusion Inversion Via Coupled Transformations IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, DDIM inversion for real images is unstable as it relies on local linearization assumptions, which result in the propagation of errors, leading to incorrect image reconstruction and loss of content. To alleviate these problems, we propose Exact Diffusion Inversion via Coupled Transformations (EDICT), an inversion method that draws inspiration from affine coupling layers. |
Bram Wallace; Akash Gokul; Nikhil Naik; | cvpr | 2023-05-17 |
135 | CAP-VSTNet: Content Affinity Preserved Versatile Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes a new framework named CAP-VSTNet, which consists of a new reversible residual network and an unbiased linear transform module, for versatile style transfer. |
Linfeng Wen; Chengying Gao; Changqing Zou; | cvpr | 2023-05-17 |
136 | Ref-NPR: Reference-Based Non-Photorealistic Radiance Fields for Controllable Scene Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Current 3D scene stylization methods transfer textures and colors as styles using arbitrary style references, lacking meaningful semantic correspondences. We introduce Reference-Based Non-Photorealistic Radiance Fields (Ref-NPR) to address this limitation. |
Yuechen Zhang; Zexin He; Jinbo Xing; Xufeng Yao; Jiaya Jia; | cvpr | 2023-05-17 |
137 | BBDM: Image-to-Image Translation With Brownian Bridge Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, a novel image-to-image translation method based on the Brownian Bridge Diffusion Model(BBDM) is proposed, which models image-to-image translation as a stochastic Brownian Bridge process, and learns the translation between two domains directly through the bidirectional diffusion process rather than a conditional generation process. |
Bo Li; Kaitao Xue; Bin Liu; Yu-Kun Lai; | cvpr | 2023-05-17 |
138 | Imagic: Text-Based Real Image Editing With Diffusion Models IF:5 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we demonstrate, for the very first time, the ability to apply complex (e.g., non-rigid) text-based semantic edits to a single real image. |
BAHJAT KAWAR et. al. | cvpr | 2023-05-17 |
139 | Modernizing Old Photos Using Multiple References Via Photorealistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In order to modernize old photos, we propose a novel multi-reference-based old photo modernization (MROPM) framework consisting of a network MROPM-Net and a novel synthetic data generation scheme. |
Agus Gunawan; Soo Ye Kim; Hyeonjun Sim; Jae-Ho Lee; Munchurl Kim; | cvpr | 2023-05-17 |
140 | Style Transfer Enabled Sim2Real Framework for Efficient Learning of Robotic Ultrasound Image Analysis Using Simulated Data Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work presents a Sim2Real framework to efficiently learn robotic US image analysis tasks based only on simulated data for real-world deployment. |
KEYU LI et. al. | arxiv-cs.RO | 2023-05-16 |
141 | Wavelet-based Unsupervised Label-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: State-of-the-art conditional Generative Adversarial Networks (GANs) need a huge amount of paired data to accomplish this task while generic unpaired image-to-image translation frameworks underperform in comparison, because they color-code semantic layouts and learn correspondences in appearance instead of semantic content. Starting from the assumption that a high quality generated image should be segmented back to its semantic layout, we propose a new Unsupervised paradigm for SIS (USIS) that makes use of a self-supervised segmentation loss and whole image wavelet based discrimination. |
George Eskandar; Mohamed Abdelsamad; Karim Armanious; Shuai Zhang; Bin Yang; | arxiv-cs.CV | 2023-05-16 |
142 | Realization RGBD Image Stylization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel method that incorporates the depth map and a heatmap of the RGB image to generate more realistic style transfer results. |
Bhavya Sehgal; Vaishnavi Mendu; Aparna Mendu; | arxiv-cs.CV | 2023-05-11 |
143 | Analysis of Adversarial Image Manipulations Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper investigates how simple, accessible image manipulation techniques affect the accuracy of facial recognition software in identifying an individual’s various face images based on one unique image. |
Ahsi Lo; Gabriella Pangelinan; Michael C. King; | arxiv-cs.CV | 2023-05-10 |
144 | Style-A-Video: Agile Diffusion for Arbitrary Text-based Video Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes a zero-shot video stylization method named Style-A-Video, which utilizes a generative pre-trained transformer with an image latent diffusion model to achieve a concise text-controlled video stylization. |
Nisha Huang; Yuxin Zhang; Weiming Dong; | arxiv-cs.CV | 2023-05-09 |
145 | Multi-Teacher Knowledge Distillation For Text Image Machine Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel Multi-Teacher Knowledge Distillation (MTKD) method to effectively distillate knowledge into the end-to-end TIMT model from the pipeline model. |
CONG MA et. al. | arxiv-cs.CL | 2023-05-09 |
146 | Joint Multi-scale Cross-lingual Speaking Style Transfer with Bidirectional Attention Mechanism for Automatic Dubbing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a joint multi-scale cross-lingual speaking style transfer framework to simultaneously model the bidirectional speaking style transfer between languages at both global (i.e. utterance level) and local (i.e. word level) scales. |
JINGBEI LI et. al. | arxiv-cs.SD | 2023-05-09 |
147 | Transplayer: Timbre Style Transfer with Flexible Timbre Control Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Inspired by the practice in voice conversion, we propose TransPlayer, which uses an autoencoder model with one-hot representations of instruments as the condition, and a Diffwave model trained especially for music synthesis. |
Y. Wu; Y. He; X. Liu; Y. Wang; R. B. Dannenberg; | icassp | 2023-04-27 |
148 | CPD-GAN: Cascaded Pyramid Deformation GAN for Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Existing work often failed to transfer complex textures to generated images well. To solve this problem, we propose a novel network for this task. |
Y. Huang; Y. Tang; X. Zheng; J. Tang; | icassp | 2023-04-27 |
149 | MSNet: A Deep Architecture Using Multi-Sentiment Semantics for Sentiment-Aware Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To incorporate the sentiment information into the image style transfer task for better sentiment-aware performance, we introduce a new task named sentiment-aware image style transfer. |
S. Sun; J. Jia; H. Wu; Z. Ye; J. Xing; | icassp | 2023-04-27 |
150 | Multidimensional Evaluation for Text Style Transfer Using ChatGPT Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We perform a comprehensive correlation analysis for two transfer directions (and overall) at different levels. |
Huiyuan Lai; Antonio Toral; Malvina Nissim; | arxiv-cs.CL | 2023-04-26 |
151 | Hierarchical Diffusion Autoencoders and Disentangled Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To mitigate those limitations, we propose Hierarchical Diffusion Autoencoders (HDAE) that exploit the fine-grained-to-abstract and lowlevel-to-high-level feature hierarchy for the latent space of diffusion models. |
ZEYU LU et. al. | arxiv-cs.CV | 2023-04-24 |
152 | Unsupervised Style-based Explicit 3D Face Reconstruction from Single Image Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a general adversarial learning framework for solving Unsupervised 2D to Explicit 3D Style Transfer (UE3DST). |
Heng Yu; Zoltan A. Milacski; Laszlo A. Jeni; | arxiv-cs.CV | 2023-04-24 |
153 | Spectral Normalized Dual Contrastive Regularization for Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a new unpaired I2I translation framework based on dual contrastive regularization and spectral normalization, namely SN-DCR. |
Chen Zhao; Wei-Ling Cai; Zheng Yuan; | arxiv-cs.CV | 2023-04-22 |
154 | A Plug-and-Play Defensive Perturbation for Copyright Protection of DNN-based Applications Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel plug-and-play invisible copyright protection method based on defensive perturbation for DNN-based applications (i.e., style transfer). |
DONGHUA WANG et. al. | arxiv-cs.CV | 2023-04-20 |
155 | Any-to-Any Style Transfer: Making Picasso and Da Vinci Collaborate Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In either case, only one result can be generated for a specific pair of content and style images, which therefore lacks flexibility and is hard to satisfy different users with different preferences. We propose here a novel strategy termed Any-to-Any Style Transfer to address this drawback, which enables users to interactively select styles of regions in the style image and apply them to the prescribed content regions. |
Songhua Liu; Jingwen Ye; Xinchao Wang; | arxiv-cs.CV | 2023-04-19 |
156 | UPGPT: Universal Diffusion Model for Person Image Generation, Editing and Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: The appearance inconsistency makes T2I unsuitable for pose transfer. We address this by proposing a multimodal diffusion model that accepts text, pose, and visual prompting. |
Soon Yau Cheong; Armin Mustafa; Andrew Gilbert; | arxiv-cs.CV | 2023-04-18 |
157 | Delta Denoising Score Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce Delta Denoising Score (DDS), a novel scoring function for text-based image editing that guides minimal modifications of an input image towards the content described in a target prompt. |
Amir Hertz; Kfir Aberman; Daniel Cohen-Or; | arxiv-cs.CV | 2023-04-14 |
158 | ALADIN-NST: Self-supervised Disentangled Representation Learning of Artistic Style Through Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our paper aims to learn a representation of visual artistic style more strongly disentangled from the semantic content depicted in an image. |
Dan Ruta; Gemma Canet Tarres; Alexander Black; Andrew Gilbert; John Collomosse; | arxiv-cs.CV | 2023-04-12 |
159 | Panoramic Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we tackle the challenging task of Panoramic Image-to-Image translation (Pano-I2I) for the first time. |
SOOHYUN KIM et. al. | arxiv-cs.CV | 2023-04-11 |
160 | Improving Diffusion Models for Scene Text Editing with Dual Encoders Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, our empirical analysis reveals that state-of-the-art diffusion models struggle with rendering correct text and controlling text style. To address these problems, we propose DIFFSTE to improve pre-trained diffusion models with a dual encoder design, which includes a character encoder for better text legibility and an instruction encoder for better style control. |
JIABAO JI et. al. | arxiv-cs.CV | 2023-04-11 |
161 | NeAT: Neural Artistic Tracing for Beautiful Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present NeAT, a new state-of-the art feed-forward style transfer method. |
Dan Ruta; Andrew Gilbert; John Collomosse; Eli Shechtman; Nicholas Kolkin; | arxiv-cs.CV | 2023-04-11 |
162 | ITportrait: Image-Text Coupled 3D Portrait Domain Adaptation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose an Image-Text multi-modal framework, namely Image and Text portrait (ITportrait), for 3D portrait domain adaptation. |
XIANGWEN DENG et. al. | arxiv-cs.MM | 2023-04-09 |
163 | One-shot Unsupervised Domain Adaptation with Personalized Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Departing from the common notion of transferring only the target “texture” information, we leverage text-to-image diffusion models (e.g., Stable Diffusion) to generate a synthetic target dataset with photo-realistic images that not only faithfully depict the style of the target domain, but are also characterized by novel scenes in diverse contexts. |
Yasser Benigmim; Subhankar Roy; Slim Essid; Vicky Kalogeiton; Stéphane Lathuilière; | arxiv-cs.CV | 2023-03-31 |
164 | Semantic Image Translation for Repairing The Texture Defects of Building Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In order to preserve fine details and regular structures, we propose a regularity-aware multi-domain method that capitalizes on frequency information and corner maps. |
QISEN SHANG et. al. | arxiv-cs.CV | 2023-03-30 |
165 | Instant Neural Radiance Fields Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present Instant Neural Radiance Fields Stylization, a novel approach for multi-view image stylization for the 3D scene. |
Shaoxu Li; Ye Pan; | arxiv-cs.CV | 2023-03-29 |
166 | Training-free Style Transfer Emerges from H-space in Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we discover that DMs inherently have disentangled representations for content and style of the resulting images: $h$-space contains the content and the skip connections convey the style. |
Jaeseok Jeong; Mingi Kwon; Youngjung Uh; | arxiv-cs.CV | 2023-03-27 |
167 | Neural Preset for Color Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a Neural Preset technique to address the limitations of existing color style transfer methods, including visual artifacts, vast memory requirement, and slow style switching speed. |
Zhanghan Ke; Yuhao Liu; Lei Zhu; Nanxuan Zhao; Rynson W. H. Lau; | arxiv-cs.CV | 2023-03-23 |
168 | Open-World Pose Transfer Via Sequential Test-Time Adaption Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: A typical pose transfer framework usually employs representative datasets to train a discriminative model, which is often violated by out-of-distribution (OOD) instances. |
JUNYANG CHEN et. al. | arxiv-cs.CV | 2023-03-20 |
169 | StyleRF: Zero-shot 3D Style Transfer of Neural Radiance Fields Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose StyleRF (Style Radiance Fields), an innovative 3D style transfer technique that resolves the three-way dilemma by performing style transformation within the feature space of a radiance field. |
KUNHAO LIU et. al. | arxiv-cs.CV | 2023-03-19 |
170 | DialogPaint: A Dialog-based Image Editing Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present DialogPaint, an innovative framework that employs an interactive conversational approach for image editing. |
Jingxuan Wei; Shiyu Wu; Xin Jiang; Yequan Wang; | arxiv-cs.CV | 2023-03-17 |
171 | Style Transfer for 2D Talking Head Animation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a new method to generate talking head animation with learnable style references. |
TRONG-THANG PHAM et. al. | arxiv-cs.CV | 2023-03-17 |
172 | NLUT: Neural-based 3D Lookup Tables for Video Photorealistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, existing methods obtain stylized video sequences by performing frame-by-frame photorealistic style transfer, which is inefficient and does not ensure the temporal consistency of the stylized video. To address this issue, we use neural network-based 3D Lookup Tables (LUTs) for the photorealistic transfer of videos, achieving a balance between efficiency and effectiveness. |
YAOSEN CHEN et. al. | arxiv-cs.CV | 2023-03-16 |
173 | SpectralCLIP: Preventing Artifacts in Text-Guided Style Transfer from A Spectral Perspective Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose SpectralCLIP, which implements a spectral filtering layer on top of the CLIP vision encoder, to alleviate the artifact issue. |
Zipeng Xu; Songlong Xing; Enver Sangineto; Nicu Sebe; | arxiv-cs.CV | 2023-03-16 |
174 | Class-Guided Image-to-Image Diffusion: Cell Painting from Brightfield Images with Class Labels Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce and implement a model which combines image-to-image and class-guided denoising diffusion probabilistic models. |
JAN OSCAR CROSS-ZAMIRSKI et. al. | arxiv-cs.CV | 2023-03-15 |
175 | 3D Face Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, previous methods mainly use images of artistic faces for style transfer while ignoring arbitrary style images such as abstract paintings. To solve this problem, we propose a novel method, namely Face-guided Dual Style Transfer (FDST). |
XIANGWEN DENG et. al. | arxiv-cs.CV | 2023-03-14 |
176 | SEM-CS: Semantic CLIPStyler for Text-Based Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, the ground semantics of objects in style transfer output is lost due to style spillover on salient and background objects (content mismatch) or over-stylization. To solve this, we propose Semantic CLIPStyler (Sem-CS) that performs semantic style transfer. |
Chanda G Kamra; Indra Deep Mastan; Debayan Gupta; | arxiv-cs.CV | 2023-03-11 |
177 | AptSim2Real: Approximately-Paired Sim-to-Real Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Unpaired image translation, while more suitable for sim-to-real transfer, is still challenging to learn for complex natural scenes. To address these challenges, we propose a third category: approximately-paired sim-to-real translation, where the source and target images do not need to be exactly paired. |
Charles Y Zhang; Ashish Shrivastava; | arxiv-cs.CV | 2023-03-09 |
178 | A Unified Arbitrary Style Transfer Framework Via Adaptive Contrastive Learning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present Unified Contrastive Arbitrary Style Transfer (UCAST), a novel style representation learning and transfer framework, which can fit in most existing arbitrary image style transfer models, e.g., CNN-based, ViT-based, and flow-based methods. |
YUXIN ZHANG et. al. | arxiv-cs.CV | 2023-03-08 |
179 | End-to-end Face-swapping Via Adaptive Latent Representation Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes a novel and end-to-end integrated framework for high resolution and attribute preservation face swapping via Adaptive Latent Representation Learning. |
Chenhao Lin; Pengbin Hu; Chao Shen; Qian Li; | arxiv-cs.CV | 2023-03-07 |
180 | Guided Image-to-Image Translation By Discriminator-Generator Communication Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This formulation illustrates the information insufficiency in the GAN training. To mitigate this problem, we propose to add a communication channel between discriminators and generators. |
Yuanjiang Cao; Lina Yao; Le Pan; Quan Z. Sheng; Xiaojun Chang; | arxiv-cs.CV | 2023-03-06 |
181 | Towards Enhanced Controllability of Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose two generic sampling techniques for improving controllability. |
WONWOONG CHO et. al. | arxiv-cs.CV | 2023-02-28 |
182 | Cross-modal Face- and Voice-style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a cross-modal style transfer framework called XFaVoT that jointly learns four tasks: image translation and voice conversion tasks with audio or image guidance, which enables the generation of “face that matches given voice and “voice that matches given face, and intra-modality translation tasks with a single framework. |
Naoya Takahashi; Mayank K. Singh; Yuki Mitsufuji; | arxiv-cs.CV | 2023-02-27 |
183 | ACE: Zero-Shot Image to Image Translation Via Pretrained Auto-Contrastive-Encoder Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, learning such mapping between domains is challenging because data from different domains can be highly unbalanced in terms of both quality and quantity. To address this problem, we propose a new approach to extract image features by learning the similarities and differences of samples within the same data distribution via a novel contrastive learning framework, which we call Auto-Contrastive-Encoder (ACE). |
Sihan Xu; Zelong Jiang; Ruisi Liu; Kaikai Yang; Zhijie Huang; | arxiv-cs.CV | 2023-02-22 |
184 | Entity-Level Text-Guided Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we study a novel task on text-guided image manipulation on the entity level in the real world (eL-TGIM). |
YIKAI WANG et. al. | arxiv-cs.CV | 2023-02-22 |
185 | Paint It Black: Generating Paintings from Text Descriptions Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, the intersection of these two, i.e., generating paintings from a given caption, is a relatively unexplored area with little data available. In this paper, we have explored two distinct strategies and have integrated them together. |
Mahnoor Shahid; Mark Koch; Niklas Schneider; | arxiv-cs.CV | 2023-02-17 |
186 | Conversation Style Transfer Using Few-Shot Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel in-context learning approach to solve the task with style-free dialogues as a pivot. |
SHAMIK ROY et. al. | arxiv-cs.CL | 2023-02-16 |
187 | Transformer-based Generative Adversarial Networks in Computer Vision: A Comprehensive Survey Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a comprehensive survey on the developments and advancements in GANs utilizing the Transformer networks for computer vision applications. |
Shiv Ram Dubey; Satish Kumar Singh; | arxiv-cs.CV | 2023-02-16 |
188 | DiffFaceSketch: High-Fidelity Face Image Synthesis with Sketch-Guided Latent Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce a Sketch-Guided Latent Diffusion Model (SGLDM), an LDM-based network architect trained on the paired sketch-face dataset. |
Yichen Peng; Chunqi Zhao; Haoran Xie; Tsukasa Fukusato; Kazunori Miyata; | arxiv-cs.CV | 2023-02-14 |
189 | DiffFashion: Reference-based Fashion Design with Structure-aware Transfer By Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Although diffusion-based image translation or neural style transfer (NST) has enabled flexible style transfer, it is often difficult to maintain the original structure of the image realistically during the reverse diffusion, especially when the referenced appearance image greatly differs from the common clothing appearance. To tackle this issue, we present a novel diffusion model-based unsupervised structure-aware transfer method to semantically generate new clothes from a given clothing image and a reference appearance image. |
SHIDONG CAO et. al. | arxiv-cs.CV | 2023-02-13 |
190 | Unified Vision-Language Representation Modeling for E-Commerce Same-Style Products Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Common methods use the image as the detected object, but they only consider the visual features and overlook the attribute information contained in the textual descriptions, and perform weakly for products in image less important industries like machinery, hardware tools and electronic component, even if an additional text matching module is added. In this paper, we propose a unified vision-language modeling method for e-commerce same-style products retrieval, which is designed to represent one product with its textual descriptions and visual contents. |
BEN CHEN et. al. | arxiv-cs.IR | 2023-02-10 |
191 | Neural Artistic Style Transfer with Conditional Adversaria Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present two methods that step toward the style image independent neural style transfer model. |
P. N. Deelaka; | arxiv-cs.CV | 2023-02-07 |
192 | Design Booster: A Text-Guided Diffusion Model for Image Translation with Spatial Layout Preservation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Besides, existing methods are mainly based on test-time optimization or fine-tuning model for each input image, which are extremely time-consuming for practical applications. To address these issues, we propose a new approach for flexible image translation by learning a layout-aware image condition together with a text condition. |
Shiqi Sun; Shancheng Fang; Qian He; Wei Liu; | arxiv-cs.CV | 2023-02-04 |
193 | Multi-domain Image Generation and Translation with Identifiability Guarantees Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Inspired by the recent advances in nonlinear Independent Component Analysis (ICA) theory, we propose a new method to learn the joint distribution from the marginals by enforcing a specific type of minimal change across domains. |
Shaoan Xie; Lingjing Kong; Mingming Gong; Kun Zhang; | iclr | 2023-02-01 |
194 | Flow Straight and Fast: Learning to Generate and Transfer Data with Rectified Flow IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present rectified flow, a simple approach to learning (neural) ordinary differential equation (ODE) models to transport between two empirically observed distributions $\pi_0$ and $\pi_1$, hence providing a unified solution to generative modeling and domain transfer, among various other tasks involving distribution transport. |
Xingchao Liu; Chengyue Gong; qiang liu; | iclr | 2023-02-01 |
195 | Edge-guided Multi-domain RGB-to-TIR Image Translation for Training Vision Tasks with Challenging Labels Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: As a remedy, we propose a modified multidomain RGB to TIR image translation model focused on edge preservation to employ annotated RGB images with challenging labels. |
Dong-Guw Lee; Myung-Hwan Jeon; Younggun Cho; Ayoung Kim; | arxiv-cs.CV | 2023-01-30 |
196 | Few-shot Face Image Translation Via GAN Prior Distillation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose GAN Prior Distillation (GPD) to enable effective few-shot face image translation. |
Ruoyu Zhao; Mingrui Zhu; Xiaoyu Wang; Nannan Wang; | arxiv-cs.CV | 2023-01-28 |
197 | ITstyler: Image-optimized Text-based Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we achieve a data-efficient text-based style transfer method that does not require optimization at the inference stage. |
Yunpeng Bai; Jiayue Liu; Chao Dong; Chun Yuan; | arxiv-cs.CV | 2023-01-25 |
198 | Face Attribute Editing with Disentangled Latent Vectors Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose an image-to-image translation framework for facial attribute editing with disentangled interpretable latent directions. |
Yusuf Dalva; Hamza Pehlivan; Cansu Moran; Öykü Irmak Hatipoğlu; Ayşegül Dündar; | arxiv-cs.CV | 2023-01-11 |
199 | Tackling Data Bias in Painting Classification with Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a system to handle data bias in small paintings datasets like the Kaokore dataset while simultaneously accounting for domain adaptation in fine-tuning a model trained on real world images. |
Mridula Vijendran; Frederick W. B. Li; Hubert P. H. Shum; | arxiv-cs.CV | 2023-01-06 |
200 | Edge Enhanced Image Style Transfer Via Transformers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To stylize the image with sufficient style patterns, the content details may be damaged and sometimes the objects of images can not be distinguished clearly. For this reason, we present a new transformer-based method named STT for image style transfer and an edge loss which can enhance the content details apparently to avoid generating blurred results for excessive rendering on style features. |
Chiyu Zhang; Jun Yang; Zaiyan Dai; Peng Cao; | arxiv-cs.CV | 2023-01-02 |
201 | Interactive Control Over Temporal Consistency While Stylizing Video Streams Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Moreover, making this control interactive is paramount from a usability perspective. To achieve the above requirements, we propose an approach that stylizes video streams in real-time at full HD resolutions while providing interactive consistency control. |
SUMIT SHEKHAR et. al. | arxiv-cs.GR | 2023-01-02 |
202 | Treatment Learning Causal Transformer for Noisy Image Classification Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Current top-notch deep learning (DL) based vision models are primarily based on exploring and exploiting the inherent correlations between training data samples and their … |
C. Yang; I-Te Danny Hung; Yi-Chieh Liu; Pin-Yu Chen; | 2023 IEEE/CVF Winter Conference on Applications of Computer … | 2023-01-01 |
203 | Image-to-image Translation with Generative Adversarial Networks Via Retinal Masks for Realistic Optical Coherence Tomography Imaging of Diabetic Macular Edema Disorders Related Papers Related Patents Related Grants Related Venues Related Experts View |
P. Vidal; J. de Moura; J. Novo; M. G. Penedo; M. Ortega; | Biomed. Signal Process. Control. | 2023-01-01 |
204 | Distill The Image to Nowhere: Inversion Knowledge Distillation for Multimodal Machine Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Thus, in this work, we introduce IKD-MMT, a novel MMT framework to support the image-free inference phase via an inversion knowledge distillation scheme. |
Ru Peng; Yawen Zeng; Jake Zhao; | emnlp | 2022-12-30 |
205 | Self-Supervised Pre-training for 3D Point Clouds Via View-Specific Point-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Different from existing pre-training paradigms designed for deep point cloud feature extractors that fall into the scope of generative modeling or contrastive learning, this paper proposes a translative pre-training framework, namely PointVST, driven by a novel self-supervised pretext task of cross-modal translation from 3D point clouds to their corresponding diverse forms of 2D rendered images. |
Qijian Zhang; Junhui Hou; | arxiv-cs.CV | 2022-12-29 |
206 | Scaling Painting Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper provides a solution to solve the original global optimization for ultra-high resolution images, enabling multiscale style transfer at unprecedented image sizes. |
Bruno Galerne; Lara Raad; José Lezama; Jean-Michel Morel; | arxiv-cs.CV | 2022-12-27 |
207 | DSI2I: Dense Style for Unpaired Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Here, by contrast, we propose to represent style as a dense feature map, allowing for a finer-grained transfer to the source image without requiring any external semantic information. |
Baran Ozaydin; Tong Zhang; Sabine Süsstrunk; Mathieu Salzmann; | arxiv-cs.CV | 2022-12-26 |
208 | Meta-Learning for Color-to-Infrared Cross-Modal Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our analysis reveals that existing data-driven methods are either too simplistic or introduce significant artifacts into the imagery. To overcome these limitations, we propose meta-learning style transfer (MLST), which learns a stylization by composing and tuning well-behaved analytic functions. |
Evelyn A. Stump; Francesco Luzi; Leslie M. Collins; Jordan M. Malof; | arxiv-cs.CV | 2022-12-24 |
209 | Artistic Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Despite all the efforts, it’s still a major challenge to apply the artistic style that was originally created on top of the structure of the content image while maintaining consistency. In this work, we solved these problems by using a Deep Learning approach using Convolutional Neural Networks. |
Weiting Li; Rahul Vyas; Ramya Sree Penta; | arxiv-cs.CV | 2022-12-21 |
210 | QuantArt: Quantizing Image Style Transfer Towards High Visual Fidelity Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we devise a new style transfer framework called QuantArt for high visual-fidelity stylization. |
Siyu Huang; Jie An; Donglai Wei; Jiebo Luo; Hanspeter Pfister; | arxiv-cs.CV | 2022-12-20 |
211 | StyleTRF: Stylizing Tensorial Radiance Fields Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we present StyleTRF, a compact, quick-to-optimize strategy for stylized view generation using TensoRF. |
Rahul Goel; Sirikonda Dhawal; Saurabh Saini; P. J. Narayanan; | arxiv-cs.CV | 2022-12-19 |
212 | ColoristaNet for Photorealistic Video Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To avoid employing the popular Gram loss, we propose a self-supervised style transfer framework, which contains a style removal part and a style restoration part. |
XIAOWEN QIU et. al. | arxiv-cs.CV | 2022-12-18 |
213 | Low-Resource Authorship Style Transfer: Can Non-Famous Authors Be Imitated? Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce the \textit{low-resource authorship style transfer} task, a more challenging class of authorship style transfer where only a limited amount of text in the target author’s style may exist. |
Ajay Patel; Nicholas Andrews; Chris Callison-Burch; | arxiv-cs.CL | 2022-12-17 |
214 | Deep Image Style Transfer from Freeform Text Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper creates a novel method of deep neural style transfer by generating style images from freeform user text input. |
Tejas Santanam; Mengyang Liu; Jiangyue Yu; Zhaodong Yang; | arxiv-cs.CV | 2022-12-13 |
215 | CLIPVG: Text-Guided Image Manipulation Using Differentiable Vector Graphics Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce CLIPVG, a text-guided image manipulation framework using differentiable vector graphics, which is also the first CLIP-based general image manipulation framework that does not require any additional generative models. |
YIREN SONG et. al. | arxiv-cs.CV | 2022-12-05 |
216 | Neural Photo-Finishing Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Image processing pipelines are ubiquitous and we rely on them either directly, by filtering or adjusting an image post-capture, or indirectly, as image signal processing (ISP) … |
ETHAN TSENG et. al. | ACM Transactions on Graphics (TOG) | 2022-11-30 |
217 | Touch and Go: Learning from Human-Collected Vision and Touch Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a dataset with paired visual and tactile data called Touch and Go, in which human data collectors probe objects in natural environments using tactile sensors, while simultaneously recording egocentric video. |
FENGYU YANG et. al. | arxiv-cs.CV | 2022-11-22 |
218 | LISA: Localized Image Stylization with Audio Via Implicit Neural Representation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a novel framework, Localized Image Stylization with Audio (LISA) which performs audio-driven localized image stylization. |
SEUNG HYUN LEE et. al. | arxiv-cs.CV | 2022-11-21 |
219 | DiffStyler: Controllable Dual Diffusion for Text-Driven Image Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present DiffStyler on the basis of diffusion models. |
NISHA HUANG et. al. | arxiv-cs.CV | 2022-11-19 |
220 | Single Stage Multi-Pose Virtual Try-On Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel single stage model for MPVTON. |
Sen He; Yi-Zhe Song; Tao Xiang; | arxiv-cs.CV | 2022-11-19 |
221 | Unsupervised 3D Pose Transfer with Cross Consistency and Dual Reconstruction Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we present X-DualNet, a simple yet effective approach that enables unsupervised 3D pose transfer. |
Chaoyue Song; Jiacheng Wei; Ruibo Li; Fayao Liu; Guosheng Lin; | arxiv-cs.CV | 2022-11-18 |
222 | Replacing Language Model for Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce replacing language model (RLM), a sequence-to-sequence language modeling framework for text style transfer. |
Pengyu Cheng; Ruineng Li; | arxiv-cs.CL | 2022-11-14 |
223 | Arbitrary Style Guidance for Enhanced Diffusion-Based Text-to-Image Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Here we propose a novel style guidance method to support generating images using arbitrary style guided by a reference image. |
Zhihong Pan; Xin Zhou; Hao Tian; | arxiv-cs.CV | 2022-11-14 |
224 | Learning Visual Representation of Underwater Acoustic Imagery Using Transformer-Based Style Transfer Method Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This letter proposed a framework for learning the visual representation of underwater acoustic imageries, which takes a transformer-based style transfer model as the main body. |
XIAOTENG ZHOU et. al. | arxiv-cs.CV | 2022-11-10 |
225 | Text-driven Photorealistic 3D Stylization For Arbitrary Meshes Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Technically, we propose to disentangle the appearance style as the spatially varying bidirectional reflectance distribution function, the local geometric variation, and the lighting condition, which are jointly optimized, via supervision of the CLIP loss, by a spherical Gaussians based differentiable renderer. |
yongwei chen; chen rui; Jiabao Lei; Yabin Zhang; Kui Jia; | nips | 2022-11-06 |
226 | Unsupervised Image-to-Image Translation with Density Changing Regularization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we make a density changing assumption where image patches of high probability density should be mapped to patches of high probability density in another domain. |
Shaoan Xie; Qirong Ho; Kun Zhang; | nips | 2022-11-06 |
227 | GenerSpeech: Towards Style Transfer for Generalizable Out-Of-Domain Text-to-Speech Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes GenerSpeech, a text-to-speech model towards high-fidelity zero-shot style transfer of OOD custom voice. |
Rongjie Huang; Yi Ren; Jinglin Liu; Chenye Cui; Zhou Zhao; | nips | 2022-11-06 |
228 | Generalized One-shot Domain Adaption of Generative Adversarial Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Besides, to realize cross-domain correspondence, we propose the variational Laplacian regularization to constrain the smoothness of the adapted generator. |
ZICHENG ZHANG et. al. | nips | 2022-11-06 |
229 | Dense Interspecies Face Embedding Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce a new task of cross-domain face understanding, and propose a dense interspecies face embedding (DIFE) learned in an unsupervised manner by our multi-teacher knowledge distillation and pseudo-paired data synthesis. |
Sejong Yang; Subin Jeon; Seonghyeon Nam; Seon Joo Kim; | nips | 2022-11-06 |
230 | Text-Only Training for Image Captioning Using Noise-Injected CLIP IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We consider the task of image-captioning using only the CLIP model and additional text data at training time, and no additional captioned images. |
David Nukrai; Ron Mokady; Amir Globerson; | arxiv-cs.CV | 2022-11-01 |
231 | Image-to-Image Translation-Based Data Augmentation for Improving Crop/Weed Classification Models for Precision Agriculture Applications IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Applications of deep-learning models in machine visions for crop/weed identification have remarkably upgraded the authenticity of precise weed management. However, compelling data … |
L. G. DIVYANTH et. al. | Algorithms | 2022-10-30 |
232 | MagicMix: Semantic Mixing with Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Have you ever imagined what a corgi-alike coffee machine or a tiger-alike rabbit would look like? In this work, we attempt to answer these questions by exploring a new task called semantic mixing, aiming at blending two different semantics to create a new concept (e.g., corgi + coffee machine — > corgi-alike coffee machine). |
Jun Hao Liew; Hanshu Yan; Daquan Zhou; Jiashi Feng; | arxiv-cs.CV | 2022-10-28 |
233 | Efficient Hair Style Transfer with Generative Adversarial Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The current state-of-the-art hair synthesis approaches struggle to maintain global composition of the target style and cannot be used in real-time applications due to their high running costs on high-resolution portrait images. Therefore, We propose a novel hairstyle transfer method, called EHGAN, which reduces computational costs to enable real-time processing while improving the transfer of hairstyle with better global structure compared to the other state-of-the-art hair synthesis methods. |
Muhammed Pektas; Baris Gecer; Aybars Ugur; | arxiv-cs.CV | 2022-10-22 |
234 | S2WAT: Image Style Transfer Via Hierarchical Vision Transformer Using Strips Window Attention Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper presents a new hierarchical vision Transformer for image style transfer, called Strips Window Attention Transformer (S2WAT), which serves as an encoder of encoder-transfer-decoder architecture. |
Chiyu Zhang; Jun Yang; Lei Wang; Zaiyan Dai; | arxiv-cs.CV | 2022-10-22 |
235 | TANGO: Text-driven Photorealistic and Robust 3D Stylization Via Lighting Decomposition IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we focus on stylizing photorealistic appearance renderings of a given surface mesh of arbitrary topology. |
Yongwei Chen; Rui Chen; Jiabao Lei; Yabin Zhang; Kui Jia; | arxiv-cs.CV | 2022-10-20 |
236 | Image-Based CLIP-Guided Essence Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Our blending operator combines the powerful StyleGAN generator and the semantic encoder of CLIP in a novel way that is simultaneously additive in both latent spaces, resulting in a mechanism that guarantees both identity preservation and high-level feature transfer without relying on a facial recognition network. |
Hila Chefer; Sagie Benaim; Roni Paiss; Lior Wolf; | eccv | 2022-10-19 |
237 | Multi-Curve Translator for High-Resolution Photorealistic Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To this end, we present the Multi-Curve Translator (MCT), which not only predicts the translated pixels for the corresponding input pixels but also for their neighboring pixels. |
Yuda Song; Hui Qian; Xin Du; | eccv | 2022-10-19 |
238 | Learning Visual Styles from Audio-Visual Associations IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a method for learning visual styles from unlabeled audio-visual data. |
Tingle Li; Yichen Liu; Andrew Owens; Hang Zhao; | eccv | 2022-10-19 |
239 | WISE: Whitebox Image Stylization By Example-Based Learning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, adapting or extending these techniques to produce new styles is often a tedious and error-prone task that requires expert knowledge. We propose a new paradigm to alleviate this problem: implementing algorithmic image filtering techniques as differentiable operations that can learn parametrizations aligned to certain reference styles. |
TZSCH WINFRIED L&OUML et. al. | eccv | 2022-10-19 |
240 | Unpaired Image Translation Via Vector Symbolic Architectures IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, if the source and target domains have a large semantic mismatch, existing techniques often suffer from source content corruption aka semantic flipping. To address this problem, we propose a new paradigm for image-to-image translation using Vector Symbolic Architectures (VSA), a theoretical framework which defines algebraic operations in a high-dimensional vector (hypervector) space. |
Justin Theiss; Jay Leverett; Daeil Kim; Aayush Prakash; | eccv | 2022-10-19 |
241 | K-SALSA: K-Anonymous Synthetic Averaging of Retinal Images Via Local Style Alignment Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: While prior works have explored image de-identification strategies based on synthetic averaging of images in other domains (e.g. facial images), existing techniques face difficulty in preserving both privacy and clinical utility in retinal images, as we demonstrate in our work. We therefore introduce k-SALSA, a generative adversarial network (GAN)-based framework for synthesizing retinal fundus images that summarize a given private dataset while satisfying the privacy notion of k-anonymity. |
Minkyu Jeon; Hyeonjin Park; Hyunwoo J. Kim; Michael Morley; Hyunghoon Cho; | eccv | 2022-10-19 |
242 | TIPS: Text-Induced Pose Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we first present the shortcomings of current pose transfer algorithms and then propose a novel text-based pose transfer technique to address those issues. |
Prasun Roy; Subhankar Ghosh; Saumik Bhattacharya; Umapada Pal; Michael Blumenstein; | eccv | 2022-10-19 |
243 | Cross Attention Based Style Distribution for Controllable Person Image Synthesis IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a cross attention based style distribution module that computes between the source semantic styles and target pose for pose transfer. |
XINYUE ZHOU et. al. | eccv | 2022-10-19 |
244 | SCAM! Transferring Humans Between Images with Semantic Cross Attention Modulation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we introduce SCAM (Semantic Cross Attention Modulation), a system that encodes rich and diverse information in each semantic region of the image (including foreground and background), thus achieving precise generation with emphasis on fine details. |
Nicolas Dufour; David Picard; Vicky Kalogeiton; | eccv | 2022-10-19 |
245 | CCPL: Contrastive Coherence Preserving Loss for Versatile Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we aim to devise a universally versatile style transfer method capable of performing artistic, photo-realistic, and video style transfer jointly, without seeing videos during training. |
Zijie Wu; Zhen Zhu; Junping Du; Xiang Bai; | eccv | 2022-10-19 |
246 | Skeleton-Free Pose Transfer for Stylized 3D Characters IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present the first method that automatically transfers poses between stylized 3D characters without skeletal rigging. |
Zhouyingcheng Liao; Jimei Yang; Jun Saito; Gerard Pons-Moll; Yang Zhou; | eccv | 2022-10-19 |
247 | Generative Domain Adaptation for Face Anti-Spoofing IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, insufficient supervision of unlabeled target domains and neglect of low-level feature alignment degrade the performances of existing methods. To address these issues, we propose a novel perspective of UDA FAS that directly fits the target data to the models, i.e., stylizes the target data to the source-domain style via image translation, and further feeds the stylized data into the well-trained source model for classification. |
QIANYU ZHOU et. al. | eccv | 2022-10-19 |
248 | Language-Driven Artistic Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce a new task—language-driven artistic style transfer (LDAST)—to manipulate the style of a content image, guided by a text. |
Tsu-Jui Fu; Xin Eric Wang; William Yang Wang; | eccv | 2022-10-19 |
249 | ManiFest: Manifold Deformation for Few-Shot Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We instead propose ManiFest: a framework for few-shot image translation that learns a context-aware representation of a target domain from a few images only. |
Fabio Pizzati; ois Lalonde Jean-Franç Raoul de Charette; | eccv | 2022-10-19 |
250 | Vector Quantized Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose introducing the vector quantization technique into the image-to-image translation framework. |
Yu-Jie Chen; Shin-I Cheng; Wei-Chen Chiu; Hung-Yu Tseng; Hsin-Ying Lee; | eccv | 2022-10-19 |
251 | Harmonizer: Learning to Perform White-Box Image and Video Harmonization IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we observe that adjusting the input arguments of basic image filters, e.g., brightness and contrast, is sufficient for humans to produce realistic images from the composite ones. |
Zhanghan Ke; Chunyi Sun; Lei Zhu; Ke Xu; Rynson W.H. Lau; | eccv | 2022-10-19 |
252 | Bi-Level Feature Alignment for Versatile Image Translation and Manipulation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper presents a versatile image translation and manipulation framework that achieves accurate semantic and style guidance in image generation by explicitly building a correspondence. |
FANGNENG ZHAN et. al. | eccv | 2022-10-19 |
253 | Interpolated SelectionConv for Spherical Images and Surfaces Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present a new and general framework for convolutional neural network operations on spherical (or omnidirectional) images. |
David Hart; Michael Whitney; Bryan Morse; | arxiv-cs.CV | 2022-10-18 |
254 | Controllable Style Transfer Via Test-time Training of Implicit Neural Representation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a controllable style transfer framework based on Implicit Neural Representation that pixel-wisely controls the stylized output via test-time training. |
Sunwoo Kim; Youngjo Min; Younghun Jung; Seungryong Kim; | arxiv-cs.CV | 2022-10-14 |
255 | Synthetic-to-real Composite Semantic Segmentation in Additive Manufacturing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work demonstrates the possibilities of using physics-based rendering for labeled image dataset generation, as well as image-to-image translation capabilities to improve the accuracy of real image segmentation for AM systems. |
Aliaksei Petsiuk; Harnoor Singh; Himanshu Dadhwal; Joshua M. Pearce; | arxiv-cs.CV | 2022-10-13 |
256 | Line Search-Based Feature Transformation for Fast, Stable, and Tunable Content-Style Control in Photorealistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce a general-purpose transformation that enables controlling the balance between how much content is preserved and the strength of the infused style. |
Tai-Yin Chiu; Danna Gurari; | arxiv-cs.CV | 2022-10-12 |
257 | Fine-Grained Image Style Transfer with Visual Transformers Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Such a design usually destroys the spatial information of the input images and fails to transfer fine-grained style patterns into style transfer results. To solve this problem, we propose a novel STyle TRansformer (STTR) network which breaks both content and style images into visual tokens to achieve a fine-grained style transformation. |
Jianbo Wang; Huan Yang; Jianlong Fu; Toshihiko Yamasaki; Baining Guo; | arxiv-cs.CV | 2022-10-11 |
258 | D2Animator: Dual Distillation of StyleGAN For High-Resolution Face Animation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The style-based generator architectures (e.g. StyleGAN v1, v2) largely promote the controllability and explainability of Generative Adversarial Networks (GANs). Many researchers … |
Zhuo Chen; Chaoyue Wang; Haimei Zhao; Bo Yuan; Xiu Li; | Proceedings of the 30th ACM International Conference on … | 2022-10-10 |
259 | Detach and Attach: Stylized Image Captioning Without Paired Stylized Dataset Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Stylized Image Captioning aims to generate captions with accurate image content and stylized elements simultaneously. However, large-scaled image and stylized caption pairs cost … |
YUTONG TAN et. al. | Proceedings of the 30th ACM International Conference on … | 2022-10-10 |
260 | Distill The Image to Nowhere: Inversion Knowledge Distillation for Multimodal Machine Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Thus, in this work, we introduce IKD-MMT, a novel MMT framework to support the image-free inference phase via an inversion knowledge distillation scheme. |
Ru Peng; Yawen Zeng; Junbo Zhao; | arxiv-cs.CL | 2022-10-10 |
261 | Adaptive Affine Transformation: A Simple and Effective Operation for Spatial Misaligned Image Generation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: One challenging problem, named spatial misaligned image generation, describing a translation between two face/pose images with large spatial deformation, is widely faced in tasks … |
Zhimeng Zhang; Yu-qiong Ding; | Proceedings of the 30th ACM International Conference on … | 2022-10-10 |
262 | Bridging CLIP and StyleGAN Through Latent Alignment for Image Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we manage to achieve inference-time optimization-free diverse manipulation direction mining by bridging CLIP and StyleGAN through Latent Alignment (CSLA). |
Wanfeng Zheng; Qiang Li; Xiaoyan Guo; Pengfei Wan; Zhongyuan Wang; | arxiv-cs.CV | 2022-10-10 |
263 | MultiStyleGAN: Multiple One-shot Image Stylizations Using A Single GAN Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we present a MultiStyleGAN method that is capable of producing multiple different stylizations at once by fine-tuning a single generator. |
Viraj Shah; Ayush Sarkar; Sudharsan Krishnakumar Anitha; Svetlana Lazebnik; | arxiv-cs.CV | 2022-10-08 |
264 | CLIP-PAE: Projection-Augmentation Embedding to Extract Relevant Features for A Disentangled, Interpretable, and Controllable Text-Guided Face Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Disentanglement, interpretability, and controllability are also hard to guarantee for manipulation. To alleviate these problems, we propose to define corpus subspaces spanned by relevant prompts to capture specific image characteristics. |
Chenliang Zhou; Fangcheng Zhong; Cengiz Oztireli; | arxiv-cs.CV | 2022-10-08 |
265 | Pose Guided Human Image Synthesis with Partially Decoupled GAN Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, it is difficult to recover the detailed texture of the whole human image. To alleviate this problem, we propose a method by decoupling the human body into several parts (\eg, hair, face, hands, feet, \etc) and then using each of these parts to guide the synthesis of a realistic image of the person, which preserves the detailed information of the generated images. |
Jianhan Wu; Jianzong Wang; Shijing Si; Xiaoyang Qu; Jing Xiao; | arxiv-cs.CV | 2022-10-07 |
266 | Improving End-to-End Text Image Translation From The Auxiliary Text Translation Task Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a novel text translation enhanced text image translation, which trains the end-to-end model with text translation as an auxiliary task. |
CONG MA et. al. | arxiv-cs.CL | 2022-10-07 |
267 | FastCLIPstyler: Optimisation-free Text-based Image Style Transfer Using Style Representations Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we create a generalised text-based style transfer network capable of stylising images in a single forward pass for an arbitrary text input making the image stylisation process around 1000 times more efficient than CLIPstyler. |
Ananda Padhmanabhan Suresh; Sanjana Jain; Pavit Noinongyao; Ankush Ganguly; | arxiv-cs.CV | 2022-10-07 |
268 | LDEdit: Towards Generalized Text Guided Image Manipulation Via Latent Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose an optimization-free method for the task of generic image manipulation from text prompts. |
Paramanand Chandramouli; Kanchana Vaishnavi Gandikota; | arxiv-cs.CV | 2022-10-05 |
269 | Collecting The Puzzle Pieces: Disentangled Self-Driven Human Pose Transfer By Permuting Textures Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose Pose Transfer by Permuting Textures (PT$^2$), an approach for self-driven human pose transfer that disentangles pose from texture at the patch-level. |
Nannan Li; Kevin J. Shih; Bryan A. Plummer; | arxiv-cs.CV | 2022-10-04 |
270 | Federated Domain Generalization for Image Recognition Via Cross-Client Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a novel domain generalization method for image recognition under federated learning through cross-client style transfer (CCST) without exchanging data samples. |
Junming Chen; Meirui Jiang; Qi Dou; Qifeng Chen; | arxiv-cs.CV | 2022-10-03 |
271 | Diffusion-based Image Translation Using Disentangled Style and Content Representation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Unfortunately, due to the stochastic nature of diffusion models, it is often difficult to maintain the original content of the image during the reverse diffusion. To address this, here we present a novel diffusion-based unsupervised image translation method using disentangled style and content representation. |
Gihyun Kwon; Jong Chul Ye; | arxiv-cs.CV | 2022-09-30 |
272 | PerSign: Personalized Bangladeshi Sign Letters Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Bangladeshi Sign Language (BdSL) – like other sign languages – is tough to learn for general people, especially when it comes to expressing letters. In this poster, we propose PerSign, a system that can reproduce a person’s image by introducing sign gestures in it. |
Mohammad Imrul Jubair; Ali Ahnaf; Tashfiq Nahiyan Khan; Ullash Bhattacharjee; Tanjila Joti; | arxiv-cs.CV | 2022-09-29 |
273 | Image-to-Image Translation for Autonomous Driving from Coarsely-Aligned Image Pairs Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we investigate the idea of turning sensor inputs (i.e., images) captured in an adverse condition into a benign one (i.e., sunny), upon which the downstream tasks (e.g., semantic segmentation) can attain high accuracy. |
YOUYA XIA et. al. | arxiv-cs.CV | 2022-09-23 |
274 | StyleTime: Style Transfer for Synthetic Time Series Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, a novel formulation of time series style transfer is proposed for the purpose of synthetic data generation and enhancement. |
Yousef El-Laham; Svitlana Vyetrenko; | arxiv-cs.LG | 2022-09-22 |
275 | VToonify: Controllable High-Resolution Portrait Video Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we investigate the challenging controllable high-resolution portrait video style transfer by introducing a novel VToonify framework. |
Shuai Yang; Liming Jiang; Ziwei Liu; Chen Change Loy; | arxiv-cs.CV | 2022-09-22 |
276 | T2V-DDPM: Thermal to Visible Face Translation Using Denoising Diffusion Probabilistic Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a Denoising Diffusion Probabilistic Model (DDPM) based solution for T2V translation specifically for facial images. |
Nithin Gopalakrishnan Nair; Vishal M. Patel; | arxiv-cs.CV | 2022-09-19 |
277 | High-resolution Semantically-consistent Image-to-image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work proposes an unsupervised domain adaptation model that preserves semantic consistency and per-pixel quality for the images during the style-transferring phase. |
MIKHAIL SOKOLOV et. al. | arxiv-cs.CV | 2022-09-13 |
278 | Time-of-Day Neural Style Transfer for Architectural Photographs Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we specialize a neural style transfer method for architectural photography. |
Yingshu Chen; Tuan-Anh Vu; Ka-Chun Shum; Binh-Son Hua; Sai-Kit Yeung; | arxiv-cs.CV | 2022-09-13 |
279 | Generalized One-shot Domain Adaptation of Generative Adversarial Networks IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we focus on the one-shot case, which is more challenging and rarely explored in previous works. |
ZICHENG ZHANG et. al. | arxiv-cs.CV | 2022-09-08 |
280 | AWADA: Attention-Weighted Adversarial Domain Adaptation for Object Detection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose AWADA, an Attention-Weighted Adversarial Domain Adaptation framework for creating a feedback loop between style-transformation and detection task. |
Maximilian Menke; Thomas Wenzel; Andreas Schwung; | arxiv-cs.CV | 2022-08-31 |
281 | Robust Sound-Guided Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel approach that first extends the image-text joint embedding space with sound and applies a direct latent optimization method to manipulate a given image based on audio input, e.g., the sound of rain. |
SEUNG HYUN LEE et. al. | arxiv-cs.CV | 2022-08-30 |
282 | Prior-Aware Synthetic Data to The Rescue: Animal Pose Estimation with Very Limited Real Data Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, when faced with serious data scarcity (i.e., $<10^2$ real images), the model performance stays unsatisfactory, particularly for limbs with considerable flexibility and several comparable parts. We therefore introduce a prior-aware synthetic animal data generation pipeline called PASyn to augment the animal pose data essential for robust pose estimation. |
Le Jiang; Shuangjun Liu; Xiangyu Bai; Sarah Ostadabbas; | arxiv-cs.CV | 2022-08-29 |
283 | AesUST: Towards Aesthetic-Enhanced Universal Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, existing approaches suffer from the aesthetic-unrealistic problem that introduces disharmonious patterns and evident artifacts, making the results easy to spot from real paintings. To address this limitation, we propose AesUST, a novel Aesthetic-enhanced Universal Style Transfer approach that can generate aesthetically more realistic and pleasing results for arbitrary styles. |
ZHIZHONG WANG et. al. | arxiv-cs.CV | 2022-08-27 |
284 | Selective Manipulation of Disentangled Representations for Privacy-aware Facial Image Processing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, users are increasingly concerned about privacy issues such as function creep and malicious usage by third-party cloud providers. To alleviate this, we propose an edge-based filtering stage that removes privacy-sensitive attributes before the sensor data are transmitted to the cloud. |
Sander De Coninck; Wei-Cheng Wang; Sam Leroux; Pieter Simoens; | arxiv-cs.CV | 2022-08-26 |
285 | Unsupervised Structure-Consistent Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We improve this work by introducing a simple yet effective auxiliary module based on gradient reversal layers. |
Shima Shahfar; Charalambos Poullis; | arxiv-cs.CV | 2022-08-24 |
286 | Dance Style Transfer with Cross-modal Transformer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present CycleDance, a dance style transfer system to transform an existing motion clip in one dance style to a motion clip in another dance style while attempting to preserve motion context of the dance. |
Wenjie Yin; Hang Yin; Kim Baraka; Danica Kragic; Mårten Björkman; | arxiv-cs.LG | 2022-08-19 |
287 | UPST-NeRF: Universal Photorealistic Style Transfer of Neural Radiance Fields for 3D Scene IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Therefore, a styling image must retrain a 3D scene representation network based on a neural radiation field. We propose a novel 3D scene photorealistic style transfer framework to address these issues. |
YAOSEN CHEN et. al. | arxiv-cs.CV | 2022-08-15 |
288 | Style Spectroscope: Improve Interpretability and Controllability Through Fourier Analysis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we provide a systematic Fourier analysis on a general framework for UST. |
Zhiyu Jin; Xuli Shen; Bin Li; Xiangyang Xue; | arxiv-cs.CV | 2022-08-12 |
289 | Towards Cross-speaker Reading Style Transfer on Audiobook Dataset Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper aims to introduce a chunk-wise multi-scale cross-speaker style model to capture both the global genre and the local prosody in audiobook speeches. |
XIANG LI et. al. | arxiv-cs.SD | 2022-08-10 |
290 | Language-Guided Face Animation By Recurrent StyleGAN-based Generator Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we leverage the motion information and study a novel task, language-guided face animation, that aims to animate a static face image with the help of languages. |
TIANKAI HANG et. al. | arxiv-cs.CV | 2022-08-10 |
291 | HyperNST: Hyper-Networks for Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present HyperNST; a neural style transfer (NST) technique for the artistic stylization of images, based on Hyper-networks and the StyleGAN2 architecture. |
DAN RUTA et. al. | arxiv-cs.CV | 2022-08-09 |
292 | Neural Style Transfer for Image Steganography and Destylization with Supervised Image to Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View |
Mallika Garg; J. S. Ubhi; A. Aggarwal; | Multimedia Tools and Applications | 2022-08-08 |
293 | Learning to Incorporate Texture Saliency Adaptive Attention to Image Cartoonization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Existing advanced models try to enhance cartoonization effect by learning to promote edges adversarially, introducing style transfer loss, or learning to align style from multiple representation space. |
Xiang Gao; Yuqi Zhang; Yingjie Tian; | arxiv-cs.CV | 2022-08-02 |
294 | AdaWCT: Adaptive Whitening and Coloring Style Injection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we present a generalization of AdaIN which relies on the whitening and coloring transformation (WCT) which we dub AdaWCT, that we apply for style injection in large GANs. |
ANTOINE DUFOUR et. al. | arxiv-cs.CV | 2022-08-01 |
295 | Quality Evaluation of Arbitrary Style Transfer: Subjective Study and Objective Metric Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To quantitatively measure the quality of AST image, we propose a new sparse representation-based method, which computes the quality according to the sparse feature similarity. |
HANGWEI CHEN et. al. | arxiv-cs.CV | 2022-08-01 |
296 | WISE: Whitebox Image Stylization By Example-based Learning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, adapting or extending these techniques to produce new styles is often a tedious and error-prone task that requires expert knowledge. We propose a new paradigm to alleviate this problem: implementing algorithmic image filtering techniques as differentiable operations that can learn parametrizations aligned to certain reference styles. |
WINFRIED LÖTZSCH et. al. | arxiv-cs.CV | 2022-07-29 |
297 | Domain Enhanced Arbitrary Image Style Transfer Via Contrastive Learning IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we tackle the challenging problem of arbitrary image style transfer using a novel style feature representation learning method. |
YUXIN ZHANG et. al. | siggraph | 2022-07-28 |
298 | SNeRF: Stylized Neural Implicit Representations for 3D Scenes IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a stylized novel view synthesis method. |
Thu Nguyen-Phuoc; Feng Liu; Lei Xiao; | siggraph | 2022-07-28 |
299 | Photo-to-shape Material Transfer for Diverse Structures Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce a method for assigning photorealistic relightable materials to 3D shapes in an automatic manner. |
Ruizhen Hu; Xiangyu Su; Xiangkai Chen; Oliver Van Kaick; Hui Huang; | siggraph | 2022-07-28 |
300 | Palette: Image-to-Image Diffusion Models IF:6 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper develops a unified framework for image-to-image translation based on conditional diffusion models and evaluates this framework on four challenging image-to-image translation tasks, namely colorization, inpainting, uncropping, and JPEG restoration. |
CHITWAN SAHARIA et. al. | siggraph | 2022-07-28 |
301 | DCT-net: Domain-calibrated Translation for Portrait Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper introduces DCT-Net, a novel image translation architecture for few-shot portrait stylization. |
Yifang Men; Yuan Yao; Miaomiao Cui; Zhouhui Lian; Xuansong Xie; | siggraph | 2022-07-28 |
302 | SAR-to-EO Image Translation with Multi-Conditional Adversarial Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper explores the use of multi-conditional adversarial networks for SAR-to-EO image translation. |
Armando Cabrera; Miriam Cha; Prafull Sharma; Michael Newey; | arxiv-cs.CV | 2022-07-26 |
303 | PTGCF: Printing Texture Guided Color Fusion for Impressionism Oil Painting Style Rendering Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: As a major branch of Non-Photorealistic Rendering (NPR), image stylization mainly uses the computer algorithms to render a photo into an artistic painting. |
Jing Geng; Li’e Ma; Xiaoquan Li; Yijun Yan; | arxiv-cs.CV | 2022-07-25 |
304 | ArtFID: Quantitative Evaluation of Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To support the automatic comparison of different style transfer approaches and to study their respective strengths and weaknesses, the field would greatly benefit from a quantitative measurement of stylization performance. Therefore, we propose a method to complement the currently mostly qualitative evaluation schemes. |
Matthias Wright; Björn Ommer; | arxiv-cs.CV | 2022-07-25 |
305 | Learning Graph Neural Networks for Image Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we study a novel semi-parametric neural style transfer framework that alleviates the deficiency of both parametric and non-parametric stylization. |
YONGCHENG JING et. al. | arxiv-cs.CV | 2022-07-24 |
306 | Generative Artisan: A Semantic-Aware and Controllable CLIPstyler Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Although CLIPstyler can achieve good performance when landscapes or portraits appear alone, it can blur the people and lose the original semantics when people and landscapes coexist. Based on these issues, we demonstrate a novel framework that uses a pre-trained CLIP text-image embedding model and guides image style transfer through an FCN semantic segmentation network. |
Zhenling Yang; Huacheng Song; Qiunan Wu; | arxiv-cs.CV | 2022-07-23 |
307 | Arbitrary Style Transfer with Structure Enhancement By Combining The Global and Local Loss Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: As the result, the stylized images distribute style elements evenly throughout the image and make the overall image structure unrecognizable. To solve this problem, we introduce a novel arbitrary style transfer method with structure enhancement by combining the global and local loss. |
Lizhen Long; Chi-Man Pun; | arxiv-cs.CV | 2022-07-23 |
308 | Learning Cooperative Neural Modules for Stylized Image Captioning Related Papers Related Patents Related Grants Related Venues Related Experts View |
Xinxiao Wu; Wentian Zhao; Jiebo Luo; | International Journal of Computer Vision | 2022-07-22 |
309 | Few-shot Image Generation Using Discrete Content Representation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we make the first attempt to adapt few-shot image translation method to few-shot image generation task. |
Yan Hong; Li Niu; Jianfu Zhang; Liqing Zhang; | arxiv-cs.CV | 2022-07-21 |
310 | GenText: Unsupervised Artistic Text Generation Via Decoupled Font and Texture Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Existing artistic text generation models usually focus on manipulating one aspect of the above components, which is a sub-optimal solution for controllable general artistic text generation. To remedy this issue, we propose a novel approach, namely GenText, to achieve general artistic text style transfer by separably migrating the font and texture styles from the different source images to the target images in an unsupervised manner. |
Qirui Huang; Bin Fu; Aozhong zhang; Yu Qiao; | arxiv-cs.CV | 2022-07-20 |
311 | Capabilities, Limitations and Challenges of Style Transfer with CycleGANs: A Study on Automatic Ring Design Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our contribution entails a crucial part of the process as it allows the customer to see the final product before buying. |
Tomas Cabezon Pedroso; Javier Del Ser; Natalia Diaz-Rodrıguez; | arxiv-cs.CV | 2022-07-18 |
312 | Effect of Instance Normalization on Fine-Grained Control for Sketch-Based Face Image Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we comprehensively investigate the effect of instance normalization on generating photorealistic face images from hand-drawn sketches. |
Zhihua Cheng; Xuejin Chen; | arxiv-cs.CV | 2022-07-17 |
313 | Experiments on Anomaly Detection in Autonomous Driving By Forward-Backward Style Transfers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Especially inspired by works on anomaly detection based on image resynthesis, we propose a novel approach for anomaly detection through style transfer. |
Daniel Bogdoll; Meng Zhang; Maximilian Nitsche; J. Marius Zöllner; | arxiv-cs.CV | 2022-07-13 |
314 | DynaST: Dynamic Sparse Transformer for Exemplar-Guided Image Generation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a dynamic sparse attention based Transformer model, termed Dynamic Sparse Transformer (DynaST), to achieve fine-level matching with favorable efficiency. |
Songhua Liu; Jingwen Ye; Sucheng Ren; Xinchao Wang; | arxiv-cs.CV | 2022-07-13 |
315 | Unsupervised Joint Image Transfer and Uncertainty Quantification Using Patch Invariant Networks Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper presents a novel method for uni-directional domain mapping that does not rely on any paired training data. |
Christoph Angermann; Markus Haltmeier; Ahsan Raza Siyal; | arxiv-cs.CV | 2022-07-09 |
316 | Towards Robust and Semantically Organised Latent Representations for Unsupervised Text Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce EPAAEs (Embedding Perturbed Adversarial AutoEncoders) which completes this perturbation model, by adding a finely adjustable noise component on the continuous embeddings space. |
Sharan Narasimhan; Suvodip Dey; Maunendra Desarkar; | naacl | 2022-07-09 |
317 | StyleFlow For Content-Fixed Image to Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To achieve content-preserving transfer in strongly constrained and normally constrained tasks, we propose StyleFlow, a new I2I translation model that consists of normalizing flows and a novel Style-Aware Normalization (SAN) module. |
Weichen Fan; Jinghuan Chen; Jiabin Ma; Jun Hou; Shuai Yi; | arxiv-cs.CV | 2022-07-05 |
318 | Harmonizer: Learning to Perform White-Box Image and Video Harmonization IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we observe that adjusting the input arguments of basic image filters, e.g., brightness and contrast, is sufficient for humans to produce realistic images from the composite ones. |
Zhanghan Ke; Chunyi Sun; Lei Zhu; Ke Xu; Rynson W. H. Lau; | arxiv-cs.CV | 2022-07-04 |
319 | NARRATE: A Normal Assisted Free-View Portrait Stylizer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose NARRATE, a novel pipeline that enables simultaneously editing portrait lighting and perspective in a photorealistic manner. |
YOUJIA WANG et. al. | arxiv-cs.CV | 2022-07-03 |
320 | Real-Time Portrait Stylization on The Edge Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work we demonstrate real-time portrait stylization, specifically, translating self-portrait into cartoon or anime style on mobile devices. |
YANYU LI et. al. | ijcai | 2022-07-01 |
321 | DivSwapper: Towards Diversified Patch-based Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, as another widespread research interest, the diversity of patch-based methods remains challenging due to the stereotyped style swapping process based on nearest patch matching. To resolve this dilemma, in this paper, we dive into the crux of existing patch-based methods and propose a universal and efficient module, termed DivSwapper, for diversified patch-based arbitrary style transfer. |
ZHIZHONG WANG et. al. | ijcai | 2022-07-01 |
322 | Unsupervised Multi-Modal Medical Image Registration Via Discriminator-Free Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a novel translation-based unsupervised deformable image registration approach to convert the multi-modal registration problem to a mono-modal one. |
Zekang Chen; Jia Wei; Rui Li; | ijcai | 2022-07-01 |
323 | Image Translation Based Synthetic Data Generation for Industrial Object Detection and Pose Estimation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Deep learning-based methods have shown excellent potential on object detection and pose estimation with vast amounts of training data to achieve good performance. Obtaining enough … |
Xu Yang; X. Fan; Jinge Wang; K. Lee; | IEEE Robotics and Automation Letters | 2022-07-01 |
324 | Unsupervised Misaligned Infrared and Visible Image Fusion Via Cross-Modality Image Generation and Registration IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To overcome the obstacles, in this paper, we present a robust cross-modality generation-registration paradigm for unsupervised misaligned infrared and visible image fusion (IVIF). |
Di Wang; Jinyuan Liu; Xin Fan; Risheng Liu; | ijcai | 2022-07-01 |
325 | SNeRF IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: This paper presents a stylized novel view synthesis method. Applying state-of-the-art stylization methods to novel views frame by frame often causes jittering artifacts due to the … |
Thu Nguyen-Phuoc; Feng Liu; Lei Xiao; | ACM Transactions on Graphics (TOG) | 2022-07-01 |
326 | Noise and Edge Based Dual Branch Image Manipulation Detection Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, the noise image extracted by the improved constrained convolution is used as the input of the model instead of the original image to obtain more subtle traces of manipulation. |
Zhongyuan Zhang; Yi Qian; Yanxiang Zhao; Lin Zhu; Jinjin Wang; | arxiv-cs.CV | 2022-07-01 |
327 | StyleCLIPDraw: Coupling Content and Style in Text-to-Drawing Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present an approach for generating styled drawings for a given text description where a user can specify a desired drawing style using a sample image. |
Peter Schaldenbrand; Zhixuan Liu; Jean Oh; | ijcai | 2022-07-01 |
328 | Universal Video Style Transfer Via Crystallization, Separation, and Blending Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, how to maintain the temporal consistency of videos while achieving high-quality arbitrary style transfer is still a hard nut to crack. To resolve this dilemma, in this paper, we propose the CSBNet which involves three key modules: 1) the Crystallization (Cr) Module that generates several orthogonal crystal nuclei, representing hierarchical stability-aware content and style components, from raw VGG features; 2) the Separation (Sp) Module that separates these crystal nuclei to generate the stability-enhanced content and style features; 3) the Blending (Bd) Module to cross-blend these stability-enhanced content and style features, producing more stable and higher-quality stylized videos. |
Haofei Lu; Zhizhong Wang; | ijcai | 2022-07-01 |
329 | Neural Neural Textures Make Sim2Real Consistent Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a new approach that combines differentiable rendering with image translation to achieve temporal consistency over indefinite timescales, using surface consistency losses and \emph{neural neural textures}. |
Ryan Burgert; Jinghuan Shang; Xiang Li; Michael Ryoo; | arxiv-cs.CV | 2022-06-27 |
330 | Non-Parametric Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we extend the second-order statistical feature matching into a general distribution matching based on the understanding that style of an image is represented by the distribution of responses from receptive fields. |
Jeong-Sik Lee; Hyun-Chul Choi; | arxiv-cs.CV | 2022-06-26 |
331 | Review Neural Networks About Image Transformation Based on IGC Learning Framework with Annotated Information Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes a novel learning framework including Independent learning, Guided learning, and Cooperative learning, called the IGC learning framework. |
Yuanjie Yan; Suorong Yang; Yan Wang; Jian Zhao; Furao Shen; | arxiv-cs.CV | 2022-06-21 |
332 | Studying The Role of Named Entities for Content Preservation in Text Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Besides, we perform an error analysis of a pre-trained formality transfer model and introduce a simple technique to use information about named entities to enhance the performance of baseline content similarity measures used in text style transfer. |
Nikolay Babakov; David Dale; Varvara Logacheva; Irina Krotova; Alexander Panchenko; | arxiv-cs.CL | 2022-06-20 |
333 | ConFUDA: Contrastive Fewshot Unsupervised Domain Adaptation for Medical Image Segmentation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: As a result, in this work, we tackle a more challenging UDA task when there are only a few (fewshot) or a single (oneshot) image available from the target domain. |
MINGXUAN GU et. al. | arxiv-cs.CV | 2022-06-08 |
334 | SpaceEdit: Learning A Unified Editing Space for Open-Domain Image Color Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Recently, large pretrained models (e.g., BERT, StyleGAN, CLIP) show great knowledge transfer and generalization capability on various downstream tasks within their domains. Inspired by these efforts, in this paper we propose a unified model for open-domain image editing focusing on color and tone adjustment of open-domain images while keeping their original content and structure. |
JING SHI et. al. | cvpr | 2022-06-07 |
335 | InOut: Diverse Image Outpainting Via GAN Inversion IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we formulate the problem from the perspective of inverting generative adversarial networks. |
YEN-CHI CHENG et. al. | cvpr | 2022-06-07 |
336 | Polymorphic-GAN: Generating Aligned Samples Across Multiple Domains With Learned Morph Maps Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we introduce a generative adversarial network that can simultaneously generate aligned image samples from multiple related domains. |
Seung Wook Kim; Karsten Kreis; Daiqing Li; Antonio Torralba; Sanja Fidler; | cvpr | 2022-06-07 |
337 | Proactive Image Manipulation Detection IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: By contrast, we propose a proactive scheme to image manipulation detection. |
Vishal Asnani; Xi Yin; Tal Hassner; Sijia Liu; Xiaoming Liu; | cvpr | 2022-06-07 |
338 | StylizedNeRF: Consistent 3D Scene Stylization As Stylized NeRF Via 2D-3D Mutual Learning IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, there is a significant domain gap between style examples which are 2D images and NeRF which is an implicit volumetric representation. To address this problem, we propose a novel mutual learning framework for 3D scene stylization that combines a 2D image stylization network and NeRF to fuse the stylization ability of 2D stylization network with the 3D consistency of NeRF. |
Yi-Hua Huang; Yue He; Yu-Jie Yuan; Yu-Kun Lai; Lin Gao; | cvpr | 2022-06-07 |
339 | Exact Feature Distribution Matching for Arbitrary Style Transfer and Domain Generalization IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we, for the first time to our best knowledge, propose to perform Exact Feature Distribution Matching (EFDM) by exactly matching the empirical Cumulative Distribution Functions (eCDFs) of image features, which could be implemented by applying the Exact Histogram Matching (EHM) in the image feature space. |
Yabin Zhang; Minghan Li; Ruihuang Li; Kui Jia; Lei Zhang; | cvpr | 2022-06-07 |
340 | Style-ERD: Responsive and Coherent Online Motion Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, for online animation applications, such as real-time avatar animation from motion capture, motions need to be processed as a stream with minimal latency. In this work, we realize a flexible, high-quality motion style transfer method for this setting. |
Tianxin Tao; Xiaohang Zhan; Zhongquan Chen; Michiel van de Panne; | cvpr | 2022-06-07 |
341 | Predict, Prevent, and Evaluate: Disentangled Text-Driven Image Manipulation Empowered By Pre-Trained Vision-Language Model IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel framework, i.e., Predict, Prevent, and Evaluate (PPE), for disentangled text-driven image manipulation that requires little manual annotation while being applicable to a wide variety of manipulations. |
ZIPENG XU et. al. | cvpr | 2022-06-07 |
342 | FlexIT: Towards Flexible Semantic Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose FlexIT, a novel method which can take any input image and a user-defined text instruction for editing. |
Guillaume Couairon; Asya Grechka; Jakob Verbeek; Holger Schwenk; Matthieu Cord; | cvpr | 2022-06-07 |
343 | Self-Supervised Dense Consistency Regularization for Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a simple but effective regularization technique for improving GAN-based image-to-image translation. |
MINSU KO et. al. | cvpr | 2022-06-07 |
344 | Drop The GAN: In Defense of Patches Nearest Neighbors As Single Image Generative Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, despite their impressiveness, single-image GANs require long training time (usually hours) for each image and each task and often suffer from visual artifacts. In this paper we revisit the classical patch-based methods, and show that – unlike previously believed — classical methods can be adapted to tackle these novel "GAN-only" tasks. |
Niv Granot; Ben Feinstein; Assaf Shocher; Shai Bagon; Michal Irani; | cvpr | 2022-06-07 |
345 | SketchEdit: Mask-Free Local Image Manipulation With Partial Sketches IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Although this setup simplifies data preparation and model design, it complicates user interaction and discards useful information in masked regions. To this end, we propose a new framework for sketch-based image manipulation that only requires sketch inputs from users and utilizes the entire original image. |
Yu Zeng; Zhe Lin; Vishal M. Patel; | cvpr | 2022-06-07 |
346 | StyTr2: Image Style Transfer With Transformers IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Therefore, traditional neural style transfer methods face biased content representation. To address this critical issue, we take long-range dependencies of input images into account for image style transfer by proposing a transformer-based approach called StyTr^2. |
YINGYING DENG et. al. | cvpr | 2022-06-07 |
347 | PCA-Based Knowledge Distillation Towards Lightweight and Content-Style Balanced Photorealistic Style Transfer Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce PCA-based knowledge distillation to distill lightweight models and show it is motivated by theory. |
Tai-Yin Chiu; Danna Gurari; | cvpr | 2022-06-07 |
348 | Maximum Spatial Perturbation Consistency for Unpaired Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes a universal regularization technique called maximum spatial perturbation consistency (MSPC), which enforces a spatial perturbation function (T) and the translation operator (G) to be commutative (i.e., T \circ G = G \circ T ). |
YANWU XU et. al. | cvpr | 2022-06-07 |
349 | Style-Structure Disentangled Features and Normalizing Flows for Diverse Icon Colorization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this study, we present a colorization network that generates flat-color icons according to given sketches and semantic colorization styles. |
Yuan-kui Li; Yun-Hsuan Lien; Yu-Shuen Wang; | cvpr | 2022-06-07 |
350 | Unsupervised Image-to-Image Translation With Generative Prior IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we present a novel framework, Generative Prior-guided UNsupervised Image-to-image Translation (GP-UNIT), to improve the overall quality and applicability of the translation algorithm. |
Shuai Yang; Liming Jiang; Ziwei Liu; Chen Change Loy; | cvpr | 2022-06-07 |
351 | Industrial Style Transfer With Large-Scale Geometric Warping and Content Preservation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel style transfer method to quickly create a new visual product with a nice appearance for industrial designers’ reference. |
Jinchao Yang; Fei Guo; Shuo Chen; Jun Li; Jian Yang; | cvpr | 2022-06-07 |
352 | Exploring Patch-Wise Semantic Relation for Contrastive Learning in Image-to-Image Translation Tasks IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, the methods often ignore the diverse semantic relation within the images. To address this, here we propose a novel semantic relation consistency (SRC) regularization along with the decoupled contrastive learning (DCL), which utilize the diverse semantics by focusing on the heterogeneous semantics between the image patches of a single image. |
Chanyong Jung; Gihyun Kwon; Jong Chul Ye; | cvpr | 2022-06-07 |
353 | 3D Photo Stylization: Learning To Generate Stylized Novel Views From A Single Image IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we make a connection between the two, and address the challenging task of 3D photo stylization – generating stylized novel views from a single image given an arbitrary style. |
Fangzhou Mu; Jian Wang; Yicheng Wu; Yin Li; | cvpr | 2022-06-07 |
354 | ManiTrans: Entity-Level Text-Guided Image Manipulation Via Token-Wise Semantic Alignment and Generation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we study a novel task on text-guided image manipulation on the entity level in the real world. |
JIANAN WANG et. al. | cvpr | 2022-06-07 |
355 | StyleMesh: Style Transfer for Indoor 3D Scene Reconstructions Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We apply style transfer on mesh reconstructions of indoor scenes. |
Lukas Höllein; Justin Johnson; Matthias Nießner; | cvpr | 2022-06-07 |
356 | Artistic Style Discovery With Independent Components Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In particular, we take a closer look into the mechanism of style transfer and obtain different artistic style components from the latent space consisting of different style features. |
XIN XIE et. al. | cvpr | 2022-06-07 |
357 | Few-Shot Incremental Learning for Label-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce a few-shot incremental learning method for label-to-image translation. |
Pei Chen; Yangkang Zhang; Zejian Li; Lingyun Sun; | cvpr | 2022-06-07 |
358 | InstaFormer: Instance-Aware Image-to-Image Translation With Transformer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present a novel Transformer-based network architecture for instance-aware image-to-image translation, dubbed InstaFormer, to effectively integrate global- and instance-level information. |
Soohyun Kim; Jongbeom Baek; Jihye Park; Gyeongnyeon Kim; Seungryong Kim; | cvpr | 2022-06-07 |
359 | Pastiche Master: Exemplar-Based High-Resolution Portrait Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we explore more challenging exemplar-based high-resolution portrait style transfer by introducing a novel DualStyleGAN with flexible control of dual styles of the original face domain and the extended artistic portrait domain. |
Shuai Yang; Liming Jiang; Ziwei Liu; Chen Change Loy; | cvpr | 2022-06-07 |
360 | A Style-Aware Discriminator for Controllable Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This limitation largely arises because labels do not consider the semantic distance. To mitigate such problems, we propose a style-aware discriminator that acts as a critic as well as a style encoder to provide conditions. |
Kunhee Kim; Sanghun Park; Eunyeong Jeon; Taehun Kim; Daijin Kim; | cvpr | 2022-06-07 |
361 | CLIPstyler: Image Style Transfer With A Single Text Condition IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, in many practical situations, users may not have reference style images but still be interested in transferring styles by just imagining them. In order to deal with such applications, we propose a new framework that enables a style transfer ‘without’ a style image, but only with a text description of the desired style. |
Gihyun Kwon; Jong Chul Ye; | cvpr | 2022-06-07 |
362 | Style Transformer for Image Inversion and Editing IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper presents a transformer-based image inversion and editing model for pretrained StyleGAN which is not only with less distortions, but also of high quality and flexibility for editing. |
XUEQI HU et. al. | cvpr | 2022-06-07 |
363 | Artistic Style Novel View Synthesis Based on A Single Image Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Recent progress in 3D display technologies has raised the demand for stylized 3D digital content. Previous approaches either perform style transfer on stereoscopic image pairs or … |
Kuan-Wei Tseng; Yao Lee; Chu-Song Chen; | 2022 IEEE/CVF Conference on Computer Vision and Pattern … | 2022-06-01 |
364 | Variational Transfer Learning Using Cross-Domain Latent Modulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose to introduce a novel cross-domain latent modulation mechanism to a variational autoencoder framework so as to achieve effective transfer learning. |
Jinyong Hou; Jeremiah D. Deng; Stephen Cranefield; Xuejie Din; | arxiv-cs.LG | 2022-05-30 |
365 | Multiscale Voxel Based Decoding For Enhanced Natural Image Reconstruction From Brain Activity Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we present a novel approach for enhanced image reconstruction, in which existing methods for object decoding and image reconstruction are merged together. |
Mali Halac; Murat Isik; Hasan Ayaz; Anup Das; | arxiv-cs.CV | 2022-05-27 |
366 | Pretraining Is All You Need for Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we regard each image-to-image translation problem as a downstream task and introduce a simple and generic framework that adapts a pretrained diffusion model to accommodate various kinds of image-to-image translation. |
TENGFEI WANG et. al. | arxiv-cs.CV | 2022-05-25 |
367 | Region-aware Knowledge Distillation for Efficient Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose Region-aware Knowledge Distillation ReKo to compress image-to-image translation models. |
Linfeng Zhang; Xin Chen; Runpei Dong; Kaisheng Ma; | arxiv-cs.CV | 2022-05-24 |
368 | Improving The Latent Space of Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In addition, the lack of content details in the features extracted by the pre-trained encoder also leads to the content leak problem. In order to solve these issues in the latent space used by style transfer, we propose two contrastive training schemes to get a refined encoder that is more suitable for this task. |
Yunpeng Bai; Cairong Wang; Chun Yuan; Yanbo Fan; Jue Wang; | arxiv-cs.CV | 2022-05-24 |
369 | Mind The Gap: Alleviating Local Imbalance for Unsupervised Cross-Modality Medical Image Segmentation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: While this operation may cause a deficiency of critical information from contexts. To tackle this limitation, we propose a novel strategy to alleviate the domain gap imbalance considering the characteristics of medical images, namely Global-Local Union Alignment. |
ZIXIAN SU et. al. | arxiv-cs.CV | 2022-05-24 |
370 | Improving Human Image Synthesis with Residual Fast Fourier Transformation and Wasserstein Distance Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: And the second is that the training of GAN is unstable and slow to converge, such as model collapse. Based on the above two problems, we propose several methods to solve them. |
Jianhan Wu; Shijing Si; Jianzong Wang; Jing Xiao; | arxiv-cs.CV | 2022-05-24 |
371 | Multi-Domain Unsupervised Image-to-Image Translation with Appearance Adaptive Convolution Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Despite the impressive results, they mainly focus on the I2I translation between two domains, so the multi-domain I2I translation still remains a challenge. To address this problem, we propose a novel multi-domain unsupervised image-to-image translation (MDUIT) framework that leverages the decomposed content feature and appearance adaptive convolution to translate an image into a target appearance while preserving the given geometric content. |
S. Jeong; J. Lee; K. Sohn; | icassp | 2022-05-22 |
372 | Generative Adversarial Network Including Referring Image Segmentation For Text-Guided Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes a novel generative adversarial network to improve the performance of image manipulation using natural language descriptions that contain desired attributes. |
Y. Watanabe; R. Togo; K. Maeda; T. Ogawa; M. Haseyama; | icassp | 2022-05-22 |
373 | Wavelet-Based Unsupervised Label-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: State-of-the-art conditional Generative Adversarial Networks (GANs) need a huge amount of paired data to accomplish this task while generic un-paired image-to-image translation frameworks underperform in comparison, because they color-code semantic layouts and learn correspondences in appearance instead of semantic content. Starting from the assumption that a high quality generated image should be segmented back to its semantic layout, we propose a new Unsupervised paradigm for SIS (USIS) that makes use of a self-supervised segmentation loss and whole image wavelet based discrimination. |
G. Eskandar; M. Abdelsamad; K. Armanious; S. Zhang; B. Yang; | icassp | 2022-05-22 |
374 | Variational Bayesian Framework for Advanced Image Generation with Domain-Related Variables Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, it remains challenging for existing methods to address advanced conditional generative problems without annotations, which can enable multiple applications like image-to-image translation and image editing. We present a unified Bayesian framework for such problems, which introduces an inference stage on latent variables within the learning process. |
Y. Li; S. Mazuelas; Y. Shen; | icassp | 2022-05-22 |
375 | Towards Using Clothes Style Transfer for Scenario-Aware Person Video Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To further improve the generation performance, we propose a novel framework with disentangled multi-branch encoders and a shared decoder. |
J. Xu; et al. | icassp | 2022-05-22 |
376 | Parameter-Free Style Projection for Arbitrary Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Existing feature transformation algorithms often suffer from loss of content or style details, non-natural stroke patterns, and unstable training. To mitigate these issues, this paper proposes a new feature-level style transformation technique, named Style Projection, for parameter-free, fast, and effective content-style transformation. |
S. Huang; et al. | icassp | 2022-05-22 |
377 | AutoLink: Self-supervised Learning of Human Skeletons and Object Outlines By Linking Keypoints Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a self-supervised method that learns to disentangle object structure from the appearance with a graph of 2D keypoints linked by straight edges. |
Xingzhe He; Bastian Wandt; Helge Rhodin; | arxiv-cs.CV | 2022-05-21 |
378 | Exploiting Social Media Content for Self-Supervised Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper we present a novel Self-Supervised Style Transfer (3ST) model, which augments SSNMT with UNMT methods in order to identify and efficiently exploit supervisory signals in non-parallel social media posts. |
Dana Ruiter; Thomas Kleinbauer; Cristina España-Bonet; Josef van Genabith; Dietrich Klakow; | arxiv-cs.CL | 2022-05-18 |
379 | Multilingual Pre-training with Language and Task Adaptation for Multilingual Text Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Besides, in view of the general scarcity of parallel data, we propose a modular approach for multilingual formality transfer, which consists of two training strategies that target adaptation to both language and task. |
Huiyuan Lai; Antonio Toral; Malvina Nissim; | acl | 2022-05-17 |
380 | Few-shot Controllable Style Transfer for Low-Resource Multilingual Settings IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work we study a relevant low-resource setting: style transfer for languages where no style-labelled corpora are available. |
Kalpesh Krishna; Deepak Nathani; Xavier Garcia; Bidisha Samanta; Partha Talukdar; | acl | 2022-05-17 |
381 | A Recipe for Arbitrary Text Style Transfer with Large Language Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we leverage large language models (LLMs) to perform zero-shot text style transfer. |
EMILY REIF et. al. | acl | 2022-05-17 |
382 | Paired Image-to-Image Translation Quality Assessment Using Multi-Method Fusion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes a novel approach that combines signals of image quality between paired source and transformation to predict the latter’s similarity with a hypothetical ground truth. |
Stefan Borasinski; Esin Yavuz; Sébastien Béhuret; | arxiv-cs.CV | 2022-05-09 |
383 | WeatherGAN: Unsupervised Multi-weather Image-to-image Translation Via Single Content-preserving UResNet Generator Related Papers Related Patents Related Grants Related Venues Related Experts View |
Sunhee Hwang; Seogkyu Jeon; Yuhui Ma; H. Byun; | Multimedia Tools and Applications | 2022-05-07 |
384 | Text to Artistic Image Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Painting is one of the ways for people to express their ideas, but what if people with disabilities in hands want to paint? To tackle this challenge, we create an end-to-end solution that can generate artistic images from text descriptions. |
Qinghe Tian; Jean-Claude Franchitti; | arxiv-cs.CV | 2022-05-05 |
385 | Generate and Edit Your Own Character in A Canonical View Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel and unified framework which generates stylized portraits in canonical view. |
Jeong-gi Kwak; Yuanming Li; Dongsik Yoon; David Han; Hanseok Ko; | arxiv-cs.CV | 2022-05-05 |
386 | Hypercomplex Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose to leverage hypercomplex algebra properties to define lightweight I2I generative models capable of preserving pre-existing relations among image dimensions, thus exploiting additional input information. |
Eleonora Grassucci; Luigi Sigillo; Aurelio Uncini; Danilo Comminiello; | arxiv-cs.CV | 2022-05-04 |
387 | Towards Robust and Semantically Organised Latent Representations for Unsupervised Text Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This definition of similarity does not look into the underlying nuances of the constituent words while mapping latent space neighbourhoods and therefore fails to recognise sentences with different style-based semantics while mapping latent neighbourhoods. We introduce EPAAEs (Embedding Perturbed Adversarial AutoEncoders) which completes this perturbation model, by adding a finely adjustable noise component on the continuous embeddings space. |
Sharan Narasimhan; Suvodip Dey; Maunendra Sankar Desarkar; | arxiv-cs.CL | 2022-05-04 |
388 | Hypercomplex Image- To- Image Translation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Image-to-image translation (I2I) aims at transferring the content representation from an input domain to an output one, bouncing along different target domains. Recent I2I … |
Eleonora Grassucci; Luigi Sigillo; A. Uncini; D. Comminiello; | 2022 International Joint Conference on Neural Networks … | 2022-05-04 |
389 | Diverse Image Captioning with Grounded Style Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: Stylized image captioning as presented in prior work aims to generate captions that reflect characteristics beyond a factual description of the scene composition, such as … |
Franz Klein; Shweta Mahajan; Stefan Roth; | arxiv-cs.CV | 2022-05-03 |
390 | Segmentation Mask and Feature Similarity Loss Guided GAN for Object-oriented Image-to-image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View |
ZHEN QIN et. al. | Inf. Process. Manag. | 2022-05-01 |
391 | An Overview of Color Transfer and Style Transfer for Images and Videos Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Given a source image (video) and a target image (video), the image (video) color transfer technique aims to process the color of the source image or video (note that the source image or video is also referred to the reference image or video in some literature) to make it look like that of the target image or video, i.e., transferring the appearance of the target image or video to that of the source image or video, which can thereby change one’s perception of the source image or video. |
Shiguang Liu; | arxiv-cs.CV | 2022-04-28 |
392 | ProCST: Boosting Semantic Segmentation Using Progressive Cyclic Style-Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a novel two-stage framework for improving domain adaptation techniques on image data. |
Shahaf Ettedgui; Shady Abu-Hussein; Raja Giryes; | arxiv-cs.CV | 2022-04-25 |
393 | Exploring Negatives in Contrastive Learning for Unpaired Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Unlike previous CL approaches that use negatives as much as possible, in this paper, we study the negatives from an information-theoretic perspective and introduce a new negative Pruning technology for Unpaired image-to-image Translation (PUT) by sparsifying and ranking the patches. |
YUPEI LIN et. al. | arxiv-cs.CV | 2022-04-23 |
394 | SE-GAN: Skeleton Enhanced GAN-based Model for Brush Handwriting Font Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: There is rare research on brush handwriting font generation, which involves holistic structure changes and complex strokes transfer. To address this issue, we propose a novel GAN-based image translation model by integrating the skeleton information. |
SHAOZU YUAN et. al. | arxiv-cs.CV | 2022-04-21 |
395 | ChildPredictor: A Child Face Prediction Framework with Disentangled Learning Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In order to obtain accurate genetic factors and perform the mapping, we propose a ChildPredictor framework. |
YUZHI ZHAO et. al. | arxiv-cs.CV | 2022-04-21 |
396 | Heavy Rain Face Image Restoration: Integrating Physical Degradation Model and Facial Component Guided Adversarial Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Unlike traditional image degradation models (IDM), such as rain removal and superresolution, this study addresses a new IDM referred to as a scale-aware heavy rain model and proposes a method for restoring high-resolution face images (HR-FIs) from low-resolution heavy rain face images (LRHR-FI). |
Chang-Hwan Son; Da-Hee Jeong; | arxiv-cs.CV | 2022-04-18 |
397 | Non-Parallel Text Style Transfer with Self-Parallel Supervision Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose LaMer, a novel text style transfer framework based on large-scale language models. |
Ruibo Liu; Chongyang Gao; Chenyan Jia; Guangxuan Xu; Soroush Vosoughi; | arxiv-cs.CL | 2022-04-17 |
398 | The Swiss Army Knife for Image-to-Image Translation: Multi-Task Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We build on a method for image-to-image translation using denoising diffusion implicit models and include a regression problem and a segmentation problem for guiding the image generation to the desired output. |
Julia Wolleb; Robin Sandkühler; Florentin Bieder; Philippe C. Cattin; | arxiv-cs.CV | 2022-04-06 |
399 | Flexible Portrait Image Editing with Fine-Grained Control Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We develop a new method for portrait image editing, which supports fine-grained editing of geometries, colors, lights and shadows using a single neural network model. |
Linlin Liu; Qian Fu; Fei Hou; Ying He; | arxiv-cs.CV | 2022-04-04 |
400 | Direct Dense Pose Estimation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We therefore propose a novel alternative method for solving the dense pose estimation problem, called Direct Dense Pose (DDP). |
Liqian Ma; Lingjie Liu; Christian Theobalt; Luc Van Gool; | arxiv-cs.CV | 2022-04-04 |
401 | Unsupervised Coherent Video Cartoonization with Perceptual Motion Consistency Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a spatially-adaptive semantic alignment framework with perceptual motion consistency for coherent video cartoonization in an unsupervised manner. |
ZHENHUAN LIU et. al. | arxiv-cs.CV | 2022-04-02 |
402 | IR-GAN: Image Manipulation with Linguistic Instruction By Increment Reasoning IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, traditional conditional image generation models mainly focus on generating high-quality and visually realistic images, and lack resolving the partial consistency between image and instruction. To address this issue, we propose an Increment Reasoning Generative Adversarial Network (IR-GAN), which aims to reason the consistency between visual increment in images and semantic increment in instructions. |
ZHENHUAN LIU et. al. | arxiv-cs.CV | 2022-04-02 |
403 | Treatment Learning Transformer for Noisy Image Classification Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we incorporate this binary information of existence of noise as treatment into image classification tasks to improve prediction accuracy by jointly estimating their treatment effects. |
Chao-Han Huck Yang; I-Te Danny Hung; Yi-Chieh Liu; Pin-Yu Chen; | arxiv-cs.CV | 2022-03-29 |
404 | StyleFool: Fooling Video Classification Systems Via Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we focus on unrestricted perturbations and propose StyleFool, a black-box video adversarial attack via style transfer to fool the video classification system. |
YUXIN CAO et. al. | arxiv-cs.CV | 2022-03-29 |
405 | ITTR: Unpaired Image-to-Image Translation with Transformers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose an effective and efficient architecture for unpaired Image-to-Image Translation with Transformers (ITTR). |
Wanfeng Zheng; Qiang Li; Guoxin Zhang; Pengfei Wan; Zhongyuan Wang; | arxiv-cs.CV | 2022-03-29 |
406 | Semi-Supervised Image-to-Image Translation Using Latent Space Mapping Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We therefore introduce a general framework for semi-supervised image translation. |
Pan Zhang; Jianmin Bao; Ting Zhang; Dong Chen; Fang Wen; | arxiv-cs.CV | 2022-03-29 |
407 | Adversarial Contrastive Fourier Domain Adaptation for Polyp Segmentation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Due to the shortage of experienced endoscopists, Computer-Aided Diagnosis (CAD) systems for colonoscopy have recently attracted many research interests. There exist several public … |
TA DUC HUY et. al. | 2022 IEEE 19th International Symposium on Biomedical … | 2022-03-28 |
408 | Interactive Style Transfer: All Is Your Palette Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To mitigate this limitation, we propose a drawing-like interactive style transfer (IST) method, by which users can interactively create a harmonious-style image. |
Zheng Lin; Zhao Zhang; Kang-Rui Zhang; Bo Ren; Ming-Ming Cheng; | arxiv-cs.CV | 2022-03-25 |
409 | Playing Lottery Tickets in Style Transfer Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Recently, the lottery ticket hypothesis (LTH) has shown great potential in finding extremely sparse matching subnetworks which can achieve on par or even better performance than the original full networks when trained in isolation. In this work, we for the first time perform an empirical study to verify whether such trainable matching subnetworks also exist in style transfer models. |
MEIHAO KONG et. al. | arxiv-cs.CV | 2022-03-25 |
410 | 3D GAN Inversion for Controllable Portrait Image Animation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we leverage newly developed 3D GANs, which allow explicit control over the pose of the image subject with multi-view consistency. |
Connor Z. Lin; David B. Lindell; Eric R. Chan; Gordon Wetzstein; | arxiv-cs.CV | 2022-03-25 |
411 | Neural Neighbor Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose Neural Neighbor Style Transfer (NNST), a pipeline that offers state-of-the-art quality, generalization, and competitive efficiency for artistic style transfer. |
NICHOLAS KOLKIN et. al. | arxiv-cs.CV | 2022-03-24 |
412 | Beyond A Video Frame Interpolator: A Space Decoupled Learning Approach to Continuous Image Transition Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we rethink the VFI problem and formulate it as a continuous image transition (CIT) task, whose key issue is to transition an image from one space to another space continuously. |
Tao Yang; Peiran Ren; Xuansong Xie; Xiansheng Hua; Lei Zhang; | arxiv-cs.CV | 2022-03-18 |
413 | Deepfake Style Transfer Mixture: A First Forensic Ballistics Study on Synthetic Images Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To this aim, in this paper a first approach to investigate the image ballistics on deepfake images subject to style-transfer manipulations is proposed. |
Luca Guarnera; Oliver Giudice; Sebastiano Battiato; | arxiv-cs.CV | 2022-03-18 |
414 | Depth-aware Neural Style Transfer Using Instance Normalization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our approach uses a deep residual convolutional network with instance normalization layers that utilizes an advanced depth prediction network to integrate depth preservation as an additional loss function to content and style. |
Eleftherios Ioannou; Steve Maddock; | arxiv-cs.CV | 2022-03-17 |
415 | CtlGAN: Few-shot Artistic Portraits Generation with Contrastive Transfer Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose CtlGAN, a new few-shot artistic portraits generation model with a novel contrastive transfer learning strategy. |
Yue Wang; Ran Yi; Ying Tai; Chengjie Wang; Lizhuang Ma; | arxiv-cs.CV | 2022-03-16 |
416 | Dual Diffusion Implicit Bridges for Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present Dual Diffusion Implicit Bridges (DDIBs), an image translation method based on diffusion models, that circumvents training on domain pairs. |
Xuan Su; Jiaming Song; Chenlin Meng; Stefano Ermon; | arxiv-cs.CV | 2022-03-16 |
417 | Image Style Transfer: from Artistic to Photorealistic Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this review, we reviewed the development of photorealistic style transfer starting from artistic style transfer and the contribution of traditional image processing techniques on photorealistic style transfer, including some work that had been completed in the Multimedia lab at the University of Alberta. |
Chenggui Sun; Li Bin Song; | arxiv-cs.CV | 2022-03-11 |
418 | Membership Privacy Protection for Image Translation Models Via Adversarial Knowledge Distillation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose adversarial knowledge distillation (AKD) as a defense method against MIAs for image-to-image translation models. |
Saeed Ranjbar Alvar; Lanjun Wang; Jian Pei; Yong Zhang; | arxiv-cs.CV | 2022-03-10 |
419 | Image Steganography Based on Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To improve upon this, we propose image steganography network based on style transfer, and the embedding of secret messages can be disguised as image stylization. |
Donghui Hu; Yu Zhang; Cong Yu; Jian Wang; Yaofei Wang; | arxiv-cs.CV | 2022-03-08 |
420 | Multi-granularity Brushstrokes Network for Universal Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Neural style transfer has been developed in recent years, where both performance and efficiency have been greatly improved. However, most existing methods do not transfer the … |
Quan Wang; Sheng Li; Xinpeng Zhang; Guorui Feng; | ACM Transactions on Multimedia Computing, Communications, … | 2022-03-04 |
421 | Sim2Real Instance-Level Style Transfer for 6D Pose Estimation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To solve this problem, we introduce a simulation to reality (sim2real) instance-level style transfer for 6D pose estimation network training. |
TAKUYA IKEDA et. al. | arxiv-cs.CV | 2022-03-03 |
422 | PetsGAN: Rethinking Priors for Single Image Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: The main contributions of this paper include: 1) We introduce to SIG a regularized latent variable model. |
ZICHENG ZHANG et. al. | arxiv-cs.CV | 2022-03-02 |
423 | GRA-GAN: Generative Adversarial Network for Image Style Transfer of Gender, Race, and Age IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
Y. Kim; S. Nam; Seung Baek Hong; K. Park; | Expert Syst. Appl. | 2022-03-01 |
424 | Styleverse: Towards Identity Stylization Across Heterogeneous Domains Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a new challenging task namely IDentity Stylization (IDS) across heterogeneous domains. |
Jia Li; Jie Cao; JunXian Duan; Ran He; | arxiv-cs.CV | 2022-03-01 |
425 | Name Your Style: An Arbitrary Artist-aware Image Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a text-driven image style transfer (TxST) that leverages advanced image-text encoders to control arbitrary style transfer. |
Zhi-Song Liu; Li-Wen Wang; Wan-Chi Siu; Vicky Kalogeiton; | arxiv-cs.CV | 2022-02-28 |
426 | Multi-scale Attention Guided Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present an improved network architecture for pose transfer by introducing attention links at every resolution level of the encoder and decoder. |
Prasun Roy; Saumik Bhattacharya; Subhankar Ghosh; Umapada Pal; | arxiv-cs.CV | 2022-02-14 |
427 | Motion Puzzle: Arbitrary Motion Style Transfer By Body Part IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper presents Motion Puzzle, a novel motion style transfer network that advances the state-of-the-art in several important respects. |
Deok-Kyeong Jang; Soomin Park; Sung-Hee Lee; | arxiv-cs.GR | 2022-02-10 |
428 | Deep Feature Rotation for Multimodal Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a simple method for representing style features in many ways called Deep Feature Rotation (DFR), while not only producing diverse outputs but also still achieving effective stylization compared to more complex methods. |
Son Truong Nguyen; Nguyen Quang Tuyen; Nguyen Hong Phuc; | arxiv-cs.CV | 2022-02-09 |
429 | Distilling GANs with Style-Mixed Triplets for X2I Translation with Limited Data Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: One transfer learning method generalizes varying kinds of conditional image synthesization tasks. |
Yaxing Wang; Joost van de weijer; Lu Yu; SHANGLING JUI; | iclr | 2022-02-08 |
430 | Residual Aligned: Gradient Optimization for Non-Negative Image Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we address an important problem of optical see through (OST) augmented reality: non-negative image synthesis. |
Flora Yu Shen; Katie Luo; Guandao Yang; Harald Haraldsson; Serge Belongie; | arxiv-cs.CV | 2022-02-08 |
431 | Deep Translation Prior: Test-Time Training for Photorealistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Recent techniques to solve photorealistic style transfer within deep convolutional neural networks (CNNs) generally require intensive training from large-scale datasets, thus having limited applicability and poor generalization ability to unseen images or styles. To overcome this, we propose a novel framework, dubbed Deep Translation Prior (DTP), to accomplish photorealistic style transfer through test-time training on given input image pair with untrained networks, which learns an image pair-specific translation prior and thus yields better performance and generalization. |
Sunwoo Kim; Soohyun Kim; Seungryong Kim; | aaai | 2022-02-07 |
432 | TiGAN: Text-Based Interactive Image Generation and Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel framework for Text-based Interactive image generation and manipulation (TiGAN) that responds to users’ natural-language feedback. |
YUFAN ZHOU et. al. | aaai | 2022-02-07 |
433 | Style Mixing and Patchwise Prototypical Matching for One-Shot Unsupervised Domain Adaptive Semantic Segmentation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we tackle the problem of one-shot unsupervised domain adaptation (OSUDA) for semantic segmentation where the segmentors only see one unlabeled target image during training. |
Xinyi Wu; Zhenyao Wu; Yuhang Lu; Lili Ju; Song Wang; | aaai | 2022-02-07 |
434 | MOST-GAN: 3D Morphable StyleGAN for Disentangled Face Image Manipulation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In contrast, we propose a framework that a priori models physical attributes of the face such as 3D shape, albedo, pose, and lighting explicitly, thus providing disentanglement by design. |
SAFA C. MEDIN et. al. | aaai | 2022-02-07 |
435 | MOST-Net: A Memory Oriented Style Transfer Network for Face Sketch Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Current image-to-image translation-based face sketch synthesis frequently encounters over-fitting problems when it comes to small-scale datasets. To tackle this problem, we present an end-to-end Memory Oriented Style Transfer Network (MOST-Net) for face sketch synthesis which can produce high-fidelity sketches with limited data. |
Fan Ji; Muyi Sun; Xingqun Qi; Qi Li; Zhenan Sun; | arxiv-cs.CV | 2022-02-07 |
436 | Style-Guided and Disentangled Representation for Robust Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Based on two ideas, this paper proposes Style-Guided and Disentangled Representation for Robust Image-to-Image Translation (SRIT). |
Jaewoong Choi; Daeha Kim; Byung Cheol Song; | aaai | 2022-02-07 |
437 | Multi-domain Unsupervised Image-to-Image Translation with Appearance Adaptive Convolution Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To address this problem, we propose a novel multi-domain unsupervised image-to-image translation (MDUIT) framework that leverages the decomposed content feature and appearance adaptive convolution to translate an image into a target appearance while preserving the given geometric content. |
Somi Jeong; Jiyoung Lee; Kwanghoon Sohn; | arxiv-cs.CV | 2022-02-06 |
438 | Stay Positive: Non-Negative Image Synthesis for Augmented Reality Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Our key insight is that one can leverage this behavior to produce high quality images with negligible artifacts. |
KATIE LUO et. al. | arxiv-cs.CV | 2022-02-01 |
439 | An Improved Image Steganography Framework Based on Y Channel Information for Neural Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Neural style transfer has effectively assisted artistic design in recent years, but it has also accelerated the tampering, synthesis, and dissemination of a large number of … |
WEN-BIN LIN et. al. | Secur. Commun. Networks | 2022-01-29 |
440 | Paired Image to Image Translation for Strikethrough Removal From Handwritten Words Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper investigates the use of paired image to image translation approaches to remove strikethrough strokes from handwritten words. |
Raphaela Heil; Ekta Vats; Anders Hast; | arxiv-cs.CV | 2022-01-24 |
441 | Disentangling Style and Speaker Attributes for TTS Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a new approach to seen and unseen style transfer training on disjoint, multi-style datasets, i.e., datasets of different styles are recorded, one individual style by one speaker in multiple utterances. |
Xiaochun An; Frank K. Soong; Lei Xie; | arxiv-cs.SD | 2022-01-24 |
442 | Generative Adversarial Network Applications in Creating A Meta-Universe Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we discuss how GANs can be used to create an artificial world. |
Soheyla Amirian; Thiab R. Taha; Khaled Rasheed; Hamid R. Arabnia; | arxiv-cs.CV | 2022-01-22 |
443 | Can Machines Generate Personalized Music? A Hybrid Favorite-aware Method for User Preference Music Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: User preference music transfer (UPMT) is a new problem in music style transfer that can be applied to many scenarios but remains understudied. … |
Zhejing Hu; Yan Liu; Gong Chen; Yongxu Liu; | arxiv-cs.SD | 2022-01-20 |
444 | Extending The Vocabulary of Fictional Languages Using Neural Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a deep learning solution to the problem. |
Thomas Zacharias; Ashutosh Taklikar; Raja Giryes; | arxiv-cs.CL | 2022-01-18 |
445 | Arbitrary Handwriting Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposed a method to imitate handwriting style by style transfer. |
Kai Yang; Xiaoman Liang; Huihuang Zhao; | arxiv-cs.CV | 2022-01-14 |
446 | Deep Convolutional Nets Learning Classification for Artistic Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Humans have mastered the skill of creativity for many decades. The process of replicating this mechanism is introduced recently by using neural networks which replicate the … |
R. D. KUMAR et. al. | Sci. Program. | 2022-01-10 |
447 | GMFIM: A Generative Mask-guided Facial Image Manipulation Model for Privacy Preservation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a Generative Mask-guided Face Image Manipulation (GMFIM) model based on GANs to apply imperceptible editing to the input face image to preserve the privacy of the person in the image. |
Mohammad Hossein Khojaste; Nastaran Moradzadeh Farid; Ahmad Nickabadi; | arxiv-cs.CV | 2022-01-10 |
448 | Consistent Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Recently, attentional arbitrary style transfer methods have been proposed to achieve fine-grained results, which manipulates the point-wise similarity between content and style features for stylization. |
Xuan Luo; Zhen Han; Lingkang Yang; Lingling Zhang; | arxiv-cs.CV | 2022-01-06 |
449 | Probing TryOnGAN Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this article, we reproduce the TryOnGAN implementation and probe it along diverse angles: impact of transfer learning, variants of conditioning image generation with poses and properties of latent space interpolation. |
Saurabh Kumar; Nishant Sinha; | arxiv-cs.CV | 2022-01-05 |
450 | A Comparative Analysis of GAN-Based Methods for SAR-to-Optical Image Translation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Unlike optical sensors, synthetic aperture radar (SAR) sensors acquire images of the Earth’s surface with all-weather and all-time capabilities, which is vital in a situation such … |
Yitao Zhao; T. Çelik; Nanqing Liu; Hengchao Li; | IEEE Geoscience and Remote Sensing Letters | 2022-01-01 |
451 | Disentangling Noise Patterns From Seismic Images: Noise Reduction and Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Seismic interpretation is a fundamental approach for obtaining static and dynamic information about subsurface reservoirs, such as geological faults/salt bodies and associated … |
HAIWEN DU et. al. | IEEE Transactions on Geoscience and Remote Sensing | 2022-01-01 |
452 | Memory-Modulated Transformer Network for Heterogeneous Face Recognition Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Heterogeneous face recognition (HFR) aims at matching face images across different domains. It is challenging due to the severe domain discrepancies and overfitting caused by … |
Mandi Luo; Haoxue Wu; Huaibo Huang; Weizan He; Ran He; | IEEE Transactions on Information Forensics and Security | 2022-01-01 |
453 | Inversion-Based Creativity Transfer with Diffusion Models IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In this paper, we introduce the task of “Creativity Trans-fer”. The artistic creativity within a painting is the means of expression, which includes not only the painting … |
YU-XIN ZHANG et. al. | ArXiv | 2022-01-01 |
454 | MIST-Tacotron: End-to-End Emotional Speech Synthesis Using Mel-Spectrogram Image Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: With the development of voice synthesis technology using deep learning, voice synthesis research that expresses the characteristics and emotions of speakers is actively being … |
Sung-Woo Moon; Sunghyun Kim; Yong-Hoon Choi; | IEEE Access | 2022-01-01 |
455 | BiFDANet: Unsupervised Bidirectional Domain Adaptation for Semantic Segmentation of Remote Sensing Images IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: When segmenting massive amounts of remote sensing images collected from different satellites or geographic locations (cities), the pre-trained deep learning models cannot always … |
YUXIANG CAI et. al. | Remote. Sens. | 2022-01-01 |
456 | Loss Functions for Pose Guided Person Image Generation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Pose guided person image generation aims to transform a source person image to a target pose. It is an ill-posed problem as we often need to generate pixels that are invisible in … |
Haoyue Shi; Le Wang; Nanning Zheng; Gang Hua; Wei Tang; | Pattern Recognition | 2022-01-01 |
457 | A Unified Framework for Bidirectional Prototype Learning From Contaminated Faces Across Heterogeneous Domains Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Existing heterogeneous face synthesis (HFS) methods focus on performing accurate image-to-image translation across domains, while they cannot effectively remove the nuisance … |
Meng Pang; Binghui Wang; Siyu Huang; Y. Cheung; B. Wen; | IEEE Transactions on Information Forensics and Security | 2022-01-01 |
458 | MSRD-CNN: Multi-Scale Residual Deep CNN for General-Purpose Image Manipulation Detection Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The authenticity of digital images is a major concern in multimedia forensics due to the availability of advanced photo editing tools/devices. In the literature, several image … |
Kapil Rana; Gurinder Singh; P. Goyal; | IEEE Access | 2022-01-01 |
459 | Intelligent Matching Method for Heterogeneous Remote Sensing Images Based on Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Intelligent matching of heterogeneous remote sensing images is a common basic problem in the field of intelligent remote sensing image processing. Aiming at the difficulty of … |
Jiawei Zhao; Dongfang Yang; Yongfei Li; Peng Xiao; Jinglan Yang; | IEEE Journal of Selected Topics in Applied Earth … | 2022-01-01 |
460 | SC-UDA: Style and Content Gaps Aware Unsupervised Domain Adaptation for Object Detection IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Current state-of-the-art object detectors can have significant performance drop when deployed in the wild due to domain gaps with training data. Unsupervised Domain Adaptation … |
FUXUN YU et. al. | 2022 IEEE/CVF Winter Conference on Applications of Computer … | 2022-01-01 |
461 | VQBB: Image-to-image Translation with Vector Quantized Brownian Bridge Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Image-to-image translation is an important and chal-lenging problem in computer vision. Existing approaches like Pixel2Pixel [7], DualGAN [12] suffer from the instabil-ity of GAN … |
Bo Li; Kaitao Xue; Bin Liu; Yunyu Lai; | ArXiv | 2022-01-01 |
462 | Neural Style Transfer and Unpaired Image-to-Image Translation to Deal with The Domain Shift Problem on Spheroid Segmentation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This is common in the context of biomedical image segmentation due to the variance of experimental conditions, equipment, and capturing settings. In this work, we address this challenge by studying both neural style transfer algorithms and unpaired image-to-image translation methods in the context of the segmentation of tumour spheroids. |
Manuel García-Domínguez; César Domínguez; Jónathan Heras; Eloy Mata; Vico Pascual; | arxiv-cs.CV | 2021-12-16 |
463 | How to Augment Your ViTs? Consistency Loss and StyleAug, A Random Style Transfer Augmentation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we empirical evaluated how different data augmentation strategies performed on CNN (e.g., ResNet) versus ViT architectures for image classification. |
Akash Umakantha; Joao D. Semedo; S. Alireza Golestaneh; Wan-Yi S. Lin; | arxiv-cs.CV | 2021-12-16 |
464 | Explainable Thermal to Visible Face Recognition Using Latent-Guided Generative Adversarial Network Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: One of the main challenges in performing thermal-to-visible face image translation is preserving the identity across different spectral bands. Existing work does not effectively … |
David Anghelone; Cunjian Chen; Philippe Faure; A. Ross; A. Dantcheva; | 2021 16th IEEE International Conference on Automatic Face … | 2021-12-15 |
465 | Stochastic Actor-Executor-Critic for Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we draw inspiration from the recent success of the maximum entropy reinforcement learning framework designed for challenging continuous control problems to develop stochastic policies over high dimensional continuous spaces including image representation, generation, and control simultaneously. |
ZIWEI LUO et. al. | arxiv-cs.CV | 2021-12-14 |
466 | A Domain Adaptive Person Re-Identification Based on Dual Attention Mechanism and Camstyle Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Due to the variation in the image capturing process, the difference between source and target sets causes a challenge in unsupervised domain adaptation (UDA) on person … |
CHENGYAN ZHONG et. al. | Algorithms | 2021-12-13 |
467 | Image-to-Image Translation-based Data Augmentation for Robust EV Charging Inlet Detection Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Towards robust EV charging inlet detection, we propose a new dataset (EVCI dataset) and a novel data augmentation method that is based on image-to-image translation where typical image-to-image translation methods synthesize a new image in a different domain given an image. |
Yeonjun Bang; Yeejin Lee; Byeongkeun Kang; | arxiv-cs.CV | 2021-12-09 |
468 | Adverse Weather Image Translation with Asymmetric and Uncertainty-aware GAN Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To address this issue, we propose a novel GAN model, i.e., AU-GAN, which has an asymmetric architecture for adverse domain translation. |
JEONG-GI KWAK et. al. | arxiv-cs.CV | 2021-12-08 |
469 | Neural Photometry-guided Visual Attribute Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a deep learning-based method for propagating spatially-varying visual material attributes (e.g. texture maps or image stylizations) to larger samples of the same or similar materials. |
Carlos Rodriguez-Pardo; Elena Garces; | arxiv-cs.CV | 2021-12-05 |
470 | Panoptic-based Object Style-Align for Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose panoptic-based object style-align generative adversarial networks (POSA-GANs) for image-to-image translation together with a compact panoptic segmentation dataset. |
LIYUN ZHANG et. al. | arxiv-cs.CV | 2021-12-03 |
471 | StyleMesh: Style Transfer for Indoor 3D Scene Reconstructions IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We apply style transfer on mesh reconstructions of indoor scenes. |
Lukas Höllein; Justin Johnson; Matthias Nießner; | arxiv-cs.CV | 2021-12-02 |
472 | FaceTuneGAN: Face Autoencoder for Convolutional Expression Transfer Using Neural Generative Adversarial Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present FaceTuneGAN, a new 3D face model representation decomposing and encoding separately facial identity and facial expression. |
Nicolas Olivier; Kelian Baert; Fabien Danieau; Franck Multon; Quentin Avril; | arxiv-cs.CV | 2021-12-01 |
473 | Semi-supervised Video-driven Facial Animation Transfer for Production IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: We propose a simple algorithm for automatic transfer of facial expressions, from videos to a 3D character, as well as between distinct 3D characters through their rendered … |
LUCIO MOSER et. al. | ACM Transactions on Graphics (TOG) | 2021-12-01 |
474 | FDA-GAN: Flow-based Dual Attention GAN for Human Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To this end, we propose the Flow-based Dual Attention GAN (FDA-GAN) to apply occlusion- and deformation-aware feature fusion for higher generation quality. |
Liyuan Ma; Kejie Huang; Dongxu Wei; Zhaoyan Ming; Haibin Shen; | arxiv-cs.CV | 2021-12-01 |
475 | SpaceEdit: Learning A Unified Editing Space for Open-Domain Image Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Inspired by these efforts, in this paper we propose a unified model for open-domain image editing focusing on color and tone adjustment of open-domain images while keeping their original content and structure. |
JING SHI et. al. | arxiv-cs.CV | 2021-11-30 |
476 | EdiBERT, A Generative Model for Image Editing Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we aim at making a step towards a unified approach for image editing. |
Thibaut Issenhuth; Ugo Tanielian; Jérémie Mary; David Picard; | arxiv-cs.CV | 2021-11-30 |
477 | TRIP: Refining Image-to-Image Translation Via Rival Preferences Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a new model TRIP to coordinate these two goals for high-quality fine-grained translation. |
Yinghua Yao; Yuangang Pan; Ivor W. Tsang; Xin Yao; | arxiv-cs.LG | 2021-11-26 |
478 | ManiFest: Manifold Deformation for Few-shot Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We instead propose ManiFest: a framework for few-shot image translation that learns a context-aware representation of a target domain from a few images only. |
Fabio Pizzati; Jean-François Lalonde; Raoul de Charette; | arxiv-cs.CV | 2021-11-26 |
479 | Attribute-specific Control Units in StyleGAN for Fine-grained Image Manipulation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We evaluate our proposed method in various face attribute manipulation tasks. |
RUI WANG et. al. | arxiv-cs.CV | 2021-11-25 |
480 | EgoRenderer: Rendering Human Avatars from Egocentric Camera Images Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present EgoRenderer, a system for rendering full-body neural avatars of a person captured by a wearable, egocentric fisheye camera that is mounted on a cap or a VR headset. |
Tao Hu; Kripasindhu Sarkar; Lingjie Liu; Matthias Zwicker; Christian Theobalt; | arxiv-cs.CV | 2021-11-24 |
481 | PT-VTON: An Image-Based Virtual Try-On Network with Progressive Pose Attention Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present PT-VTON, a novel pose-transfer-based framework for cloth transfer that enables virtual try-on with arbitrary poses. |
Hanhan Zhou; Tian Lan; Guru Venkataramani; | arxiv-cs.CV | 2021-11-23 |
482 | Delving Into Rectifiers in Style-Based Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we show that the activation function is one of the crucial components in controlling the direction of image synthesis. |
Yipeng Zhang; Bingliang Hu; Hailong Ning; Quang Wang; | arxiv-cs.CV | 2021-11-20 |
483 | Artistic Style Transfer with Internal-external Learning and Contrastive Learning IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Motivated by this, we propose an internal-external style transfer method with two contrastive losses. |
HAIBO CHEN et. al. | nips | 2021-11-20 |
484 | Breaking The Dilemma of Medical Image-to-image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In order to break the dilemma of the existing modes, we propose a new unsupervised mode called RegGAN for medical image-to-image translation. |
LINGKE KONG et. al. | nips | 2021-11-20 |
485 | Global and Local Alignment Networks for Unpaired Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In the paper, to address this issue, we introduce a novel approach, Global and Local Alignment Networks (GLA-Net). |
GUANGLEI YANG et. al. | arxiv-cs.CV | 2021-11-19 |
486 | Palette: Image-to-Image Diffusion Models IF:6 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper develops a unified framework for image-to-image translation based on conditional diffusion models and evaluates this framework on four challenging image-to-image translation tasks, namely colorization, inpainting, uncropping, and JPEG restoration. |
CHITWAN SAHARIA et. al. | arxiv-cs.CV | 2021-11-10 |
487 | Style Transfer with Target Feature Palette and Attention Coloring Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, to solve these problems, a novel artistic stylization method with target feature palettes is proposed, which can transfer key features accurately. |
Suhyeon Ha; Guisik Kim; Junseok Kwon; | arxiv-cs.CV | 2021-11-07 |
488 | Semantically Consistent Image-to-Image Translation for Unsupervised Domain Adaptation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we investigate the problem of UDA from a synthetic computer-generated domain to a similar but real-world domain for learning semantic segmentation. |
Stephan Brehm; Sebastian Scherer; Rainer Lienhart; | arxiv-cs.CV | 2021-11-05 |
489 | Evaluating The Evaluation Metrics for Style Transfer: A Case Study in Multilingual Formality Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we evaluate leading automatic metrics on the oft-researched task of formality style transfer. |
Eleftheria Briakou; Sweta Agrawal; Joel Tetreault; Marine Carpuat; | emnlp | 2021-11-05 |
490 | Collaborative Learning of Bidirectional Decoders for Unsupervised Text Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a collaborative learning framework for unsupervised text style transfer using a pair of bidirectional decoders, one decoding from left to right while the other decoding from right to left. |
Yun Ma; Yangbin Chen; Xudong Mao; Qing Li; | emnlp | 2021-11-05 |
491 | Seamless Satellite-image Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce Seamless Satellite-image Synthesis (SSS), a novel neural architecture to create scale-and-space continuous satellite textures from cartographic data. |
Jialin Zhu; Tom Kelly; | arxiv-cs.CV | 2021-11-05 |
492 | Generic Resources Are What You Need: Style Transfer Tasks Without Task-specific Parallel Training Data IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel approach to this task that leverages generic resources, and without using any task-specific parallel (source-target) data outperforms existing unsupervised approaches on the two most popular style transfer tasks: formality transfer and polarity swap. |
Huiyuan Lai; Antonio Toral; Malvina Nissim; | emnlp | 2021-11-05 |
493 | StyleCLIPDraw: Coupling Content and Style in Text-to-Drawing Synthesis IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce StyleCLIPDraw which adds a style loss to the CLIPDraw text-to-drawing synthesis model to allow artistic control of the synthesized drawings in addition to control of the content via text. |
Peter Schaldenbrand; Zhixuan Liu; Jean Oh; | arxiv-cs.CV | 2021-11-04 |
494 | StyleGAN of All Trades: Image Manipulation with Only Pretrained StyleGAN IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we take a deeper look at the spatial properties of StyleGAN. |
Min Jin Chong; Hsin-Ying Lee; David Forsyth; | arxiv-cs.CV | 2021-11-02 |
495 | Learning Co-segmentation By Segment Swapping for Retrieval and Discovery Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: The goal of this work is to efficiently identify visually similar patterns in images, e.g. identifying an artwork detail copied between an engraving and an oil painting, or recognizing parts of a night-time photograph visible in its daytime counterpart. |
Xi Shen; Alexei A. Efros; Armand Joulin; Mathieu Aubry; | arxiv-cs.CV | 2021-10-29 |
496 | Separating Content and Style for Unsupervised Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose to separate the content code and style code simultaneously in a unified framework. |
Yunfei Liu; Haofei Wang; Yang Yue; Feng Lu; | arxiv-cs.CV | 2021-10-27 |
497 | Recurrent SinGAN: Towards Scale-Agnostic Single Image GANs Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Learning a deep generative model on a single image has attracted considerable attention recently. In this paper, we present a single image generative model, named recurrent … |
Xiaoyu He; Zhenyong Fu; | Proceedings of the 2021 5th International Conference on … | 2021-10-22 |
498 | Fusion of Complementary 2D and 3D Mesostructural Datasets Using Generative Adversarial Networks Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a method for combining information from pairs of distinct but complementary imaging techniques in order to accurately reconstruct the desired multi-phase, high resolution, representative, 3D images. |
Amir Dahari; Steve Kench; Isaac Squires; Samuel J. Cooper; | arxiv-cs.CV | 2021-10-21 |
499 | StyleAlign: Analysis and Applications of Aligned StyleGAN Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we perform an in-depth study of the properties and applications of aligned generative models. |
Zongze Wu; Yotam Nitzan; Eli Shechtman; Dani Lischinski; | arxiv-cs.CV | 2021-10-21 |
500 | STALP: Style Transfer with Auxiliary Limited Pairing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present an approach to example-based stylization of images that uses a single pair of a source image and its stylized counterpart. |
DAVID FUTSCHIK et. al. | arxiv-cs.CV | 2021-10-20 |
501 | Style Agnostic 3D Reconstruction Via Adversarial Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose an approach that enables a differentiable rendering-based learning of 3D objects from images with backgrounds without the need for silhouette supervision. |
Felix Petersen; Bastian Goldluecke; Oliver Deussen; Hilde Kuehne; | arxiv-cs.CV | 2021-10-20 |
502 | Unified Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, a novel approach, the Unified Style Transfer (UST) model, is proposed. |
Guanjie Huang; Hongjian He; Xiang Li; Xingchen Li; Ziang Liu; | arxiv-cs.CV | 2021-10-20 |
503 | FacialGAN: Style Transfer and Attribute Manipulation on Synthetic Faces Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To address this issue, we propose FacialGAN, a novel framework enabling simultaneous rich style transfers and interactive facial attributes manipulation. |
Ricard Durall; Jireh Jam; Dominik Strassel; Moi Hoon Yap; Janis Keuper; | arxiv-cs.CV | 2021-10-18 |
504 | Domain Adaptation for Viewpoint Estimation with Image Generation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Viewpoint estimation is the pre-procedure for purposive perception and fine pose estimation, which has vital applications in robot manipulation and grasping. Unfortunately, the … |
Xunjin Wu; Changsheng Lu; Chaochen Gu; Kaijie Wu; Shanying Zhu; | 2021 International Conference on Control, Automation and … | 2021-10-14 |
505 | Towards Using Clothes Style Transfer for Scenario-aware Person Video Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To further improve the generation performance, we propose a novel framework with disentangled multi-branch encoders and a shared decoder. |
JINGNING XU et. al. | arxiv-cs.CV | 2021-10-14 |
506 | Music Sentiment Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Music sentiment transfer attempts to apply the high level objective of sentiment transfer to the domain of music. |
Miles Sigel; Michael Zhou; Jiebo Luo; | arxiv-cs.SD | 2021-10-12 |
507 | Exploring Content Based Image Retrieval for Highly Imbalanced Melanoma Data Using Style Transfer, Semantic Image Segmentation and Ensemble Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper explores this domain and proposes multiple similarity measures which uses Style Loss and Dice Coefficient via a novel similarity measure called I1-Score. |
Priyam Mehta; | arxiv-cs.CV | 2021-10-12 |
508 | Harnessing The Conditioning Sensorium for Improved Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We evaluate our method on traditional, well-aligned, datasets such as CelebA-HQ, and propose two novel datasets for evaluation on more complex scenes: ClassicTV and FFHQ-Wild. |
Cooper Nederhood; Nicholas Kolkin; Deqing Fu; Jason Salavon; | arxiv-cs.CV | 2021-10-12 |
509 | Bridging The Gap Between Label- and Reference-based Synthesis in Multi-attribute Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: The image-to-image translation (I2IT) model takes a target label or a reference image as the input, and changes a source into the specified target domain. |
Qiusheng Huang; Zhilin Zheng; Xueqi Hu; Li Sun; Qingli Li; | arxiv-cs.CV | 2021-10-11 |
510 | LSC-GAN: Latent Style Code Modeling for Continuous Image-to-image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper intends to build the model for I2I translation among continuous varying domains. |
Qiusheng Huang; Xueqi Hu; Li Sun; Qingli Li; | arxiv-cs.CV | 2021-10-11 |
511 | Digging Into Self-Supervised Learning of Feature Descriptors Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we focus on understanding the limitations of existing self-supervised approaches and propose a set of improvements that combined lead to powerful feature descriptors. |
Iaroslav Melekhov; Zakaria Laskar; Xiaotian Li; Shuzhe Wang; Juho Kannala; | arxiv-cs.CV | 2021-10-10 |
512 | Long-Term Temporally Consistent Unpaired Video Translation From Simulated Surgical 3D Data IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel approach which combines unpaired image translation with neural rendering to transfer simulated to photorealistic surgical abdominal scenes. |
DOMINIK RIVOIR et. al. | iccv | 2021-10-08 |
513 | TransferI2I: Transfer Learning for Image-to-Image Translation From Small Datasets Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a new transfer learning for I2I translation (TransferI2I). |
Yaxing Wang; Hector Laria; Joost van de Weijer; Laura Lopez-Fuentes; Bogdan Raducanu; | iccv | 2021-10-08 |
514 | DRB-GAN: A Dynamic ResBlock Generative Adversarial Network for Artistic Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a Dynamic ResBlock Generative Adversarial Network (DRB-GAN) for artistic style transfer. |
Wenju Xu; Chengjiang Long; Ruisheng Wang; Guanghui Wang; | iccv | 2021-10-08 |
515 | Instance-Wise Hard Negative Example Generation for Contrastive Learning in Unpaired Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we uncover that the negative examples play a critical role in the performance of contrastive learning for image translation. |
Weilun Wang; Wengang Zhou; Jianmin Bao; Dong Chen; Houqiang Li; | iccv | 2021-10-08 |
516 | STRIVE: Scene Text Replacement in Videos Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose replacing scene text in videos using deep style transfer and learned photometric transformations. |
VIJAY KUMAR B G et. al. | iccv | 2021-10-08 |
517 | SurfaceNet: Adversarial SVBRDF Estimation From A Single Image IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper we present SurfaceNet, an approach for estimating spatially-varying bidirectional reflectance distribution function (SVBRDF) material properties from a single image. |
Giuseppe Vecchio; Simone Palazzo; Concetto Spampinato; | iccv | 2021-10-08 |
518 | Manifold Alignment for Semantically Aligned Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Based on this assumption, the style transfer problem is formulated as aligning two multi-manifold distributions and a Manifold Alignment based Style Transfer (MAST) framework is proposed. |
JING HUO et. al. | iccv | 2021-10-08 |
519 | Reality Transform Adversarial Generators for Image Splicing Forgery Detection and Localization IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We view this retouching process as image style transfer and then proposed the fake-to-realistic transformation generator GT. |
Xiuli Bi; Zhipeng Zhang; Bin Xiao; | iccv | 2021-10-08 |
520 | Diverse Image Style Transfer Via Invertible Cross-Space Mapping IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To address this limitation, we propose a Diverse Image Style Transfer (DIST) framework which achieves significant diversity by enforcing an invertible cross-space mapping. |
HAIBO CHEN et. al. | iccv | 2021-10-08 |
521 | Frequency Domain Image Translation: More Photo-Realistic, Better Identity-Preserving IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To solve these challenges, we propose a novel frequency domain image translation (FDIT) framework, exploiting frequency information for enhancing the image generation process. |
MU CAI et. al. | iccv | 2021-10-08 |
522 | StyleFormer: Real-Time Arbitrary Style Transfer Via Parametric Style Composition IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a new feed-forward arbitrary style transfer method, referred to as StyleFormer, which can simultaneously fulfill fine-grained style diversity and semantic content coherency. |
Xiaolei Wu; Zhihao Hu; Lu Sheng; Dong Xu; | iccv | 2021-10-08 |
523 | SPatchGAN: A Statistical Feature Based Discriminator for Unsupervised Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: For unsupervised image-to-image translation, we propose a discriminator architecture which focuses on the statistical features instead of individual patches. |
Xuning Shao; Weidong Zhang; | iccv | 2021-10-08 |
524 | Attack As The Best Defense: Nullifying Image-to-Image Translation GANs Via Limit-Aware Adversarial Attack Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we tackle the problem by a new adversarial attack scheme, namely the Nullifying Attack, which cancels the image translation process and proposes a corresponding framework, the Limit-Aware Self-Guiding Gradient Sliding Attack (LaS-GSA) under a black-box setting. |
Chin-Yuan Yeh; Hsi-Wen Chen; Hong-Han Shuai; De-Nian Yang; Ming-Syan Chen; | iccv | 2021-10-08 |
525 | Unaligned Image-to-Image Translation By Learning to Reweight IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we consider the task of image translation between two unaligned domains, which may arise for various possible reasons. |
Shaoan Xie; Mingming Gong; Yanwu Xu; Kun Zhang; | iccv | 2021-10-08 |
526 | AdaAttN: Revisit Attention Mechanism in Arbitrary Neural Style Transfer IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To alleviate this problem, in this paper, we propose a novel Adaptive Attention Normalization (AdaAttN) module to adaptively perform attentive normalization on per-point basis. |
SONGHUA LIU et. al. | iccv | 2021-10-08 |
527 | Rethinking The Truly Unsupervised Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we tackle image-to-image translation in a fully unsupervised setting, i.e., neither paired images nor domain labels. |
Kyungjune Baek; Yunjey Choi; Youngjung Uh; Jaejun Yoo; Hyunjung Shim; | iccv | 2021-10-08 |
528 | Structure-Transformed Texture-Enhanced Network for Person Image Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we devise a structure-transformed texture-enhanced network to generate high-quality person images and construct the relationships between two tasks. |
Munan Xu; Yuanqi Chen; Shan Liu; Thomas H. Li; Ge Li; | iccv | 2021-10-08 |
529 | Domain-Aware Universal Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To overcome this limitation, we propose a unified architecture, Domain-aware Style Transfer Networks (DSTN) that transfer not only the style but also the property of domain (i.e., domainness) from a given reference image. |
Kibeom Hong; Seogkyu Jeon; Huan Yang; Jianlong Fu; Hyeran Byun; | iccv | 2021-10-08 |
530 | 3DStyleNet: Creating 3D Shapes With Geometric and Texture Style Variations IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a method to create plausible geometric and texture style variations of 3D objects in the quest to democratize 3D content creation. |
Kangxue Yin; Jun Gao; Maria Shugrina; Sameh Khamis; Sanja Fidler; | iccv | 2021-10-08 |
531 | SDA-GAN: Unsupervised Image Translation Using Spectral Domain Attention-Guided Generative Adversarial Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work introduced a novel GAN architecture for unsupervised image translation on the task of face style transform. |
Qizhou Wang; Maksim Makarenko; | arxiv-cs.CV | 2021-10-06 |
532 | DiffusionCLIP: Text-Guided Diffusion Models for Robust Image Manipulation IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Specifically, these approaches often have difficulties in reconstructing images with novel poses, views, and highly variable contents compared to the training data, altering object identity, or producing unwanted image artifacts. To mitigate these problems and enable faithful manipulation of real images, we propose a novel method, dubbed DiffusionCLIP, that performs text-driven image manipulation using diffusion models. |
Gwanghyun Kim; Taesung Kwon; Jong Chul Ye; | arxiv-cs.CV | 2021-10-06 |
533 | Self-Supervised Generative Style Transfer for One-Shot Medical Image Segmentation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Motivated by atlas-based segmentation, we propose a novel volumetric self-supervised learning for data augmentation capable of synthesizing volumetric image-segmentation pairs via learning transformations from a single labeled atlas to the unlabeled data. |
DEVAVRAT TOMAR et. al. | arxiv-cs.CV | 2021-10-05 |
534 | Voice Aging with Audio-Visual Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we apply a similar approach to age a speaker’s voice, referred to as voice aging. |
Justin Wilson; Sunyeong Park; Seunghye J. Wilson; Ming C. Lin; | arxiv-cs.SD | 2021-10-05 |
535 | Causal Representation Learning for Context-Aware Face Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel context-aware face transfer method, called CarTrans, that incorporates causal effects of contextual factors into face representation, and thus is able to be aware of the uncertainty of new contexts. |
Gege Gao; Huaibo Huang; Chaoyou Fu; Ran He; | arxiv-cs.CV | 2021-10-04 |
536 | DRAN: Detailed Region-Adaptive Normalization for Conditional Image Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To evaluate the effectiveness and show the general use of our method, we conduct a set of experiments on makeup transfer and semantic image synthesis. |
YUEMING LYU et. al. | arxiv-cs.CV | 2021-09-29 |
537 | ISF-GAN: An Implicit Style Function for High-Resolution Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This work proposes an implicit style function (ISF) to straightforwardly achieve multi-modal and multi-domain image-to-image translation from pre-trained unconditional generators. |
YAHUI LIU et. al. | arxiv-cs.CV | 2021-09-26 |
538 | Contrastive Unpaired Translation Using Focal Loss for Patch Classification Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This type of approach encourages mapping of corresponding patches to the same location in relation to other patches (negatives) while at the same time improves the output image quality and significantly decreases memory usage as well as the time required to train the model compared to CycleGAN method used as a baseline. |
Bernard Spiegl; | arxiv-cs.CV | 2021-09-25 |
539 | Layered Neural Atlases for Consistent Video Editing IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present a method that decomposes, or unwraps, an input video into a set of layered 2D atlases, each providing a unified representation of the appearance of an object (or background) over the video. |
Yoni Kasten; Dolev Ofri; Oliver Wang; Tali Dekel; | arxiv-cs.CV | 2021-09-23 |
540 | A 3D Mesh-based Lifting-and-Projection Network for Human Pose Transfer |