Paper Digest: Recent Papers on Style Transfer
Paper Digest Team extracted all recent Style Transfer related papers on our radar, and generated highlight sentences for them. The results are then sorted by relevance & date. In addition to this ‘static’ page, we also provide a real-time version of this article, which has more coverage and is updated in real time to include the most recent updates on this topic.
This list is created by the Paper Digest Team. Experience the cutting-edge capabilities of Paper Digest, an innovative AI-powered research platform that empowers you to read, write, get answers and review.
Try us today and unlock the full potential of our services for free!
TABLE 1: Paper Digest: Recent Papers on Style Transfer
Paper | Author(s) | Source | Date | |
---|---|---|---|---|
1 | StyleSSP: Sampling StartPoint Enhancement for Training-free Diffusion-based Method for Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Through a series of experiments, we discovered that an effective startpoint in the sampling stage significantly enhances the style transfer process. Based on this discovery, we propose StyleSSP, which focuses on obtaining a better startpoint to address layout changes of original content and content leakage from style image. |
Ruojun Xu; Weijie Xi; Xiaodi Wang; Yongbo Mao; Zach Cheng; | arxiv-cs.CV | 2025-01-20 |
2 | Dynamic Neural Style Transfer for Artistic Image Generation Using VGG19 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Nevertheless, a number of current techniques continue to encounter obstacles, including lengthy processing times, restricted choices of style images, and the inability to modify the weight ratio of styles. We proposed a neural style transfer system that can add various artistic styles to a desired image to address these constraints allowing flexible adjustments to style weight ratios and reducing processing time. |
Kapil Kashyap; Mehak Garg; Sean Fargose; Sindhu Nair; | arxiv-cs.CV | 2025-01-16 |
3 | Multimodal LLMs Can Reason About Aesthetics in Zero-Shot Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present the first study on how Multimodal LLMs’ (MLLMs) reasoning ability shall be elicited to evaluate the aesthetics of artworks. |
Ruixiang Jiang; Changwen Chen; | arxiv-cs.CV | 2025-01-15 |
4 | Improving Image Captioning By Mimicking Human Reformulation Feedback at Inference-time Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce a novel type of feedback — caption reformulations — and train models to mimic reformulation feedback based on human annotations. |
Uri Berger; Omri Abend; Lea Frermann; Gabriel Stanovsky; | arxiv-cs.CV | 2025-01-08 |
5 | ZDySS — Zero-Shot Dynamic Scene Stylization Using Gaussian Splatting Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce ZDySS, a zero-shot stylization framework for dynamic scenes, allowing our model to generalize to previously unseen style images at inference. |
ABHISHEK SAROHA et. al. | arxiv-cs.CV | 2025-01-07 |
6 | Aesthetic Matters in Music Perception for Image Stylization: A Emotion-driven Music-to-Visual Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Similarly, music research largely focuses on theoretical aspects, with limited exploration of its emotional dimensions and their integration with visual arts. To address these gaps, we introduce EmoMV, an emotion-driven music-to-visual manipulation method that manipulates images based on musical emotions. |
JUNJIE XU et. al. | arxiv-cs.CV | 2025-01-03 |
7 | StyleRWKV: High-Quality and High-Efficiency Style Transfer with RWKV-like Architecture Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we present a novel framework StyleRWKV, to achieve high-quality style transfer with limited memory usage and linear time complexity. |
Miaomiao Dai; Qianyu Zhou; Lizhuang Ma; | arxiv-cs.CV | 2024-12-27 |
8 | Single Trajectory Distillation for Accelerating Image and Video Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This training strategy can not ensure the consistency of whole trajectories. To address this issue, we propose single trajectory distillation (STD) starting from a specific partial noise state. |
SIJIE XU et. al. | arxiv-cs.CV | 2024-12-25 |
9 | DRDM: A Disentangled Representations Diffusion Model for Synthesizing Realistic Person Images Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, existing methods face significant challenges with details missing, limbs distortion and the garment style deviation. To address these issues, we propose a Disentangled Representations Diffusion Model (DRDM) to generate photo-realistic images from source portraits in specific desired poses and appearances. |
Enbo Huang; Yuan Zhang; Faliang Huang; Guangyu Zhang; Yang Liu; | arxiv-cs.CV | 2024-12-25 |
10 | Ensuring Consistency for In-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The former entails incorporating image information during translation, while the latter involves maintaining consistency between the style of the text-image and the original image, ensuring background integrity. To address these consistency requirements, we introduce a novel two-stage framework named HCIIT (High-Consistency In-Image Translation) which involves text-image translation using a multimodal multilingual large language model in the first stage and image backfilling with a diffusion model in the second stage. |
CHENGPENG FU et. al. | arxiv-cs.CL | 2024-12-23 |
11 | Style Transfer Dataset: What Makes A Good Stylization? Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a new dataset with the goal of advancing image style transfer – the task of rendering one image in the style of another image. |
Victor Kitov; Valentin Abramov; Mikhail Akhtyrchenko; | arxiv-cs.CV | 2024-12-22 |
12 | Diffusion-Based Conditional Image Editing Through Optimized Inference with Guidance Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a simple but effective training-free approach for text-driven image-to-image translation based on a pretrained text-to-image diffusion model. |
Hyunsoo Lee; Minsoo Kang; Bohyung Han; | arxiv-cs.CV | 2024-12-20 |
13 | Enhancing Nighttime Vehicle Detection with Day-to-Night Style Transfer and Labeling-Free Augmentation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This issue is particularly pronounced in transportation applications, such as detecting vehicles and other objects of interest on rural roads at night, where street lighting is often absent, and headlights may introduce undesirable glare. This study addresses these challenges by introducing a novel framework for labeling-free data augmentation, leveraging CARLA-generated synthetic data for day-to-night image style transfer. |
Yunxiang Yang; Hao Zhen; Yongcan Huang; Jidong J. Yang; | arxiv-cs.CV | 2024-12-20 |
14 | Can We Get Rid of Handcrafted Feature Extractors? SparseViT: Nonsemantics-Centered, Parameter-Efficient Image Manipulation Localization Through Spare-Coding Transformer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Hence, in this paper, we propose a Sparse Vision Transformer (SparseViT), which reformulates the dense, global self-attention in ViT into a sparse, discrete manner. |
LEI SU et. al. | arxiv-cs.CV | 2024-12-19 |
15 | Content-style Disentangled Representation for Controllable Artistic Image Stylization and Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, current methods for content and style disentanglement primarily rely on image information for supervision, which leads to two problems: 1) models can only support one modality for style or content input;2) incomplete disentanglement resulting in semantic interference from the reference image. To address the above issues, this paper proposes a content-style representation disentangling method for controllable artistic image stylization and generation. |
Ma Zhuoqi; Zhang Yixuan; You Zejun; Tian Long; Liu Xiyang; | arxiv-cs.CV | 2024-12-18 |
16 | Prompt Augmentation for Self-supervised Text-guided Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In response, our work introduces prompt augmentation, a method amplifying a single input prompt into several target prompts, strengthening textual context and enabling localised image editing. |
Rumeysa Bodur; Binod Bhattarai; Tae-Kyun Kim; | arxiv-cs.CV | 2024-12-17 |
17 | LineArt: A Knowledge-guided Training-free High-quality Appearance Transfer for Design Drawing with Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present LineArt, a framework that transfers complex appearance onto detailed design drawings, facilitating design and artistic creation. |
XI WANG et. al. | arxiv-cs.CV | 2024-12-16 |
18 | UnMA-CapSumT: Unified and Multi-Head Attention-driven Caption Summarization Transformer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To the best of our knowledge, no such work exists that provided a description that integrates different captioning methods to describe the contents of an image with factual and stylized (romantic and humorous) elements. To overcome these limitations, this paper presents a novel Unified Attention and Multi-Head Attention-driven Caption Summarization Transformer (UnMA-CapSumT) based Captioning Framework. |
Dhruv Sharma; Chhavi Dhiman; Dinesh Kumar; | arxiv-cs.CV | 2024-12-16 |
19 | Learning Flow Fields in Attention for Controllable Person Image Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We attribute these distortions to inadequate attention to corresponding regions in the reference image. To address this, we thereby propose learning flow fields in attention (Leffa), which explicitly guides the target query to attend to the correct reference key in the attention layer during training. |
ZIJIAN ZHOU et. al. | arxiv-cs.CV | 2024-12-11 |
20 | StyleStudio: Text-Driven Style Transfer with Selective Control of Style Elements Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Recent advancements in text-to-image models have improved the nuance of style transformations, yet significant challenges remain, particularly with overfitting to reference styles, limiting stylistic control, and misaligning with textual content. In this paper, we propose three complementary strategies to address these issues. |
Mingkun Lei; Xue Song; Beier Zhu; Hao Wang; Chi Zhang; | arxiv-cs.CV | 2024-12-11 |
21 | StyleMaster: Stylize Your Video with Artistic Generation and Translation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Style control has been popular in video generation models. Existing methods often generate videos far from the given style, cause content leakage, and struggle to transfer one … |
ZIXUAN YE et. al. | arxiv-cs.CV | 2024-12-10 |
22 | StyleMark: A Robust Watermarking Method for Art Style Images Against Black-Box Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Unfortunately, AST-generated images lose the structural and semantic information of the original style image, hindering end-to-end robust tracking by watermarks. To fill this gap, we propose StyleMark, the first robust watermarking method for black-box AST, which can be seamlessly applied to art style images achieving precise attribution of artistic styles after AST. |
Yunming Zhang; Dengpan Ye; Sipeng Shen; Jun Wang; | arxiv-cs.CV | 2024-12-09 |
23 | Continuous Video Process: Modeling Videos As Continuous Multi-Dimensional Processes for Video Prediction Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In our paper, we introduce a novel model class, that treats video as a continuous multi-dimensional process rather than a series of discrete frames. |
Gaurav Shrivastava; Abhinav Shrivastava; | arxiv-cs.CV | 2024-12-06 |
24 | Learning Artistic Signatures: Symmetry Discovery and Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Conversely, more recent work with diffusion models offers compelling empirical results but provides little theoretical grounding. To address these issues, we propose an alternative definition of artistic style. |
Emma Finn; T. Anderson Keller; Emmanouil Theodosis; Demba E. Ba; | arxiv-cs.CV | 2024-12-05 |
25 | D-LORD for Motion Stylization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper introduces a novel framework named D-LORD (Double Latent Optimization for Representation Disentanglement), which is designed for motion stylization (motion style transfer and motion retargeting). |
Meenakshi Gupta; Mingyuan Lei; Tat-Jen Cham; Hwee Kuan Lee; | arxiv-cs.CV | 2024-12-05 |
26 | SGSST: Scaling Gaussian Splatting StyleTransfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This work introduces SGSST: Scaling Gaussian Splatting Style Transfer, an optimization-based method to apply style transfer to pretrained 3DGS scenes. |
Bruno Galerne; Jianling Wang; Lara Raad; Jean-Michel Morel; | arxiv-cs.CV | 2024-12-04 |
27 | Style3D: Attention-guided Multi-view Style Transfer for 3D Object Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present Style3D, a novel approach for generating stylized 3D objects from a content image and a style image. |
Bingjie Song; Xin Huang; Ruting Xie; Xue Wang; Qing Wang; | arxiv-cs.CV | 2024-12-04 |
28 | GIST: Towards Photorealistic Style Transfer Via Multiscale Geometric Representations Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Motivated by the ability of multiscale geometric image representations to capture fine-grained details and global structure, we propose GIST: Geometric-based Image Style Transfer, a novel Style Transfer technique that exploits the geometric properties of content and style images. |
Renan A. Rojas-Gomez; Minh N. Do; | arxiv-cs.CV | 2024-12-03 |
29 | RaD: A Metric for Medical Image Distribution Comparison in Out-of-Domain Detection and Other Applications Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce a new perceptual metric tailored for medical images: Radiomic Feature Distance (RaD), which utilizes standardized, clinically meaningful and interpretable image features. |
NICHOLAS KONZ et. al. | arxiv-cs.CV | 2024-12-02 |
30 | A White-Box False Positive Adversarial Attack Method on Contrastive Loss Based Offline Handwritten Signature Verification Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we tackle the challenge of white-box false positive adversarial attacks on contrastive loss based offline handwritten signature verification models. |
Zhongliang Guo; Weiye Li; Yifei Qian; Ognjen Arandjelovic; Lei Fang; | aistats | 2024-12-01 |
31 | Z-STAR+: A Zero-shot Style Transfer Method Via Adjusting Style Distribution Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce a Cross-attention Reweighting module that utilizes local content features to query style image information best suited to the input patch, thereby aligning the style distribution of the stylized results with that of the style image. |
Yingying Deng; Xiangyu He; Fan Tang; Weiming Dong; | arxiv-cs.CV | 2024-11-28 |
32 | Music2Fail: Transfer Music to Failed Recorder Style Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we investigate another style transfer scenario called “failed-music style transfer”. |
CHON IN LEONG et. al. | arxiv-cs.SD | 2024-11-27 |
33 | Interleaved Scene Graph for Interleaved Text-and-Image Generation Assessment Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Models designed to generate interleaved text and images face challenges in ensuring consistency within and across these modalities. To address these challenges, we present ISG, a comprehensive evaluation framework for interleaved text-and-image generation. |
DONGPING CHEN et. al. | arxiv-cs.CV | 2024-11-26 |
34 | CapHDR2IR: Caption-Driven Transfer from Visible Light to Infrared Domain Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This occurs when multiple objects with similar temperatures appear indistinguishable in the training data, further exacerbating the loss of fidelity. To solve this challenge, this paper proposes CapHDR2IR, a novel framework incorporating vision-language models using high dynamic range (HDR) images as inputs to generate IR images. |
JINGCHAO PENG et. al. | arxiv-cs.CV | 2024-11-25 |
35 | A Training-Free Approach for Music Style Transfer with Latent Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper introduces a novel training-free approach leveraging pre-trained Latent Diffusion Models (LDMs). |
SOOYOUNG KIM et. al. | arxiv-cs.SD | 2024-11-24 |
36 | Omni-IML: Towards Unified Image Manipulation Localization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To this end, we propose Omni-IML, the first generalist model to unify diverse IML tasks. |
Chenfan Qu; Yiwu Zhong; Fengjun Guo; Lianwen Jin; | arxiv-cs.CV | 2024-11-22 |
37 | HyperGAN-CLIP: A Unified Framework for Domain Adaptation, Image Synthesis and Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Despite their success, adapting these models to diverse tasks such as domain adaptation, reference-guided synthesis, and text-guided manipulation with limited training data remains challenging. Towards this end, in this study, we present a novel framework that significantly extends the capabilities of a pre-trained StyleGAN by integrating CLIP space via hypernetworks. |
ABDUL BASIT ANEES et. al. | arxiv-cs.CV | 2024-11-19 |
38 | Towards Multi-View Consistent Style Transfer with One-Step Diffusion Via Vision Conditioning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To efficiently adapt the pre-trained model for multi-view style transfer on small datasets, we introduce a vision condition module to extract style information from the reference style image to serve as conditional input for the diffusion model and employ LoRA in diffusion model for adaptation. |
YUSHEN ZUO et. al. | arxiv-cs.CV | 2024-11-15 |
39 | Mechanisms of Generative Image-to-Image Translation Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a streamlined image-to-image translation network with a simpler architecture compared to existing models. |
Guangzong Chen; Mingui Sun; Zhi-Hong Mao; Kangni Liu; Wenyan Jia; | arxiv-cs.CV | 2024-11-15 |
40 | Artistic Neural Style Transfer Algorithms with Activation Smoothing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we re-implement image-based NST, fast NST, and arbitrary NST. |
XIANGTIAN LI et. al. | arxiv-cs.CV | 2024-11-12 |
41 | TCSinger: Zero-Shot Singing Voice Synthesis with Style Transfer and Multi-Level Style Control Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Furthermore, current SVS models often fail to generate singing voices rich in stylistic nuances for unseen singers. To address these challenges, we introduce TCSinger, the first zero-shot SVS model for style transfer across cross-lingual speech and singing styles, along with multi-level style control. |
YU ZHANG et. al. | emnlp | 2024-11-11 |
42 | AI-Driven Stylization of 3D Environments Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this system, we discuss methods to stylize a scene of 3D primitive objects into a higher fidelity 3D scene using novel 3D representations like NeRFs and 3D Gaussian Splatting. |
Yuanbo Chen; Yixiao Kang; Yukun Song; Cyrus Vachha; Sining Huang; | arxiv-cs.CV | 2024-11-08 |
43 | NCST: Neural-based Color Style Transfer for Video Retouching Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Typically, users cannot fine-tune the resulting images or videos. To tackle this issue, we introduce a method that predicts specific parameters for color style transfer using two images. |
Xintao Jiang; Yaosen Chen; Siqin Zhang; Wei Wang; Xuming Wen; | arxiv-cs.CV | 2024-10-31 |
44 | A Practical Style Transfer Pipeline for 3D Animation: Insights from Production R&D Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents the insights from our development process, where we explored various options to balance quality, artist control, and workload, leading to several key decisions. |
Hideki Todo; Yuki Koyama; Kunihiro Sakai; Akihiro Komiya; Jun Kato; | arxiv-cs.GR | 2024-10-31 |
45 | Kandinsky 3: Text-to-Image Synthesis for Multifunctional Generative Framework Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present Kandinsky 3, a novel T2I model based on latent diffusion, achieving a high level of quality and photorealism. |
VLADIMIR ARKHIPKIN et. al. | arxiv-cs.CV | 2024-10-28 |
46 | IconDM: Text-Guided Icon Set Expansion Using Diffusion Models Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Icons are ubiquitous visual elements in graphic design. However, their creation is non-trivial and time-consuming. To this end, we draw inspiration from the booming text-to-image … |
JIAWEI LIN et. al. | ACM Multimedia | 2024-10-28 |
47 | UniVST: A Unified Framework for Training-free Localized Video Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents UniVST, a unified framework for localized video style transfer based on diffusion model. |
QUANJIAN SONG et. al. | arxiv-cs.CV | 2024-10-26 |
48 | DiffuseST: Unleashing The Capability of The Diffusion Model for Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a novel and training-free approach for style transfer, combining textual embedding with spatial features and separating the injection of content or style. |
Ying Hu; Chenyi Zhuang; Pan Gao; | arxiv-cs.CV | 2024-10-19 |
49 | Group Diffusion Transformers Are Unsupervised Multitask Learners Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: While large language models (LLMs) have revolutionized natural language processing with their task-agnostic capabilities, visual generation tasks such as image translation, style transfer, and character customization still rely heavily on supervised, task-specific datasets. In this work, we introduce Group Diffusion Transformers (GDTs), a novel framework that unifies diverse visual generation tasks by redefining them as a group generation problem. |
LIANGHUA HUANG et. al. | arxiv-cs.CV | 2024-10-19 |
50 | 4DStyleGaussian: Zero-shot 4D Style Transfer with Gaussian Splatting Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce 4DStyleGaussian, a novel 4D style transfer framework designed to achieve real-time stylization of arbitrary style references while maintaining reasonable content affinity, multi-view consistency, and temporal coherence. |
Wanlin Liang; Hongbin Xu; Weitao Chen; Feng Xiao; Wenxiong Kang; | arxiv-cs.CV | 2024-10-14 |
51 | TextCtrl: Diffusion-based Scene Text Editing with Prior Guidance Control Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: GAN-based STE methods generally encounter a common issue of model generalization, while Diffusion-based STE methods suffer from undesired style deviations. To address these problems, we propose TextCtrl, a diffusion-based method that edits text with prior guidance control. |
Weichao Zeng; Yan Shu; Zhenhang Li; Dongbao Yang; Yu Zhou; | arxiv-cs.CV | 2024-10-13 |
52 | EBDM: Exemplar-guided Image Translation with Brownian-bridge Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel approach termed Exemplar-guided Image Translation with Brownian-Bridge Diffusion Models (EBDM). |
Eungbean Lee; Somi Jeong; Kwanghoon Sohn; | arxiv-cs.CV | 2024-10-13 |
53 | TextMaster: Universal Controllable Text Edit Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our approach employs adaptive standard letter spacing as guidance during training and uses adaptive mask boosting to prevent the leakage of text position and size information. |
AOQIANG WANG et. al. | arxiv-cs.CV | 2024-10-13 |
54 | Bridging Text and Image for Artist Style Transfer Via Contrastive Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a Contrastive Learning for Artistic Style Transfer (CLAST) that leverages advanced image-text encoders to control arbitrary style transfer. |
Zhi-Song Liu; Li-Wen Wang; Jun Xiao; Vicky Kalogeiton; | arxiv-cs.CV | 2024-10-12 |
55 | NaRCan: Natural Refined Canonical Image with Integration of Diffusion Prior for Video Editing Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Existing models often fail to maintain sequence temporal consistency, disrupting frame transitions. To tackle this issue, this paper introduces NaRCan, a video editing framework that integrates a hybrid deformation field network with diffusion priors. |
TING-HSUAN CHEN et. al. | nips | 2024-10-07 |
56 | Beyond Imperfections: A Conditional Inpainting Approach for End-to-End Artifact Removal in VTON and Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Artifacts often degrade the visual quality of virtual try-on (VTON) and pose transfer applications, impacting user experience. This study introduces a novel conditional inpainting technique designed to detect and remove such distortions, improving image aesthetics. |
Aref Tabatabaei; Zahra Dehghanian; Maryam Amirmazlaghani; | arxiv-cs.CV | 2024-10-05 |
57 | Tuning Timestep-Distilled Diffusion Model Using Pairwise Sample Optimization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present an algorithm named pairwise sample optimization (PSO), which enables the direct fine-tuning of an arbitrary timestep-distilled diffusion model. |
ZICHEN MIAO et. al. | arxiv-cs.CV | 2024-10-04 |
58 | PixelShuffler: A Simple Image Translation Through Pixel Rearrangement Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a novel pixel shuffle method that addresses the image-to-image translation problem generally with a specific demonstrative application in style transfer. |
Omar Zamzam; | arxiv-cs.CV | 2024-10-03 |
59 | Harnessing The Latent Diffusion Model for Training-Free Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a training-free style transfer algorithm, Style Tracking Reverse Diffusion Process (STRDP) for a pretrained Latent Diffusion Model (LDM). |
Kento Masui; Mayu Otani; Masahiro Nomura; Hideki Nakayama; | arxiv-cs.CV | 2024-10-02 |
60 | A Pavement Crack Translator for Data Augmentation and Pixel-Level Detection Based on Weakly Supervised Learning Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Recent state-of-the-art pavement crack detection algorithms are data-driven and domain-sensitive due to their heavy reliance on datasets. Establishing a high-quality pavement … |
JINGTAO ZHONG et. al. | IEEE Transactions on Intelligent Transportation Systems | 2024-10-01 |
61 | Revisiting Feature Disentanglement Strategy in Diffusion Training and Breaking Conditional Independence Assumption in Sampling Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a training framework for feature disentanglement of Diffusion Models (FDiff). |
WONWOONG CHO et. al. | eccv | 2024-09-30 |
62 | WaSt-3D: Wasserstein-2 Distance for Scene-to-Scene Stylization on 3D Gaussians Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In our work, we leverage an explicit Gaussian Scale (GS) representation and directly match the distributions of Gaussians between style and content scenes using the Earth Mover’s Distance (EMD). |
DMYTRO KOTOVENKO et. al. | eccv | 2024-09-30 |
63 | LEGO: Learning EGOcentric Action Frame Generation Via Visual Instruction Tuning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce a novel problem – egocentric action frame generation. |
BOLIN LAI et. al. | eccv | 2024-09-30 |
64 | Towards Compact Reversible Image Representations for Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we learn compact neural representations for style transfer motivated from an information theoretical perspective. |
XIYAO LIU et. al. | eccv | 2024-09-30 |
65 | InstaStyle: Inversion Noise of A Stylized Image Is Secretly A Style Adviser Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose InstaStyle, a novel approach that excels in generating high-fidelity stylized images with only a single reference image. |
XING CUI et. al. | eccv | 2024-09-30 |
66 | Pixel-Aware Stable Diffusion for Realistic Image Super-Resolution and Personalized Stylization IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a pixel-aware stable diffusion (PASD) network to achieve robust Real-ISR and personalized image stylization. |
Tao Yang; Rongyuan Wu; Peiran Ren; Xuansong Xie; Lei Zhang; | eccv | 2024-09-30 |
67 | Implicit Style-Content Separation Using B-LoRA IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce B-LoRA, a method that leverages LoRA (Low-Rank Adaptation) to implicitly separate the style and content components of a single image, facilitating various image stylization tasks. |
Yarden Frenkel; Yael Vinker; Ariel Shamir; Danny Cohen-Or; | eccv | 2024-09-30 |
68 | WaSt-3D: Wasserstein-2 Distance for Scene-to-Scene Stylization on 3D Gaussians Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In our work, we leverage an explicit Gaussian Splatting (GS) representation and directly match the distributions of Gaussians between style and content scenes using the Earth Mover’s Distance (EMD). |
DMYTRO KOTOVENKO et. al. | arxiv-cs.CV | 2024-09-26 |
69 | Copying Style, Extracting Value: Illustrators’ Perception of AI Style Transfer and Its Impact on Creative Labor Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We provided four illustrators with a model fine-tuned to their style and conducted semi-structured interviews about the model’s successes, limitations, and potential uses. |
Julien Porquet; Sitong Wang; Lydia B. Chilton; | arxiv-cs.HC | 2024-09-25 |
70 | Pix2Next: Leveraging Vision Foundation Models for RGB to NIR Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes Pix2Next, a novel image-to-image translation framework designed to address the challenge of generating high-quality Near-Infrared (NIR) images from RGB inputs. |
YOUNGWAN JIN et. al. | arxiv-cs.CV | 2024-09-25 |
71 | AEANet: Affinity Enhanced Attentional Networks for Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, existing style transfer methods often significantly damage the texture lines of the content image during the style transformation. To address these issues, we propose affinity-enhanced attentional network, which include the content affinity-enhanced attention (CAEA) module, the style affinity-enhanced attention (SAEA) module, and the hybrid attention (HA) module. |
Gen Li; Xianqiu Zheng; Yujian Li; | arxiv-cs.CV | 2024-09-22 |
72 | Embedded Image-to-Image Translation for Efficient Sim-to-Real Transfer in Learning-based Robot-Assisted Soft Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel approach utilizing image translation models to mitigate domain mismatches and facilitate efficient robot skill learning in a simulated environment. |
Jacinto Colan; Keisuke Sugita; Ana Davila; Yutaro Yamada; Yasuhisa Hasegawa; | arxiv-cs.RO | 2024-09-16 |
73 | Mamba-ST: State Space Model for Efficient Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To overcome the above, this paper explores a novel design of Mamba, an emergent State-Space Model (SSM), called Mamba-ST, to perform style transfer. |
FILIPPO BOTTI et. al. | arxiv-cs.CV | 2024-09-16 |
74 | One-Shot Learning for Pose-Guided Person Image Synthesis in The Wild Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, naively applying test-time tuning results in inconsistencies in facial identities and appearance attributes. To address this, we introduce a Visual Consistency Module (VCM), which enhances appearance consistency by combining the face, text, and image embedding. |
DONGQI FAN et. al. | arxiv-cs.CV | 2024-09-14 |
75 | MagicStyle: Portrait Stylization Based on Reference Image Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This challenge becomes even more pronounced when the content image is a portrait which has complex textural details. To address this challenge, we propose a diffusion model-based reference image stylization method specifically for portraits, called MagicStyle. |
Zhaoli Deng; Kaibin Zhou; Fanyi Wang; Zhenpeng Mi; | arxiv-cs.CV | 2024-09-12 |
76 | StructuReiser: A Structure-preserving Video Stylization Method Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce StructuReiser, a novel video-to-video translation method that transforms input videos into stylized sequences using a set of user-provided keyframes. |
Radim Spetlik; David Futschik; Daniel Sykora; | arxiv-cs.CV | 2024-09-09 |
77 | MRStyle: A Unified Framework for Color Style Transfer with Multi-Modality Reference Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce MRStyle, a comprehensive framework that enables color style transfer using multi-modality reference, including image and text. |
JIANCHENG HUANG et. al. | arxiv-cs.CV | 2024-09-08 |
78 | Seed-to-Seed: Image Translation in Diffusion Seed Space Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce Seed-to-Seed Translation (StS), a novel approach for Image-to-Image Translation using diffusion models (DMs), aimed at translations that require close adherence to the structure of the source image. |
Or Greenberg; Eran Kishon; Dani Lischinski; | arxiv-cs.CV | 2024-09-01 |
79 | ST2SI: Image Style Transfer Via Vision Transformer Using Spatial Interaction Related Papers Related Patents Related Grants Related Venues Related Experts View |
Wenshu Li; Yinliang Chen; Xiaoying Guo; Xiaoyu He; | Comput. Graph. | 2024-09-01 |
80 | Style Transfer: From Stitching to Neural Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This article compares two style transfer methods in image processing: the traditional method, which synthesizes new images by stitching together small patches from existing images, and a modern machine learning-based approach that uses a segmentation network to isolate foreground objects and apply style transfer solely to the background. |
XINHE XU et. al. | arxiv-cs.CV | 2024-09-01 |
81 | CSGO: Content-Style Composition in Text-to-Image Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we present a data construction pipeline for content-style-stylized image triplets that generates and automatically cleanses stylized data triplets. |
PENG XING et. al. | arxiv-cs.CV | 2024-08-29 |
82 | G3DST: Generalizing 3D Style Transfer with Neural Radiance Fields Across Scenes and Styles Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we overcome the limitations of existing methods by rendering stylized novel views from a NeRF without the need for per-scene or per-style optimization. |
Adil Meric; Umut Kocasari; Matthias Nießner; Barbara Roessle; | arxiv-cs.CV | 2024-08-24 |
83 | Prompt-Softbox-Prompt: A Free-text Embedding Control for Image Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we provide a comprehensive and in-depth analysis of text embeddings in Stable Diffusion XL, offering three key insights. |
Yitong Yang; Yinglin Wang; Jing Wang; Tian Zhang; | arxiv-cs.CV | 2024-08-24 |
84 | Query-Efficient Video Adversarial Attack with Stylized Logo Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Moreover, compared to a plethora of methods targeting image classifiers, video adversarial attacks are still not that popular. Therefore, to generate adversarial examples with a low budget and to provide them with a higher verisimilitude, we propose a novel black-box video attack framework, called Stylized Logo Attack (SLA). |
DUOXUN TANG et. al. | arxiv-cs.CV | 2024-08-21 |
85 | FAGStyle: Feature Augmentation on Geodesic Surface for Zero-shot Text-guided Diffusion Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Despite their versatility, these methods often struggle with maintaining style consistency, reflecting the described style accurately, and preserving the content of the target image. To address these challenges, we introduce FAGStyle, a zero-shot text-guided diffusion image style transfer method. |
Yuexing Han; Liheng Ruan; Bing Wang; | arxiv-cs.CV | 2024-08-20 |
86 | Structure-preserving Image Translation for Depth Estimation in Colonoscopy Video Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a general pipeline of structure-preserving synthetic-to-real (sim2real) image translation (producing a modified version of the input image) to retain depth geometry through the translation process. |
Shuxian Wang; Akshay Paruchuri; Zhaoxi Zhang; Sarah McGill; Roni Sengupta; | arxiv-cs.CV | 2024-08-19 |
87 | StyleBrush: Style Extraction and Transfer from A Single Image Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose StyleBrush, a method that accurately captures styles from a reference image and “brushes” the extracted style onto other input visual content. |
WANCHENG FENG et. al. | arxiv-cs.CV | 2024-08-18 |
88 | Learning A Low-Level Vision Generalist Via Visual Task Prompt Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In addition, these methods are sensitive to prompt image content and often struggle with low-frequency information processing. In this paper, we propose a Visual task Prompt-based Image Processing (VPIP) framework to overcome these challenges. |
XIANGYU CHEN et. al. | arxiv-cs.CV | 2024-08-16 |
89 | The Dawn of KAN in Image-to-Image (I2I) Translation: Integrating Kolmogorov-Arnold Networks with GANs for Unpaired I2I Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This study aims to demonstrate that the Kolmogorov-Arnold Network (KAN) can effectively replace the Multi-layer Perceptron (MLP) method in generative AI, particularly in the subdomain of image-to-image translation, to achieve better generative quality. |
Arpan Mahara; Naphtali D. Rishe; Liangdong Deng; | arxiv-cs.CV | 2024-08-15 |
90 | An Analysis for Image-to-Image Translation and Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: With the development of generative technologies in deep learning, a large number of image-to-image translation and style transfer models have emerged at an explosive rate in … |
Xiaoming Yu; Jie Tian; Zhenhua Hu; | arxiv-cs.CV | 2024-08-12 |
91 | InstantStyleGaussian: Efficient Art Style Transfer with 3D Gaussian Splatting Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present InstantStyleGaussian, an innovative 3D style transfer method based on the 3D Gaussian Splatting (3DGS) scene representation. |
Xin-Yi Yu; Jun-Xin Yu; Li-Bo Zhou; Yan Wei; Lin-Lin Ou; | arxiv-cs.CV | 2024-08-08 |
92 | CLIP-based Point Cloud Classification Via Point Cloud to Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Secondly, the adapter only relies on the global representation of the multi-view features. Motivated by this observation, we propose a Pretrained Point Cloud to Image Translation Network (PPCITNet) that produces generalized colored images along with additional salient visual cues to the point cloud depth maps so that it can achieve promising performance on point cloud classification and understanding. |
Shuvozit Ghose; Manyi Li; Yiming Qian; Yang Wang; | arxiv-cs.CV | 2024-08-07 |
93 | D2Styler: Advancing Arbitrary Style Transfer with Discrete Diffusion Methods Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel framework called D$^2$Styler (Discrete Diffusion Styler) that leverages the discrete representational capability of VQ-GANs and the advantages of discrete diffusion, including stable training and avoidance of mode collapse. |
Onkar Susladkar; Gayatri Deshmukh; Sparsh Mittal; Parth Shastri; | arxiv-cs.CV | 2024-08-07 |
94 | A Multi-Level Cross-Attention Image Registration Method for Visible and Infrared Small Unmanned Aerial Vehicle Targets Via Image Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Small UAV target detection and tracking based on cross-modality image fusion have gained widespread attention. Due to the limited feature information available from small UAVs in … |
WEN JIANG et. al. | Remote. Sens. | 2024-08-07 |
95 | IPAdapter-Instruct: Resolving Ambiguity in Image-based Conditioning Using Instruct Prompts Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose IPAdapter-Instruct, which combines natural-image conditioning with “Instruct” prompts to swap between interpretations for the same conditioning image: style transfer, object extraction, both, or something else still? |
CIARA ROWLES et. al. | arxiv-cs.CV | 2024-08-06 |
96 | FastEdit: Fast Text-Guided Single-Image Editing Via Semantic-Aware Diffusion Fine-Tuning Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Conventional Text-guided single-image editing approaches require a two-step process, including fine-tuning the target text embedding for over 1K iterations and the generative … |
Zhi Chen; Zecheng Zhao; Yadan Luo; Zi Huang; | ArXiv | 2024-08-06 |
97 | Who Looks Like Me: Semantic Routed Image Harmonization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Image harmonization, aiming to seamlessly blend extraneous foreground objects with background images, is a promising and challenging task.Ensuring a synthetic image appears realistic requires maintaining consistency in visual characteristics, such as texture and style, across global and semantic regions.In this paper, We approach image harmonization as a semantic routed style transfer problem, and propose an imageharmonization model by routing semantic similarity explicitly to enhance the consistency of appearance characteristics.To refine calculate the similarity between the composed foreground and background instance, we propose an InstanceSimilarity Evaluation Module(ISEM). |
JINSHENG SUN et. al. | ijcai | 2024-08-03 |
98 | Diffutoon: High-Resolution Editable Toon Shading Via Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we model the toon shading problem as four subproblems, i.e., stylization, consistency enhancement, structure guidance, and colorization. |
Zhongjie Duan; Chengyu Wang; Cen Chen; Weining Qian; Jun Huang; | ijcai | 2024-08-03 |
99 | FBSDiff: Plug-and-Play Frequency Band Substitution of Diffusion Features for Highly Controllable Text-Driven Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To guide T2I generation with a reference image, we propose to decompose diverse guiding factors with different frequency bands of diffusion features in the DCT spectral space, and accordingly devise a novel frequency band substitution layer which realizes dynamic control of the reference image to the T2I generation result in a plug-and-play manner. |
Xiang Gao; Jiaying Liu; | arxiv-cs.CV | 2024-08-02 |
100 | StyleRF-VolVis: Style Transfer of Neural Radiance Fields for Expressive Volume Visualization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper introduces StyleRF-VolVis, an innovative style transfer framework for expressive volume visualization (VolVis) via neural radiance field (NeRF). |
Kaiyuan Tang; Chaoli Wang; | arxiv-cs.GR | 2024-07-31 |
101 | Controllable Neural Style Transfer for Dynamic Meshes Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we propose a novel mesh stylization technique that improves previous NST works in several ways. |
Guilherme Gomes Haetinger; Jingwei Tang; Raphael Ortiz; Paul Kanyuk; Vinicius Azevedo; | siggraph | 2024-07-28 |
102 | Toonify3D: StyleGAN-based 3D Stylized Face Generator Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our goal is to create expressive 3D faces by turning Toonify into a 3D stylized face generator. |
WONJONG JANG et. al. | siggraph | 2024-07-28 |
103 | Color-SD: Stable Diffusion Model Already Has A Color Style Noisy Latent Space Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: We present Color-SD, a comprehensive color style transfer framework that utilizes either image or text references. Built on the pretrained Stable Diffusion Model, Color-SD … |
Jiancheng Huang; Mingfu Yan; Yifan Liu; Shifeng Chen; | 2024 IEEE International Conference on Multimedia and Expo … | 2024-07-15 |
104 | StyleSplat: 3D Object Style Transfer with Gaussian Splatting Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce StyleSplat, a lightweight method for stylizing 3D objects in scenes represented by 3D Gaussians from reference style images. |
Sahil Jain; Avik Kuthiala; Prabhdeep Singh Sethi; Prakanshul Saxena; | arxiv-cs.CV | 2024-07-12 |
105 | Deep Learning-Powered Optical Microscopy for Steel Research Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The success of machine learning (ML) models in object or pattern recognition naturally leads to ML being employed in the classification of the microstructure of steel surfaces. … |
Š. MIKMEKOVÁ et. al. | Mach. Learn. Knowl. Extr. | 2024-07-11 |
106 | Ada-adapter:Fast Few-shot Style Personlization of Diffusion Model with Pre-trained Image Encoder Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose Ada-Adapter, a novel framework for few-shot style personalization of diffusion models. |
JIA LIU et. al. | arxiv-cs.CV | 2024-07-07 |
107 | MSF: A Multi-Scale Fusion Generative Adversarial Network for SAR-to-Optical Image Translation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: This paper proposes an image translation method based on multi-scale fusion GAN (MFS) network. In MFS network, there are two modules: optical image generation sub-network (OGS), … |
YONGKANG CHEN et. al. | IGARSS 2024 – 2024 IEEE International Geoscience and Remote … | 2024-07-07 |
108 | Frequency-Controlled Diffusion Model for Versatile Text-Guided Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes frequency-controlled diffusion model (FCDiffusion), an end-to-end diffusion-based framework that contributes a novel solution to text-guided I2I from a frequency-domain perspective. |
Xiang Gao; Zhengbo Xu; Junhan Zhao; Jiaying Liu; | arxiv-cs.CV | 2024-07-03 |
109 | StyleShot: A Snapshot on Any Style Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we show that, a good style representation is crucial and sufficient for generalized style transfer without test-time tuning. |
JUNYAO GAO et. al. | arxiv-cs.CV | 2024-07-01 |
110 | Towards Diverse Image-to-image Translation Via Adaptive Normalization Layer and Contrast Learning Related Papers Related Patents Related Grants Related Venues Related Experts View |
HENG ZHANG et. al. | Comput. Graph. | 2024-07-01 |
111 | Expanding The Defect Image Dataset of Composite Material Coating with Enhanced Image-to-image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View |
Xinrui Tao; Hanjun Gao; Kai Yang; Qiong Wu; | Eng. Appl. Artif. Intell. | 2024-07-01 |
112 | InstantStyle-Plus: Style Transfer with Content-Preserving in Text-to-Image Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To address these challenges, we deconstruct the style transfer task into three core elements: 1) Style, focusing on the image’s aesthetic characteristics; 2) Spatial Structure, concerning the geometric arrangement and composition of visual elements; and 3) Semantic Content, which captures the conceptual meaning of the image. Guided by these principles, we introduce InstantStyle-Plus, an approach that prioritizes the integrity of the original content while seamlessly integrating the target style. |
HAOFAN WANG et. al. | arxiv-cs.CV | 2024-06-30 |
113 | MUMU: Bootstrapping Multimodal Image Generation from Text-to-Image Data Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We train a model to generate images from multimodal prompts of interleaved text and images such as a |
William Berman; Alexander Peysakhovich; | arxiv-cs.CV | 2024-06-26 |
114 | Conditional Face Image Manipulation Detection: Combining Algorithm and Human Examiner Decisions Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: It has been shown that digitally manipulated face images can pose a security threat to automated authentication systems (e.g., when such systems are used for border control). In … |
M. IBSEN et. al. | Proceedings of the 2024 ACM Workshop on Information Hiding … | 2024-06-24 |
115 | TinyStyler: Efficient Few-Shot Text Style Transfer with Authorship Embeddings Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce TinyStyler, a lightweight but effective approach, which leverages a small language model (800M params) and pre-trained authorship embeddings to perform efficient, few-shot text style transfer. |
ZACHARY HORVITZ et. al. | arxiv-cs.CL | 2024-06-21 |
116 | Style-NeRF2NeRF: 3D Style Transfer From Style-Aligned Multi-View Images Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a simple yet effective pipeline for stylizing a 3D scene, harnessing the power of 2D image diffusion models. |
Haruo Fujiwara; Yusuke Mukuta; Tatsuya Harada; | arxiv-cs.CV | 2024-06-19 |
117 | SDNIA-YOLO: A Robust Object Detection Model for Extreme Weather Conditions Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This study thus proposes a stylization data-driven neural-image-adaptive YOLO (SDNIA-YOLO), which improves the model’s robustness by enhancing image quality adaptively and learning valuable information related to extreme weather conditions from images synthesized by neural style transfer (NST). |
Yuexiong Ding; Xiaowei Luo; | arxiv-cs.CV | 2024-06-18 |
118 | Domain Targeted Synthetic Plant Style Transfer Using Stable Diffusion, LoRA and ControlNet Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Synthetic images can help alleviate much of the cost in the creation of training data for plant phenotyping-focused AI development. Synthetic-to-real style transfer is of … |
Zane K. J. Hartley; Rob J. Lind; Michael P. Pound; Andrew P. French; | 2024 IEEE/CVF Conference on Computer Vision and Pattern … | 2024-06-17 |
119 | Style Transfer with Multi-iteration Preference Optimization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Numerous recent techniques for text style transfer characterize their approaches as variants of reinforcement learning and preference optimization. In this work, we consider the relationship between these approaches and a class of optimization approaches developed primarily for (non-neural) statistical machine translation, formerly known as `tuning’. |
Shuai Liu; Jonathan May; | arxiv-cs.CL | 2024-06-17 |
120 | Style Transfer for 2D Talking Head Generation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Audio-driven talking head animation is a challenging research topic with many real-world applications. Recent works have focused on creating photo-realistic 2D animation, while … |
TRONG-THANG PHAM et. al. | 2024 IEEE/CVF Conference on Computer Vision and Pattern … | 2024-06-17 |
121 | Generative Latent Coding for Ultra-Low Bitrate Image Compression Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However they encounter difficulties in achieving both high-realism and high-fidelity at low bitrate as the pixel-space distortion may not align with human perception. To address this issue we introduce a Generative Latent Coding (GLC) architecture which performs transform coding in the latent space of a generative vector-quantized variational auto-encoder (VQ-VAE) instead of in the pixel space. |
Zhaoyang Jia; Jiahao Li; Bin Li; Houqiang Li; Yan Lu; | cvpr | 2024-06-13 |
122 | DynVideo-E: Harnessing Dynamic NeRF for Large-Scale Motion- and View-Change Human-Centric Video Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To provide consistent and controllable editing we propose the image-based video-NeRF editing pipeline with a set of innovative designs including multi-view multi-pose Score Distillation Sampling (SDS) from both the 2D personalized diffusion prior and 3D diffusion prior reconstruction losses text-guided local parts super-resolution and style transfer. |
JIA-WEI LIU et. al. | cvpr | 2024-06-13 |
123 | 3DToonify: Creating Your High-Fidelity 3D Stylized Avatar Easily from 2D Portrait Images Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we make a connection between the two and tackle the challenging task of 3D portrait stylization – modeling high-fidelity 3D stylized avatars from captured 2D portrait images. |
YIFANG MEN et. al. | cvpr | 2024-06-13 |
124 | Doubly Abductive Counterfactual Inference for Text-based Image Editing Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To this end we propose a Doubly Abductive Counterfactual inference framework (DAC). |
XUE SONG et. al. | cvpr | 2024-06-13 |
125 | Lift3D: Zero-Shot Lifting of Any 2D Vision Model to 3D Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we ask the question of whether any 2D vision model can be lifted to make 3D consistent predictions. |
MUKUND VARMA T et. al. | cvpr | 2024-06-13 |
126 | Puff-Net: Efficient Style Transfer with Pure Content and Style Feature Fusion Network Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: For instance it is difficult for CNN-based methods to handle global information and long-range dependencies between input images for which transformer-based methods have been proposed. |
Sizhe Zheng; Pan Gao; Peng Zhou; Jie Qin; | cvpr | 2024-06-13 |
127 | Style Injection in Diffusion: A Training-free Approach for Adapting Large-scale Diffusion Models for Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Despite the impressive generative capabilities of diffusion models existing diffusion model-based style transfer methods require inference-stage optimization (e.g. fine-tuning or textual inversion of style) which is time-consuming or fails to leverage the generative ability of large-scale diffusion models. To address these issues we introduce a novel artistic style transfer method based on a pre-trained large-scale diffusion model without any optimization. |
Jiwoo Chung; Sangeek Hyun; Jae-Pil Heo; | cvpr | 2024-06-13 |
128 | Z*: Zero-shot Style Transfer Via Attention Reweighting Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We further reveal that the cross-attention mechanism in latent diffusion models tends to blend the content and style images resulting in stylized outputs that deviate from the original content image. To overcome this limitation we introduce a cross-attention reweighting strategy. |
Yingying Deng; Xiangyu He; Fan Tang; Weiming Dong; | cvpr | 2024-06-13 |
129 | S-DyRF: Reference-Based Stylized Radiance Fields for Dynamic Scenes Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Current 3D stylization methods often assume static scenes which violates the dynamic nature of our real world. To address this limitation we present S-DyRF a reference-based spatio-temporal stylization method for dynamic neural radiance fields. |
XINGYI LI et. al. | cvpr | 2024-06-13 |
130 | Video Prediction By Modeling Videos As Continuous Multi-Dimensional Processes IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In our paper we introduce a novel model class that treats video as a continuous multi-dimensional process rather than a series of discrete frames. |
Gaurav Shrivastava; Abhinav Shrivastava; | cvpr | 2024-06-13 |
131 | ArtAdapter: Text-to-Image Style Transfer Using Multi-Level Style Encoder and Explicit Adaptation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work introduces ArtAdapter a transformative text-to-image (T2I) style transfer framework that transcends traditional limitations of color brushstrokes and object shape capturing high-level style elements such as composition and distinctive artistic expression. |
Dar-Yen Chen; Hamish Tennent; Ching-Wen Hsu; | cvpr | 2024-06-13 |
132 | Misalignment-Robust Frequency Distribution Loss for Image Transformation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper aims to address a common challenge in deep learning-based image transformation methods such as image enhancement and super-resolution which heavily rely on precisely aligned paired datasets with pixel-level alignments. |
ZHANGKAI NI et. al. | cvpr | 2024-06-13 |
133 | One-Shot Structure-Aware Stylized Image Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Recently diffusion models have been adopted for image stylization but still lack the capability to maintain the original quality of input images. Building on this we propose OSASIS: a novel one-shot stylization method that is robust in structure preservation. |
Hansam Cho; Jonghyun Lee; Seunggyu Chang; Yonghyun Jeong; | cvpr | 2024-06-13 |
134 | Geometry Transfer for Stylizing Radiance Fields Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we introduce Geometry Transfer a novel method that leverages geometric deformation for 3D style transfer. |
HYUNYOUNG JUNG et. al. | cvpr | 2024-06-13 |
135 | ICE-G: Image Conditional Editing of 3D Gaussian Splats Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce a novel approach to quickly edit a 3D model from a single reference view. |
VISHNU JAGANATHAN et. al. | arxiv-cs.CV | 2024-06-12 |
136 | TLCM: Training-efficient Latent Consistency Model for Image Generation with 2-8 Steps Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: (2) They routinely lead to quality degradation for generation, especially in text-image alignment. This paper proposes a novel training-efficient Latent Consistency Model (TLCM) to overcome these challenges. |
Qingsong Xie; Zhenyi Liao; Zhijie Deng; Chen chen; Haonan Lu; | arxiv-cs.CV | 2024-06-09 |
137 | Varying Manifolds in Diffusion: From Time-varying Geometries to Visual Saliency Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Deep generative models learn the data distribution, which is concentrated on a low-dimensional manifold. The geometric analysis of distribution transformation provides a better … |
Junhao Chen; Manyi Li; Zherong Pan; Xifeng Gao; Changhe Tu; | ArXiv | 2024-06-07 |
138 | Dream-in-Style: Text-to-3D Generation Using Stylized Score Distillation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: We present a method to generate 3D objects in styles. Our method takes a text prompt and a style reference image as input and reconstructs a neural radiance field to synthesize a … |
Hubert Kompanowski; Binh-Son Hua; | ArXiv | 2024-06-05 |
139 | Application of An Improved U-Net with Image-to-image Translation and Transfer Learning in Peach Orchard Segmentation Related Papers Related Patents Related Grants Related Venues Related Experts View |
JIAYU CHENG et. al. | Int. J. Appl. Earth Obs. Geoinformation | 2024-06-01 |
140 | MegActor: Harness The Power of Raw Video for Vivid Portrait Animation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Despite raw driving videos contain richer information on facial expressions than intermediate representations such as landmarks in the field of portrait animation, they are seldom the subject of research. |
SHURONG YANG et. al. | arxiv-cs.CV | 2024-05-31 |
141 | SeamlessExpressiveLM: Speech Language Model for Expressive Speech-to-Speech Translation with Chain-of-Thought Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work proposes SeamlessExpressiveLM, a single speech language model for expressive S2ST. |
Hongyu Gong; Bandhav Veluri; | arxiv-cs.CL | 2024-05-30 |
142 | ExpoGenius: Robust Personalized Human Image Generation Using Diffusion Model for Exposure Variation and Pose Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Diffusion models hold significant appeal within the realm of synthetic media generation and demonstrate exceptional performance in personalized human image generation. However, … |
Depei Liu; Hongjie Fan; Junfei Liu; | Proceedings of the 2024 International Conference on … | 2024-05-30 |
143 | LEAST: Local Text-conditioned Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we evaluate the text-conditioned image editing and style transfer techniques on their fine-grained understanding of user prompts for precise local style transfer. |
Silky Singh; Surgan Jandial; Simra Shahid; Abhinav Java; | arxiv-cs.CV | 2024-05-25 |
144 | Towards Evaluating The Robustness of Automatic Speech Recognition Systems Via Audio Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose an attack on ASR systems based on user-customized style transfer. |
WEIFEI JIN et. al. | arxiv-cs.SD | 2024-05-15 |
145 | Empathy Through Aesthetics: Using AI Stylization for Visual Anonymization of Interview Videos Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Faces are the primary source for identifying individuals, which need to be obstructed to achieve anonymization in images and videos. However, human faces are also one of the most … |
Ö. Yalçın; Vanessa Utz; Steve Dipaola; | Proceedings of the 3rd Empathy-Centric Design Workshop: … | 2024-05-11 |
146 | StyleMamba : State Space Model for Efficient Text-driven Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present StyleMamba, an efficient image style transfer framework that translates text prompts into corresponding visual styles while preserving the content integrity of the original images. |
Zijia Wang; Zhi-Song Liu; | arxiv-cs.CV | 2024-05-08 |
147 | TridentCap: Image-Fact-Style Trident Semantic Framework for Stylized Image Captioning Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Stylized image captioning (SIC) aims to generate captions with target style for images. The biggest challenge is that the collection and annotation of stylized data are pretty … |
LANXIAO WANG et. al. | IEEE Transactions on Circuits and Systems for Video … | 2024-05-01 |
148 | Towards Semantically Continuous Unpaired Image-to-image Translation Via Margin Adaptive Contrastive Learning and Wavelet Transform Related Papers Related Patents Related Grants Related Venues Related Experts View |
Heng Zhang; Yi-Jun Yang; Wei Zeng; | Expert Syst. Appl. | 2024-05-01 |
149 | GAN‐Based Multi‐Decomposition Photo Cartoonization Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Cartoon images play a vital role in film production, scientific and educational animation, video games, and other fields, and are one of the key visual expressions of artistic … |
Wenqing Zhao; Jianlin Zhu; Jin Huang; Ping Li; Bin Sheng; | Computer Animation and Virtual Worlds | 2024-05-01 |
150 | SRAGAN: Saliency Regularized and Attended Generative Adversarial Network for Chinese Ink-wash Painting Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Though a wide range of I2I models tackle this problem, a notable challenge is that the content details of the source image could be easily erased or corrupted due to the transfer of ink-wash style elements. To remedy this issue, we propose to incorporate saliency detection into the unpaired I2I framework to regularize image content, where the detected saliency map is utilized from two aspects: (\romannumeral1) we propose saliency IOU (SIOU) loss to explicitly regularize object content structure by enforcing saliency consistency before and after image stylization; (\romannumeral2) we propose saliency adaptive normalization (SANorm) which implicitly enhances object structure integrity of the generated paintings by dynamically injecting image saliency information into the generator to guide stylization process. |
Xiang Gao; Yuqi Zhang; | arxiv-cs.CV | 2024-04-24 |
151 | CoARF: Controllable 3D Artistic Style Transfer for Radiance Fields Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce Controllable Artistic Radiance Fields (CoARF), a novel algorithm for controllable 3D scene stylization. |
Deheng Zhang; Clara Fernandez-Labrador; Christopher Schroers; | arxiv-cs.CV | 2024-04-23 |
152 | Music Style Transfer With Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The existing music style transfer methods generate spectrograms with artifacts, leading to significant noise in the generated audio. To address these issues, this study proposes a music style transfer framework based on diffusion models (DM) and uses spectrogram-based methods to achieve multi-to-multi music style transfer. |
Hong Huang; Yuyi Wang; Luyao Li; Jun Lin; | arxiv-cs.SD | 2024-04-23 |
153 | Regional Style and Color Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Existing methods often suffer from the drawback of applying style homogeneously across the entire image, leading to stylistic inconsistencies or foreground object twisted when applied to image with foreground elements such as person figures. To address this limitation, we propose a new approach that leverages a segmentation network to precisely isolate foreground objects within the input image. |
Zhicheng Ding; Panfeng Li; Qikai Yang; Siyang Li; Qingtian Gong; | arxiv-cs.CV | 2024-04-22 |
154 | Rethink Arbitrary Style Transfer with Transformer and Contrastive Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce an innovative technique to improve the quality of stylized images. |
ZHANJIE ZHANG et. al. | arxiv-cs.CV | 2024-04-21 |
155 | Towards Highly Realistic Artistic Style Transfer Via Stable Diffusion with Step-aware and Layer-aware Prompt Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, diffusion model-based methods generally fail to preserve the content structure of input content images well, introducing some undesired content structure and style patterns. To address the above problems, we propose a novel pre-trained diffusion-based artistic style transfer method, called LSAST, which can generate highly realistic artistic stylized images while preserving the content structure of input content images well, without bringing obvious artifacts and disharmonious style patterns. |
ZHANJIE ZHANG et. al. | arxiv-cs.CV | 2024-04-17 |
156 | Improved Object-Based Style Transfer with Single Deep Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This research paper proposes a novel methodology for image-to-image style transfer on objects utilizing a single deep convolutional neural network. |
Harshmohan Kulkarni; Om Khare; Ninad Barve; Sunil Mane; | arxiv-cs.CV | 2024-04-15 |
157 | Ctrl-Adapter: An Efficient and Versatile Framework for Adapting Diverse Controls to Any Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Furthermore, applying ControlNets independently to different frames cannot effectively maintain object temporal consistency. To address these challenges, we introduce Ctrl-Adapter, an efficient and versatile framework that adds diverse controls to any image/video diffusion model through the adaptation of pretrained ControlNets. |
Han Lin; Jaemin Cho; Abhay Zala; Mohit Bansal; | arxiv-cs.CV | 2024-04-15 |
158 | Arbitrary Style Transfer Based on Content Integrity and Style Consistency Enhancement Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The existing arbitrary style transfer methods mainly suffer two challenges. One is content integrity, as most methods focus too much on style, resulting in incomplete content … |
Lu Kang; Guoqiang Xiao; Michael S. Lew; Song Wu; | ICASSP 2024 – 2024 IEEE International Conference on … | 2024-04-14 |
159 | Tuning-Free Adaptive Style Incorporation for Structure-Consistent Text-Driven Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we target the task of text-driven style transfer in the context of text-to-image (T2I) diffusion models. |
YANQI GE et. al. | arxiv-cs.CV | 2024-04-10 |
160 | StylizedGS: Controllable Stylization for 3D Gaussian Splatting IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Additionally, the ability for artists to apply flexible control over stylized scenes is considered highly desirable to foster an environment conducive to creative exploration. To address the above issues, we introduce StylizedGS, an efficient 3D neural style transfer framework with adaptable control over perceptual factors based on 3D Gaussian Splatting (3DGS) representation. |
DINGXI ZHANG et. al. | arxiv-cs.CV | 2024-04-08 |
161 | Stylizing Sparse-View 3D Scenes with Hierarchical Neural Representation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we consider the stylization of sparse-view scenes in terms of disentangling content semantics and style textures. |
Y. Wang; A. Gao; Y. Gong; Y. Zeng; | arxiv-cs.CV | 2024-04-08 |
162 | RoNet: Rotation-oriented Continuous Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel rotation-oriented solution and model the continuous generation with an in-plane rotation over the style representation of an image, achieving a network named RoNet. |
Yi Li; Xin Xie; Lina Lei; Haiyan Fu; Yanqing Guo; | arxiv-cs.CV | 2024-04-05 |
163 | Multi-Domain Image-to-Image Translation with Cross-Granularity Contrastive Learning Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The objective of multi-domain image-to-image translation is to learn the mapping from a source domain to a target domain in multiple image domains while preserving the content … |
Huiyuan Fu; Jin Liu; Tingyi Yu; Xin Wang; Huadong Ma; | ACM Transactions on Multimedia Computing, Communications … | 2024-04-04 |
164 | Pixel Art Character Generation As An Image-to-image Translation Problem Using GANs Related Papers Related Patents Related Grants Related Venues Related Experts View |
Flávio Coutinho; Luiz Chaimowicz; | Graph. Model. | 2024-04-01 |
165 | Grid Diffusion Models for Text-to-Video Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: These methods require large datasets and are limited in terms of computational costs compared to text-to-image generation. To tackle these challenges, we propose a simple but effective novel grid diffusion for text-to-video generation without temporal dimension in architecture and a large text-video paired dataset. |
Taegyeong Lee; Soyeong Kwon; Taehwan Kim; | arxiv-cs.CV | 2024-03-29 |
166 | DiffStyler: Diffusion-based Localized Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Recent developments in large-scale text-to-image diffusion models have heralded unprecedented synthesis capabilities, albeit at the expense of relying on extensive and often imprecise textual descriptions to delineate artistic styles. Addressing these limitations, this paper introduces DiffStyler, a novel approach that facilitates efficient and precise arbitrary image style transfer. |
Shaoxu Li; | arxiv-cs.CV | 2024-03-27 |
167 | Enhancing Cross-Dataset EEG Emotion Recognition: A Novel Approach with Emotional EEG Style Transfer Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To tackle the significant inter-domain differences in cross-dataset EEG emotion recognition, this paper introduces an innovative solution known as the Emotional EEG Style Transfer Network (E$^2$STN). |
YIJIN ZHOU et. al. | arxiv-cs.HC | 2024-03-25 |
168 | AnyV2V: A Tuning-Free Framework For Any Video-to-Video Editing Tasks Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Furthermore, these methods frequently rely on textual input as the editing guidance, leading to ambiguities and limiting the types of edits they can perform. Recognizing these challenges, we introduce AnyV2V, a novel tuning-free paradigm designed to simplify video editing into two primary steps: (1) employing an off-the-shelf image editing model to modify the first frame, (2) utilizing an existing image-to-video generation model to generate the edited video through temporal feature injection. |
Max Ku; Cong Wei; Weiming Ren; Harry Yang; Wenhu Chen; | arxiv-cs.CV | 2024-03-21 |
169 | Implicit Style-Content Separation Using B-LoRA IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce B-LoRA, a method that leverages LoRA (Low-Rank Adaptation) to implicitly separate the style and content components of a single image, facilitating various image stylization tasks. |
Yarden Frenkel; Yael Vinker; Ariel Shamir; Daniel Cohen-Or; | arxiv-cs.CV | 2024-03-21 |
170 | Diffusion Attack: Leveraging Stable Diffusion for Naturalistic Image Attacking Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, attackers rarely impose limitations on the naturalness and comfort of the appearance of the generated attack image, resulting in a noticeable and unnatural attack. To address this challenge, we propose a framework to incorporate style transfer to craft adversarial inputs of natural styles that exhibit minimal detectability and maximum natural appearance, while maintaining superior attack capabilities. |
Qianyu Guo; Jiaming Fu; Yawen Lu; Dongming Gan; | arxiv-cs.CV | 2024-03-21 |
171 | Diffusion-based Human Motion Style Transfer with Semantic Guidance Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, we may encounter a single unseen style example in practical scenarios, but not in sufficient quantity to constitute a style cluster for AdaIN-based methods. Therefore, in this paper, we propose a novel two-stage framework for few-shot style transfer learning based on the diffusion model. |
Lei Hu; Zihao Zhang; Yongjing Ye; Yiwen Xu; Shihong Xia; | arxiv-cs.GR | 2024-03-20 |
172 | Enhancing Fingerprint Image Synthesis with GANs, Diffusion Models, and Style Transfer Techniques Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present novel approaches involving generative adversarial networks and diffusion models in order to synthesize high quality, live and spoof fingerprint images while preserving features such as uniqueness and diversity. |
W. Tang; D. Figueroa; D. Liu; K. Johnsson; A. Sopasakis; | arxiv-cs.CV | 2024-03-20 |
173 | Generalized Consistency Trajectory Models for Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This work aims to unlock the full potential of CTMs by proposing generalized CTMs (GCTMs), which translate between arbitrary distributions via ODEs. |
Beomsu Kim; Jaemin Kim; Jeongsol Kim; Jong Chul Ye; | arxiv-cs.CV | 2024-03-19 |
174 | LocalStyleFool: Regional Video Style Transfer Attack Using Segment Anything Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To close the gap, we propose LocalStyleFool, an improved black-box video adversarial attack that superimposes regional style-transfer-based perturbations on videos. |
YUXIN CAO et. al. | arxiv-cs.CV | 2024-03-18 |
175 | Attack Deterministic Conditional Image Generative Models for Diverse and Controllable Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Given that many deterministic conditional image generative models have been able to produce high-quality yet fixed results, we raise an intriguing question: is it possible for pre-trained deterministic conditional image generative models to generate diverse results without changing network structures or parameters? To answer this question, we re-examine the conditional image generation tasks from the perspective of adversarial attack and propose a simple and efficient plug-in projected gradient descent (PGD) like method for diverse and controllable image generation. |
TIANYI CHU et. al. | arxiv-cs.CV | 2024-03-13 |
176 | Gaussian Splatting in Style Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In contrast, we propose a novel architecture trained on a collection of style images that, at test time, produces real time high-quality stylized novel views. |
ABHISHEK SAROHA et. al. | arxiv-cs.CV | 2024-03-13 |
177 | StyleDyRF: Zero-shot 4D Style Transfer for Dynamic Neural Radiance Fields Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce StyleDyRF, a method that represents the 4D feature space by deforming a canonical feature volume and learns a linear style transformation matrix on the feature volume in a data-driven fashion. |
Hongbin Xu; Weitao Chen; Feng Xiao; Baigui Sun; Wenxiong Kang; | arxiv-cs.CV | 2024-03-13 |
178 | StyleGaussian: Instant 3D Style Transfer with Gaussian Splatting IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce StyleGaussian, a novel 3D style transfer technique that allows instant transfer of any image’s style to a 3D scene at 10 frames per second (fps). |
KUNHAO LIU et. al. | arxiv-cs.CV | 2024-03-12 |
179 | Authorship Style Transfer with Policy Optimization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a simple two-stage tune-and-optimize technique for low-resource textual style transfer. |
Shuai Liu; Shantanu Agarwal; Jonathan May; | arxiv-cs.CL | 2024-03-12 |
180 | Towards Model Extraction Attacks in GAN-Based Image Translation Via Domain Shift Mitigation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Model extraction attacks (MEAs) enable an attacker to replicate the functionality of a victim deep neural network (DNN) model by only querying its API service remotely, posing a … |
DI MI et. al. | arxiv-cs.CR | 2024-03-12 |
181 | 3D-aware Image Generation and Editing with Multi-modal Conditions Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel end-to-end 3D-aware image generation and editing model incorporating multiple types of conditional inputs, including pure noise, text and reference image. |
Bo Li; Yi-ke Li; Zhi-fen He; Bin Liu; Yun-Kun Lai; | arxiv-cs.CV | 2024-03-11 |
182 | A Spatiotemporal Style Transfer Algorithm for Dynamic Visual Stimulus Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Here, we introduce the Spatiotemporal Style Transfer (STST) algorithm, a dynamic visual stimulus generation framework that allows powerful manipulation and synthesis of video stimuli for vision research. |
Antonino Greco; Markus Siegel; | arxiv-cs.CV | 2024-03-07 |
183 | Conditional Image Hiding Network Based on Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View |
FENGHUA ZHANG et. al. | Inf. Sci. | 2024-03-01 |
184 | Foreground and Background Separated Image Style Transfer with A Single Text Condition Related Papers Related Patents Related Grants Related Venues Related Experts View |
Yue Yu; Jianming Wang; Nengli Li; | Image Vis. Comput. | 2024-03-01 |
185 | TOSS: High-quality Text-guided Novel View Synthesis from A Single Image IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present TOSS, which introduces text to the task of novel view synthesis (NVS) from just a single RGB image. |
YUKAI SHI et. al. | iclr | 2024-02-26 |
186 | Image Translation As Diffusion Visual Programmers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce the novel Diffusion Visual Programmer (DVP), a neuro-symbolic image translation framework. |
CHENG HAN et. al. | iclr | 2024-02-26 |
187 | Ground-A-Video: Zero-shot Grounded Video Editing Using Text-to-image Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper introduces a novel grounding-guided video-to-video translation framework called Ground-A-Video for multi-attribute video editing. |
Hyeonho Jeong; Jong Chul Ye; | iclr | 2024-02-26 |
188 | Guiding Instruction-based Image Editing Via Multimodal Large Language Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Multimodal large language models (MLLMs) show promising capabilities in cross-modal understanding and visual-aware response generation via LMs. |
TSU-JUI FU et. al. | iclr | 2024-02-26 |
189 | On The Analysis of GAN-based Image-to-Image Translation with Gaussian Noise Injection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our contributions include connecting $f$-divergence and score matching, unveiling insights into the impact of Gaussian noise on aligning probability distributions, and demonstrating generalized robustness implications. |
CHAOHUA SHI et. al. | iclr | 2024-02-26 |
190 | IRConStyle: Image Restoration Framework Using Contrastive Learning and Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This raises a question: Why does the contrastive learning paradigm not yield satisfactory results in image restoration? In this paper, we conduct in-depth analyses and propose three guidelines to address the above question. |
Dongqi Fan; Xin Zhao; Liang Chang; | arxiv-cs.CV | 2024-02-24 |
191 | Counterfactual Generation with Identifiability Guarantees Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we tackle the domain-varying dependence between the content and the style variables inherent in the counterfactual generation task. |
HANQI YAN et. al. | arxiv-cs.LG | 2024-02-23 |
192 | Music Style Transfer with Time-Varying Inversion of Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper presents a music style transfer approach that effectively captures musical attributes using minimal data. |
SIFEI LI et. al. | arxiv-cs.SD | 2024-02-21 |
193 | S2WAT: Image Style Transfer Via Hierarchical Vision Transformer Using Strips Window Attention IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper introduces Strips Window Attention Transformer (S2WAT), a novel hierarchical vision transformer designed for style transfer. |
Chiyu Zhang; Xiaogang Xu; Lei Wang; Zaiyan Dai; Jun Yang; | aaai | 2024-02-20 |
194 | BARET: Balanced Attention Based Real Image Editing Driven By Target-Text Inversion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Image editing approaches with diffusion models have been rapidly developed, yet their applicability are subject to requirements such as specific editing types (e.g., foreground or background object editing, style transfer), multiple conditions (e.g., mask, sketch, caption), and time consuming fine-tuning of diffusion models. For alleviating these limitations and realizing efficient real image editing, we propose a novel editing technique that only requires an input image and target text for various editing types including non-rigid edits without fine-tuning diffusion model. |
YUMING QIAO et. al. | aaai | 2024-02-20 |
195 | DreamStyler: Paint By Style Inversion with Text-to-Image Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To this end, we introduce DreamStyle, a novel framework designed for artistic image synthesis, proficient in both text-to-image synthesis and style transfer. |
NAMHYUK AHN et. al. | aaai | 2024-02-20 |
196 | SEIT: Structural Enhancement for Unsupervised Image Translation in Frequency Domain Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose an unsupervised image translation method with structural enhancement in frequency domain named SEIT. |
ZHIFENG ZHU et. al. | aaai | 2024-02-20 |
197 | FontDiffuser: One-Shot Font Generation Via Denoising Diffusion with Multi-Scale Content Aggregation and Style Contrastive Learning IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Although existing font generation methods have achieved satisfactory performance, they still struggle with complex characters and large style variations. To address these issues, we propose FontDiffuser, a diffusion-based image-to-image one-shot font generation method, which innovatively models the font imitation task as a noise-to-denoise paradigm. |
ZHENHUA YANG et. al. | aaai | 2024-02-20 |
198 | FedST: Federated Style Transfer Learning for Non-IID Image Segmentation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a novel federated image segmentation method based on style transfer, FedST, by using a denoising diffusion probabilistic model to achieve feature disentanglement and image synthesis of cross-domain image data between multiple clients. |
BOYUAN MA et. al. | aaai | 2024-02-20 |
199 | ChromaFusionNet (CFNet): Natural Fusion of Fine-Grained Color Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Existing methods, including color style transfer and image harmonization, exhibit inconsistencies, especially at boundary regions. Addressing this, we present ChromaFusionNet (CFNet), a novel approach that views the color fusion problem through the lens of image color inpainting. |
YI DONG et. al. | aaai | 2024-02-20 |
200 | HyperEditor: Achieving Both Authenticity and Cross-Domain Capability in Image Editing Via Hypernetworks Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In response, we propose an innovative image editing method called HyperEditor, which utilizes weight factors generated by hypernetworks to reassign the weights of the pre-trained StyleGAN2’s generator. |
Hai Zhang; Chunwei Wu; Guitao Cao; Hailing Wang; Wenming Cao; | aaai | 2024-02-20 |
201 | Learning to Manipulate Artistic Images Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose an arbitrary Style Image Manipulation Network (SIM-Net), which leverages semantic-free information as guidance and a region transportation strategy in a self-supervised manner for image generation. |
Wei Guo; Yuqi Zhang; De Ma; Qian Zheng; | aaai | 2024-02-20 |
202 | FPRF: Feed-Forward Photorealistic Style Transfer of Large-Scale 3D Neural Radiance Fields Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present FPRF, a feed-forward photorealistic style transfer method for large-scale 3D neural radiance fields. |
GeonU Kim; Kim Youwang; Tae-Hyun Oh; | aaai | 2024-02-20 |
203 | Scalable Motion Style Transfer with Constrained Diffusion Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Recent image transfer works show the potential of independent training on each domain by leveraging implicit bridging between diffusion models, with the content preservation, however, limited to simple data patterns. We address this by imposing biased sampling in backward diffusion while maintaining the domain independence in the training stage. |
Wenjie Yin; Yi Yu; Hang Yin; Danica Kragic; Mårten Björkman; | aaai | 2024-02-20 |
204 | ArtBank: Artistic Style Transfer with Pre-trained Diffusion Model and Implicit Style Prompt Bank Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Small model-based approaches can preserve the content strucuture, but fail to produce highly realistic stylized images and introduce artifacts and disharmonious patterns; Pre-trained large-scale model-based approaches can generate highly realistic stylized images but struggle with preserving the content structure. To address the above issues, we propose ArtBank, a novel artistic style transfer framework, to generate highly realistic stylized images while preserving the content structure of the content images. |
ZHANJIE ZHANG et. al. | aaai | 2024-02-20 |
205 | Improving 2D-3D Dense Correspondences with Diffusion Models for 6D Object Pose Estimation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we compare image-to-image translation networks based on GANs and diffusion models for the downstream task of 6D object pose estimation. |
Peter Hönig; Stefan Thalhammer; Markus Vincze; | arxiv-cs.CV | 2024-02-09 |
206 | Application of Multi-level Adaptive Neural Network Based on Optimization Algorithm in Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View |
Hong-an Li; Lanye Wang; Jun Liu; | Multim. Tools Appl. | 2024-02-09 |
207 | Shape-biased Texture Agnostic Representations for Improved Textureless and Metallic Object Detection and 6D Pose Estimation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To address his issue, we propose a strategy for inducing a shape bias to CNN training. |
Peter Hönig; Stefan Thalhammer; Jean-Baptiste Weibel; Matthias Hirschmanner; Markus Vincze; | arxiv-cs.CV | 2024-02-07 |
208 | IGUANe: A 3D Generalizable CycleGAN for Multicenter Harmonization of Brain MR Images Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this study, we introduce IGUANe (Image Generation with Unified Adversarial Networks), an original 3D model that leverages the strengths of domain translation and straightforward application of style transfer methods for multicenter brain MR image harmonization. |
Vincent Roca; Grégory Kuchcinski; Jean-Pierre Pruvo; Dorian Manouvriez; Renaud Lopes; | arxiv-cs.CV | 2024-02-05 |
209 | ToonAging: Face Re-Aging Upon Artistic Portrait Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce a novel one-stage method for face re-aging combined with portrait style transfer, executed in a single generative step. |
Bumsoo Kim; Abdul Muqeet; Kyuchul Lee; Sanghyun Seo; | arxiv-cs.CV | 2024-02-05 |
210 | ConRF: Zero-shot Stylization of 3D Scenes with Conditioned Radiation Fields Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce ConRF, a novel method of zero-shot stylization. |
XINGYU MIAO et. al. | arxiv-cs.CV | 2024-02-02 |
211 | Phrase Grounding-based Style Transfer for Single-Domain Generalized Object Detection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel phrase grounding-based style transfer (PGST) approach for the task. |
HAO LI et. al. | arxiv-cs.CV | 2024-02-02 |
212 | Towards Efficient Image and Video Style Transfer Via Distillation and Learnable Feature Transformation Related Papers Related Patents Related Grants Related Venues Related Experts View |
JING HUO et. al. | Comput. Vis. Image Underst. | 2024-02-01 |
213 | Transferring Human Emotions to Robot Motions Using Neural Policy Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The proposed approach was evaluated for both platforms, performing a total of 147 questionnaires asking human subjects to recognize the human motion style transferred to the robot motion for a predefined set of actions. |
Raul Fernandez-Fernandez; Bartek Łukawski; Juan G. Victores; Claudio Pacchierotti; | arxiv-cs.RO | 2024-02-01 |
214 | Panoptic-Level Image-to-Image Translation for Object Recognition and Visual Odometry Enhancement Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Image-to-image translation methods have progressed from only considering the image-level information to integrating the global- and instance-level information. However, only the … |
LIYUN ZHANG et. al. | IEEE Transactions on Circuits and Systems for Video … | 2024-02-01 |
215 | LATENTPATCH: A Non-Parametric Approach for Face Generation and Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents LatentPatch, a new method for generating realistic images from a small dataset of only a few images. |
Benjamin Samuth; Julien Rabin; David Tschumperlé; Frédéric Jurie; | arxiv-cs.MM | 2024-01-30 |
216 | FreeStyle: Free Lunch for Text-guided Style Transfer Using Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce FreeStyle, an innovative style transfer method built upon a pre-trained large diffusion model, requiring no further optimization. |
FEIHONG HE et. al. | arxiv-cs.CV | 2024-01-28 |
217 | CreativeSynth: Creative Blending and Synthesis of Visual Arts Based on Multimodal Diffusion Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Large-scale text-to-image generative models have made impressive strides, showcasing their ability to synthesize a vast array of high-quality images. |
NISHA HUANG et. al. | arxiv-cs.CV | 2024-01-25 |
218 | BlenDA: Domain Adaptive Object Detection Through Diffusion-based Blending Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To address the large domain gap issue between the source and target domains, we propose a novel regularization method for domain adaptive object detection, BlenDA, by generating the pseudo samples of the intermediate domains and their corresponding soft domain labels for adaptation training. |
Tzuhsuan Huang; Chen-Che Huang; Chung-Hao Ku; Jun-Cheng Chen; | arxiv-cs.CV | 2024-01-18 |
219 | Image Translation As Diffusion Visual Programmers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce the novel Diffusion Visual Programmer (DVP), a neuro-symbolic image translation framework. |
CHENG HAN et. al. | arxiv-cs.CV | 2024-01-18 |
220 | Artwork Protection Against Neural Style Transfer Using Locally Adaptive Adversarial Color Attack Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose Locally Adaptive Adversarial Color Attack (LAACA), empowering artists to protect their artwork from unauthorized style transfer by processing before public release. |
ZHONGLIANG GUO et. al. | arxiv-cs.CV | 2024-01-17 |
221 | Key-point Guided Deformable Image Manipulation Using Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce a Key-point-guided Diffusion probabilistic Model (KDM) that gains precise control over images by manipulating the object’s key-point. |
SEOK-HWAN OH et. al. | arxiv-cs.CV | 2024-01-16 |
222 | CAT-LLM: Prompting Large Language Models with Text Style Definition for Chinese Article-style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, existing research mainly concentrates on style transfer within individual English sentences, while ignoring the complexity of long Chinese texts, which limits the wider applicability of style transfer in digital media realm. To bridge this gap, we propose a Chinese Article-style Transfer framework (CAT-LLM), leveraging the capabilities of Large Language Models (LLMs). |
Zhen Tao; Dinghao Xi; Zhiyu Li; Liumin Tang; Wei Xu; | arxiv-cs.CL | 2024-01-11 |
223 | HiCAST: Highly Customized Arbitrary Style Transfer with Adapter Enhanced Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In the specific, our model is constructed based on Latent Diffusion Model (LDM) and elaborately designed to absorb content and style instance as conditions of LDM. |
HANZHANG WANG et. al. | arxiv-cs.CV | 2024-01-11 |
224 | Let’s Go Shopping (LGS) – Web-Scale Image-Text Dataset for Visual Concept Understanding IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Vision and vision-language applications of neural networks, such as image classification and captioning, rely on large-scale annotated datasets that require non-trivial … |
YATONG BAI et. al. | ArXiv | 2024-01-09 |
225 | Let’s Go Shopping (LGS) — Web-Scale Image-Text Dataset for Visual Concept Understanding Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce the Let’s Go Shopping (LGS) dataset, a large-scale public dataset with 15 million image-caption pairs from publicly available e-commerce websites. |
YATONG BAI et. al. | arxiv-cs.CV | 2024-01-09 |
226 | MirrorDiffusion: Stabilizing Diffusion Process in Zero-shot Image Translation By Prompts Redescription and Beyond Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To make reconstruction explicit, we propose a prompt redescription strategy to realize a mirror effect between the source and reconstructed image in the diffusion model (MirrorDiffusion). |
Yupei Lin; Xiaoyu Xian; Yukai Shi; Liang Lin; | arxiv-cs.CV | 2024-01-06 |
227 | Unsupervised Exemplar-Based Image-to-Image Translation and Cascaded Vision Transformers for Tagged and Untagged Cardiac Cine MRI Registration Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Multi-modal registration between tagged and untagged cardiac cine magnetic resonance (MR) images remains difficult, due to the domain gap and large deformations between the two … |
Meng Ye; Mikael Kanski; Dong Yang; Leon Axel; Dimitris N. Metaxas; | 2024 IEEE/CVF Winter Conference on Applications of Computer … | 2024-01-03 |
228 | Neural Style Protection: Counteracting Unauthorized Neural Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Arbitrary neural style transfer is an advanced AI technique that can effectively synthesize pictures with an artistic style similar to a given source picture. However, if such an … |
Yaxin Li; Jie Ren; Han Xu; Hui Liu; | 2024 IEEE/CVF Winter Conference on Applications of Computer … | 2024-01-03 |
229 | FaceRefiner: High-Fidelity Facial Texture Refinement With Differentiable Rendering-Based Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Recent facial texture generation methods prefer to use deep networks to synthesize image content and then fill in the UV map, thus generating a compelling full texture from a … |
CHENGYANG LI et. al. | IEEE Transactions on Multimedia | 2024-01-01 |
230 | Open-Set: ID Card Presentation Attack Detection Using Neural Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The accurate detection of ID card Presentation Attacks (PA) is becoming increasingly important due to the rising number of online/remote services that require the presentation of … |
Reuben P. Markham; Juan M. Espín López; Mario Nieto-Hidalgo; Juan E. Tapia; | IEEE Access | 2024-01-01 |
231 | Fine-Grained Human Hair Segmentation Using A Text-to-Image Diffusion Model Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Human hair segmentation is essential for face recognition and for achieving natural transformation of style transfer. However, it remains a challenging task due to the diverse … |
Dohyun Kim; Euna Lee; Daehyun Yoo; Hongchul Lee; | IEEE Access | 2024-01-01 |
232 | Multi-level Patch Transformer for Style Transfer with Single Reference Image Related Papers Related Patents Related Grants Related Venues Related Experts View |
Yue He; Lan Chen; Yu-Jie Yuan; Shu-Yu Chen; Lin Gao; | International Conference on Computational Visual Media | 2024-01-01 |
233 | Self-Supervised Learning Guided By SAR Image Factors for Terrain Classification Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Effective feature representation is the key to synthetic aperture radar (SAR) image terrain classification. Limited by the abstract appearance and the scarcity of high-quality … |
ZHONGLE REN et. al. | IEEE Transactions on Geoscience and Remote Sensing | 2024-01-01 |
234 | GCSANet: Arbitrary Style Transfer With Global Context Self-Attentional Network Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Arbitrary style transfer is attracting increasing attention in the computer vision community due to its application flexibility. Existing approaches directly fuse deep style … |
Zhongyu Bai; Hongli Xu; X. Zhang; Qichuan Ding; | IEEE Transactions on Multimedia | 2024-01-01 |
235 | Ultrasound Despeckling With GANs and Cross Modality Transfer Learning Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Ultrasound images are corrupted by a type of signal-dependent noise, called speckle, difficult to remove or attenuate with the classical denoising methods. On the contrary, … |
DIOGO FRÓIS VIEIRA et. al. | IEEE Access | 2024-01-01 |
236 | AdvST: Generating Unrestricted Adversarial Images Via Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Recent years have witnessed extensive applications of Deep Neural Networks (DNNs) in various vision tasks. However, DNNs are vulnerable to adversarial images crafted by … |
XIAOMENG WANG et. al. | IEEE Transactions on Multimedia | 2024-01-01 |
237 | A Generative Adversarial Network AMS-CycleGAN for Multi-Style Image Transformation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The objective of image style transfer is to create an image that has the artistic features of a reference style image while also retaining the details of the original content … |
XIAODI RANG et. al. | IEEE Access | 2024-01-01 |
238 | Multi-Source Style Transfer Via Style Disentanglement Network Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Despite the great success of deep neural networks for style transfer tasks, the entanglement of content and style in images leads to more style information not being captured. To … |
Quan Wang; Sheng Li; Zichi Wang; Xinpeng Zhang; Guorui Feng; | IEEE Transactions on Multimedia | 2024-01-01 |
239 | UnlearnCanvas: A Stylized Image Dataset to Benchmark Machine Unlearning for Diffusion Models IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The rapid advancement of diffusion models (DMs) has not only transformed various real-world industries but has also introduced negative societal concerns, including the generation … |
YIHUA ZHANG et. al. | ArXiv | 2024-01-01 |
240 | Self-Supervised Underwater Image Generation for Underwater Domain Pre-Training Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The rapid progress in computer vision has presented new opportunities for enhancing the visual capabilities of underwater robots. However, most deep learning-based visual … |
Zhi-zong Wu; Zhengxing Wu; Xingyu Chen; Yue Lu; Junzhi Yu; | IEEE Transactions on Instrumentation and Measurement | 2024-01-01 |
241 | Comparison of Deep Learning Image-to-image Models for Medical Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View |
Zeyu Yang; Frank G. Zöllner; | Bildverarbeitung für die Medizin | 2024-01-01 |
242 | Disrupting Anti-Spoofing Systems By Images of Consistent Identity Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Face anti-spoofing aims to distinguish between live and spoof images to ensure the authenticity and reliability of face recognition. Methods based on convolutional neural networks … |
Feng Ding; Zihan Jiang; Yue Zhou; Jianfeng Xu; Guopu Zhu; | IEEE Signal Processing Letters | 2024-01-01 |
243 | DR-AVIT: Toward Diverse and Realistic Aerial Visible-to-Infrared Image Translation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Image-to-image (I2I) translation methods based on generative adversarial networks (GANs) have shown general solutions for aerial visible-to-infrared image translation (AVIT) task. … |
Zonghao Han; Shun Zhang; Yuru Su; Xiaoning Chen; Shaohui Mei; | IEEE Transactions on Geoscience and Remote Sensing | 2024-01-01 |
244 | Side-Scan Sonar Image Classification With Zero-Shot and Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Side-scan sonar (SSS) has become an important tool for ocean exploration due to its practicality and reliability. Existing approaches for SSS image classification mainly rely on … |
Zhongyu Bai; Hongli Xu; Qichuan Ding; Xiangyue Zhang; | IEEE Transactions on Instrumentation and Measurement | 2024-01-01 |
245 | Progressive Fourier Adversarial Domain Adaptation for Object Classification and Retrieval Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Domain adaptation has been extensively explored as a means of transferring knowledge from the labeled source domain to the unlabeled target domain with disparate data … |
TIANBAO LI et. al. | IEEE Transactions on Multimedia | 2024-01-01 |
246 | Visible-to-Infrared Image Translation for Matching Tasks Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Visible-to-infrared image translation is an important way to enrich infrared data. However, the reliability of the data generated by image translation in downstream tasks has … |
Decao Ma; Shaopeng Li; Juan Su; Yong Xian; Tao Zhang; | IEEE Journal of Selected Topics in Applied Earth … | 2024-01-01 |
247 | PhotoStyle60: A Photographic Style Dataset for Photo Authorship Attribution and Photographic Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Photography, like painting, allows artists to express themselves through their unique style. In digital photography, this is achieved not only with the choice of the subject and … |
Marco Cotogni; Marco Arazzi; Claudio Cusano; | IEEE Transactions on Multimedia | 2024-01-01 |
248 | RainSD: Rain Style Diversification Module for Image Synthesis Enhancement Using Feature-Level Style Distribution Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, a synthetic road dataset with sensor blockage generated from real road dataset BDD100K is suggested in the format of BDD100K annotation. |
HYEONJAE JEON et. al. | arxiv-cs.CV | 2023-12-31 |
249 | RAST: Restorable Arbitrary Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: The objective of arbitrary style transfer is to apply a given artistic or photo-realistic style to a target image. Although current methods have shown some success in transferring … |
Yingnan Ma; Chenqiu Zhao; Bingran Huang; Xudong Li; Anup Basu; | ACM Transactions on Multimedia Computing, Communications … | 2023-12-30 |
250 | Text Fact Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Text style transfer is a prominent task that aims to control the style of text without inherently changing its factual content. To cover more text modification applications, such as adapting past news for current events and repurposing educational materials, we propose the task of text fact transfer, which seeks to transfer the factual content of a source text between topics without modifying its style. |
Nishant Balepur; Jie Huang; Kevin Chang; | emnlp | 2023-12-22 |
251 | Open-Set: ID Card Presentation Attack Detection Using Neural Transfer Style Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work explores ID card Presentation Attack Instruments (PAI) in order to improve the generation of samples with four Generative Adversarial Networks (GANs) based image translation models and analyses the effectiveness of the generated data for training fraud detection systems. |
Reuben Markham; Juan M. Espin; Mario Nieto-Hidalgo; Juan E. Tapia; | arxiv-cs.CV | 2023-12-21 |
252 | DETER: Detecting Edited Regions for Deterring Generative Manipulations Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To counteract the shortcomings, we introduce DETER, a large-scale dataset for DETEcting edited image Regions and deterring modern advanced generative manipulations. |
SAI WANG et. al. | arxiv-cs.CV | 2023-12-16 |
253 | LogoStyleFool: Vitiating Video Recognition Systems Via Logo Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we focus on the video black-box setting and propose a novel attack framework named LogoStyleFool by adding a stylized logo to the clean video. |
YUXIN CAO et. al. | arxiv-cs.CV | 2023-12-15 |
254 | Towards Better Morphed Face Images Without Ghosting Artifacts Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a method for automatic prevention of ghosting artifacts based on a pixel-wise alignment during morph generation. |
Clemens Seibold; Anna Hilsmann; Peter Eisert; | arxiv-cs.CV | 2023-12-13 |
255 | Scalable Motion Style Transfer with Constrained Diffusion Generation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Recent image transfer works show the potential of independent training on each domain by leveraging implicit bridging between diffusion models, with the content preservation, however, limited to simple data patterns. We address this by imposing biased sampling in backward diffusion while maintaining the domain independence in the training stage. |
Wenjie Yin; Yi Yu; Hang Yin; Danica Kragic; Mårten Björkman; | arxiv-cs.CV | 2023-12-12 |
256 | Diffusion Cocktail: Mixing Domain-Specific Diffusion Models for Diversified Image Generations Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose Diffusion Cocktail (Ditail), a training-free method that transfers style and content information between multiple diffusion models. |
Haoming Liu; Yuanhe Guo; Shengjie Wang; Hongyi Wen; | arxiv-cs.CV | 2023-12-11 |
257 | ArtBank: Artistic Style Transfer with Pre-trained Diffusion Model and Implicit Style Prompt Bank Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Small model-based approaches can preserve the content strucuture, but fail to produce highly realistic stylized images and introduce artifacts and disharmonious patterns; Pre-trained large-scale model-based approaches can generate highly realistic stylized images but struggle with preserving the content structure. To address the above issues, we propose ArtBank, a novel artistic style transfer framework, to generate highly realistic stylized images while preserving the content structure of the content images. |
ZHANJIE ZHANG et. al. | arxiv-cs.CV | 2023-12-11 |
258 | Neutral Editing Framework for Diffusion-based Video Editing Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Text-conditioned image editing has succeeded in various types of editing based on a diffusion framework. Unfortunately, this success did not carry over to a video, which continues … |
Sunjae Yoon; Gwanhyeong Koo; Jiajing Hong; Changdong Yoo; | ArXiv | 2023-12-10 |
259 | AesFA: An Aesthetic Feature-Aware Arbitrary Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This work proposes a lightweight but effective model, AesFA — Aesthetic Feature-Aware NST. |
Joonwoo Kwon; Sooyoung Kim; Yuewei Lin; Shinjae Yoo; Jiook Cha; | arxiv-cs.CV | 2023-12-10 |
260 | Anything to Glyph: Artistic Font Synthesis Via Text-to-Image Diffusion Model IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The automatic generation of artistic fonts is a challenging task that attracts many research interests. Previous methods specifically focus on glyph or texture style transfer. … |
CHANGSHUO WANG et. al. | SIGGRAPH Asia 2023 Conference Papers | 2023-12-10 |
261 | BARET : Balanced Attention Based Real Image Editing Driven By Target-text Inversion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Image editing approaches with diffusion models have been rapidly developed, yet their applicability are subject to requirements such as specific editing types (e.g., foreground or background object editing, style transfer), multiple conditions (e.g., mask, sketch, caption), and time consuming fine-tuning of diffusion models. For alleviating these limitations and realizing efficient real image editing, we propose a novel editing technique that only requires an input image and target text for various editing types including non-rigid edits without fine-tuning diffusion model. |
YUMING QIAO et. al. | arxiv-cs.CV | 2023-12-09 |
262 | MuVieCAST: Multi-View Consistent Artistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce MuVieCAST, a modular multi-view consistent style transfer network architecture that enables consistent style transfer between multiple viewpoints of the same scene. |
Nail Ibrahimli; Julian F. P. Kooij; Liangliang Nan; | arxiv-cs.CV | 2023-12-08 |
263 | Reimagining Animation Making Through Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Dissolution” and “Cunabula” are experimental animation projects that explore the potential of artistic style transfer with neural networks in animation filmmaking. The objective … |
Sujin Kim; | SIGGRAPH Asia 2023 Art Papers | 2023-12-08 |
264 | Towards 4D Human Video Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present a first step towards 4D (3D and time) human video stylization, which addresses style transfer, novel view synthesis and human animation within a unified framework. |
Tiantian Wang; Xinxin Zuo; Fangzhou Mu; Jian Wang; Ming-Hsuan Yang; | arxiv-cs.CV | 2023-12-07 |
265 | LEGO: Learning EGOcentric Action Frame Generation Via Visual Instruction Tuning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we introduce a novel problem — egocentric action frame generation. |
BOLIN LAI et. al. | arxiv-cs.CV | 2023-12-06 |
266 | Geometric Style Transfer for Face Portraits Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Geometric style transfer jointly stylizes the texture and geometry of a content image to better match a style image, which has attracted widespread attention due to its various … |
Miaomiao Dai; Hao Yin; Ran Yi; Lizhuang Ma; | Proceedings of the 5th ACM International Conference on … | 2023-12-06 |
267 | Style Transfer to Calvin and Hobbes Comics Using Stable Diffusion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This project report summarizes our journey to perform stable diffusion fine-tuning on a dataset containing Calvin and Hobbes comics. |
Sloke Shrestha; Sundar Sripada V. S.; Asvin Venkataramanan; | arxiv-cs.CV | 2023-12-06 |
268 | Multimodality-guided Image Style Transfer Using Cross-modal GAN Inversion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Unfortunately, many TIST approaches produce undesirable artifacts in the transferred images. To address this issue, we present a novel method to achieve much improved style transfer based on text guidance. |
Hanyu Wang; Pengxiang Wu; Kevin Dela Rosa; Chen Wang; Abhinav Shrivastava; | arxiv-cs.CV | 2023-12-04 |
269 | MMFusion: Combining Image Forensic Filters for Visual Manipulation Detection and Localization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Recent image manipulation localization and detection techniques typically leverage forensic artifacts and traces that are produced by a noise-sensitive filter, such as SRM or Bayar convolution. In this paper, we showcase that different filters commonly used in such approaches excel at unveiling different types of manipulations and provide complementary forensic traces. |
Kostas Triaridis; Konstantinos Tsigos; Vasileios Mezaris; | arxiv-cs.CV | 2023-12-04 |
270 | SASSL: Enhancing Self-Supervised Learning Via Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This results in distorted augmented samples with compromised semantic information, ultimately impacting downstream performance. To overcome this limitation, we propose SASSL: Style Augmentations for Self Supervised Learning, a novel data augmentation technique based on Neural Style Transfer. |
RENAN A. ROJAS-GOMEZ et. al. | arxiv-cs.CV | 2023-12-02 |
271 | Beyond Entropy: Style Transfer Guided Single Image Continual Test-Time Adaptation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In detail, when using a single image, the instability caused by batch normalization layers and entropy loss significantly destabilizes many existing methods in real-world cTTA scenarios. To overcome these challenges, we present BESTTA, a novel single image continual test-time adaptation method guided by style transfer, which enables stable and efficient adaptation to the target environment by transferring the style of the input image to the source style. |
Younggeol Cho; Youngrae Kim; Dongman Lee; | arxiv-cs.CV | 2023-11-30 |
272 | InstaStyle: Inversion Noise of A Stylized Image Is Secretly A Style Adviser Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose InstaStyle, a novel approach that excels in generating high-fidelity stylized images with only a single reference image. |
XING CUI et. al. | arxiv-cs.CV | 2023-11-25 |
273 | Z*: Zero-shot Style Transfer Via Attention Rearrangement Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Despite the remarkable progress in image style transfer, formulating style in the context of art is inherently subjective and challenging. In contrast to existing learning/tuning … |
Yingying Deng; Xiangyu He; Fan Tang; Weiming Dong; | ArXiv | 2023-11-25 |
274 | Neural Style Transfer for Computer Games Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Here, we present an approach for injecting depth-aware NST as part of the 3D rendering pipeline. |
Eleftherios Ioannou; Steve Maddock; | arxiv-cs.CV | 2023-11-24 |
275 | Highly Detailed and Temporal Consistent Video Stylization Via Synchronized Multi-Frame Diffusion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a synchronized multi-frame diffusion framework to maintain both the visual details and the temporal consistency. |
Minshan Xie; Hanyuan Liu; Chengze Li; Tien-Tsin Wong; | arxiv-cs.CV | 2023-11-24 |
276 | FreePIH: Training-Free Painterly Image Harmonization with Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Unlike existing methods that require either training auxiliary networks or fine-tuning a large pre-trained backbone, or both, to harmonize a foreground object with a painterly-style background image, our FreePIH tames the denoising process as a plug-in module for foreground image style transfer. Specifically, we find that the very last few steps of the denoising (i.e., generation) process strongly correspond to the stylistic information of images, and based on this, we propose to augment the latent features of both the foreground and background images with Gaussians for a direct denoising-based harmonization. |
Ruibin Li; Jingcai Guo; Song Guo; Qihua Zhou; Jie Zhang; | arxiv-cs.CV | 2023-11-24 |
277 | A New Benchmark and Model for Challenging Image Manipulation Detection Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To investigate the State-of-The-Art (SoTA) IMD methods in those challenging conditions, we introduce a new Challenging Image Manipulation Detection (CIMD) benchmark dataset, which consists of two subsets, for evaluating editing-based and compression-based IMD methods, respectively. |
Zhenfei Zhang; Mingyang Li; Ming-Ching Chang; | arxiv-cs.CV | 2023-11-23 |
278 | 3D Face Style Transfer with A Hybrid Solution of NeRF and Mesh Rasterization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we tackle the problem of 3D face style transfer which aims at generating stylized novel views of a 3D human face with multi-view consistency. |
Jianwei Feng; Prateek Singhal; | arxiv-cs.CV | 2023-11-22 |
279 | Soulstyler: Using Large Language Model to Guide Image Style Transfer for Target Object Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, most current methods require reference to stylized images and cannot individually stylize specific objects. To overcome this limitation, we propose the Soulstyler framework, which allows users to guide the stylization of specific objects in an image through simple textual descriptions. |
JUNHAO CHEN et. al. | arxiv-cs.CV | 2023-11-22 |
280 | Hairstyle-and-identity-aware Facial Image Style Transfer with Region-guiding Masks Related Papers Related Patents Related Grants Related Venues Related Experts View |
Hsin-Ying Wang; Chiu-Wei Chien; Ming-Han Tsai; I-Chen Lin; | Multim. Tools Appl. | 2023-11-15 |
281 | FastBlend: A Powerful Model-Free Toolkit Making Video Stylization Easier Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: With the emergence of diffusion models and rapid development in image processing, it has become effortless to generate fancy images in tasks such as style transfer and image … |
ZHONGJIE DUAN et. al. | ArXiv | 2023-11-15 |
282 | PSST: A Benchmark for Evaluation-driven Text Public-Speaking Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, previous text style transfer primarily focused on sentence-level data-driven approaches, limiting exploration of potential problems in large language models (LLMs) and the ability to meet complex application needs. To overcome these limitations, we introduce a novel task called Public-Speaking Style Transfer (PSST), which aims to simulate humans to transform passage-level, official texts into a public-speaking style. |
HUASHAN SUN et. al. | arxiv-cs.CL | 2023-11-14 |
283 | STEER: Unified Style Transfer with Expert Reinforcement Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we focus on arbitrary style transfer: rewriting a text from an arbitrary, unknown style to a target style. |
SKYLER HALLINAN et. al. | arxiv-cs.CL | 2023-11-13 |
284 | ControlStyle: Text-Driven Stylized Image Generation Using Diffusion Priors IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a new task for “stylizing” text-to-image models, namely text-driven stylized image generation, that further enhances editability in content creation. |
Jingwen Chen; Yingwei Pan; Ting Yao; Tao Mei; | arxiv-cs.CV | 2023-11-09 |
285 | SCONE-GAN: Semantic Contrastive Learning-based Generative Adversarial Network for An End-to-end Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: That is because these methods learn more frequent associations rather than the scene structures. To mitigate the problem, we propose SCONE-GAN that utilises graph convolutional networks to learn the objects dependencies, maintain the image structure and preserve its semantics while transferring images into the target domain. |
IMAN ABBASNEJAD et. al. | arxiv-cs.CV | 2023-11-07 |
286 | PRISM: Progressive Restoration for Scene Graph-based Image Manipulation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Scene graphs have emerged as accurate descriptive priors for image generation and manipulation tasks, however, their complexity and diversity of the shapes and relations of objects in data make it challenging to incorporate them into the models and generate high-quality results. To address these challenges, we propose PRISM, a novel progressive multi-head image manipulation approach to improve the accuracy and quality of the manipulated regions in the scene. |
Pavel Jahoda; Azade Farshad; Yousef Yeganeh; Ehsan Adeli; Nassir Navab; | arxiv-cs.LG | 2023-11-03 |
287 | Optimal Image Transport on Sparse Dictionaries Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we derive a novel optimal image transport algorithm over sparse dictionaries by taking advantage of Sparse Representation (SR) and Optimal Transport (OT). |
Junqing Huang; Haihui Wang; Andreas Weiermann; Michael Ruzhansky; | arxiv-cs.CV | 2023-11-03 |
288 | Expanding Expressiveness of Diffusion Models with Limited Data Via Self-Distillation Based Fine-Tuning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Training diffusion models on limited datasets poses challenges in terms of limited generation capacity and expressiveness, leading to unsatisfactory results in various downstream tasks utilizing pretrained diffusion models, such as domain translation and text-guided image manipulation. In this paper, we propose Self-Distillation for Fine-Tuning diffusion models (SDFT), a methodology to address these challenges by leveraging diverse features from diffusion models pretrained on large source datasets. |
Jiwan Hur; Jaehyun Choi; Gyojin Han; Dong-Jae Lee; Junmo Kim; | arxiv-cs.CV | 2023-11-02 |
289 | Novel View Synthesis from A Single RGBD Image for Indoor Scenes Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose an approach for synthesizing novel view images from a single RGBD (Red Green Blue-Depth) input. |
Congrui Hetang; Yuping Wang; | arxiv-cs.CV | 2023-11-02 |
290 | CFA-GAN: Cross Fusion Attention and Frequency Loss for Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View |
XIANGTIAN ZHENG et. al. | Displays | 2023-11-01 |
291 | InstaFormer++: Multi-Domain Instance-Aware Image-to-Image Translation with Transformer Related Papers Related Patents Related Grants Related Venues Related Experts View |
SOOHYUN KIM et. al. | Int. J. Comput. Vis. | 2023-10-31 |
292 | Histopathological Image Analysis with Style-Augmented Feature Domain Mixing for Improved Generalization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this study, we propose a feature domain style mixing technique that uses adaptive instance normalization to generate style-augmented versions of images. |
Vaibhav Khamankar; Sutanu Bera; Saumik Bhattacharya; Debashis Sen; Prabir Kumar Biswas; | arxiv-cs.CV | 2023-10-31 |
293 | An Implementation of Multimodal Fusion System for Intelligent Digital Human Generation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, an implementation scheme of an intelligent digital human generation system with multimodal fusion is proposed. |
Yingjie Zhou; Yaodong Chen; Kaiyue Bi; Lian Xiong; Hui Liu; | arxiv-cs.MM | 2023-10-31 |
294 | Generative AI Model for Artistic Style Transfer Using Convolutional Neural Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a comprehensive overview of a novel technique for style transfer using Convolutional Neural Networks (CNNs). |
Jonayet Miah; Duc M Cao; Md Abu Sayed; Md. Sabbirul Haque; | arxiv-cs.CV | 2023-10-27 |
295 | Rethinking Neural Style Transfer: Generating Personalized and Watermarked Stylized Images Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Neural style transfer (NST) has attracted many research interests recent years. The existing NST schemes could only generate one stylized image from a content-style image pair. … |
Quan Wang; Sheng Li; Xinpeng Zhang; Guorui Feng; | Proceedings of the 31st ACM International Conference on … | 2023-10-26 |
296 | Style Transfer Meets Super-Resolution: Advancing Unpaired Infrared-to-Visible Image Translation with Detail Enhancement Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The problem of unpaired infrared-to-visible image translation has gained significant attention due to its ability to generate visible images with color information from low-detail … |
Yirui Shen; Jingxuan Kang; Shuang Li; Zhenjie Yu; Shuigen Wang; | Proceedings of the 31st ACM International Conference on … | 2023-10-26 |
297 | Interactive Image Style Transfer Guided By Graffiti Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Neural style transfer (NST) can quickly produce impressive artistic images, which allows ordinary people to become painter. The brushstrokes of stylized images created by the … |
Quan Wang; Yanli Ren; Xinpeng Zhang; Guorui Feng; | Proceedings of the 31st ACM International Conference on … | 2023-10-26 |
298 | Region-controlled Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, they fail to control the strength of textures in different regions of the content image. To address this issue, we propose a training method that uses a loss function to constrain the style intensity in different regions. |
Junjie Kang; Jinsong Wu; Shiqi Jiang; | arxiv-cs.CV | 2023-10-24 |
299 | Constructing Non-isotropic Gaussian Diffusion Model Using Isotropic Gaussian Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a Non-isotropic Gaussian Diffusion Model (NGDM) for image-to-image translation and image editing, which require translating or editing the source image while preserving the image regions irrelevant to the translation/editing task. |
Xi Yu; Xiang Gu; Haozhi Liu; Jian Sun; | nips | 2023-10-24 |
300 | ViSt3D: Video Stylization with 3D CNN Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To the best of our knowledge, we present the first approach to video stylization using 3D CNN directly, building upon insights from 2D image stylization. |
Ayush Pande; Gaurav Sharma; | nips | 2023-10-24 |
301 | ImageBrush: Learning Visual In-Context Instructions for Exemplar-Based Image Manipulation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel manipulation methodology, dubbed ImageBrush, that learns visual instructions for more accurate image editing. |
YA SHENG SUN et. al. | nips | 2023-10-24 |
302 | RePoseDM: Recurrent Pose Alignment and Gradient Guidance for Pose Guided Image Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Due to the leakage of the source pose in conditional guidance, we propose gradient guidance from pose interaction fields, which output the distance from the valid pose manifold given a predicted pose as input. |
Anant Khandelwal; | arxiv-cs.CV | 2023-10-24 |
303 | Text Fact Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Text style transfer is a prominent task that aims to control the style of text without inherently changing its factual content. To cover more text modification applications, such as adapting past news for current events and repurposing educational materials, we propose the task of text fact transfer, which seeks to transfer the factual content of a source text between topics without modifying its style. |
Nishant Balepur; Jie Huang; Kevin Chen-Chuan Chang; | arxiv-cs.CL | 2023-10-22 |
304 | Ladder Bottom-up Convolutional Bidirectional Variational Autoencoder for Image Translation of Dotted Arabic Expiration Dates Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes an approach of Ladder Bottom-up Convolutional Bidirectional Variational Autoencoder (LCBVAE) architecture for the encoder and decoder, which is trained on the image translation of the dotted Arabic expiration dates by reconstructing the Arabic dotted expiration dates into filled-in expiration dates. |
Ahmed Zidane; Ghada Soliman; | arxiv-cs.CV | 2023-10-21 |
305 | CycleNet: Rethinking Cycle Consistency in Text-Guided Diffusion for Image Manipulation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper introduces Cyclenet, a novel but simple method that incorporates cycle consistency into DMs to regularize image manipulation. |
Sihan Xu; Ziqiao Ma; Yidong Huang; Honglak Lee; Joyce Chai; | arxiv-cs.CV | 2023-10-19 |
306 | TOSS: High-quality Text-guided Novel View Synthesis from A Single Image IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In this paper, we present TOSS, which introduces text to the task of novel view synthesis (NVS) from just a single RGB image. While Zero-1-to-3 has demonstrated impressive … |
YUKAI SHI et. al. | ArXiv | 2023-10-16 |
307 | TOSS:High-quality Text-guided Novel View Synthesis from A Single Image Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present TOSS, which introduces text to the task of novel view synthesis (NVS) from just a single RGB image. |
YUKAI SHI et. al. | arxiv-cs.CV | 2023-10-16 |
308 | Chinese Painting Style Transfer Using Deep Generative Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we will study and leverage different state-of-the-art deep generative models for Chinese painting style transfer and evaluate the performance both qualitatively and quantitatively. |
Weijian Ma; Yanyang Kong; | arxiv-cs.CV | 2023-10-15 |
309 | LOVECon: Text-driven Training-Free Long Video Editing with ControlNet Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper aims to bridge the gap, establishing a simple and effective baseline for training-free diffusion model-based long video editing. |
Zhenyi Liao; Zhijie Deng; | arxiv-cs.CV | 2023-10-14 |
310 | Does Resistance to Style-transfer Equal Global Shape Bias? Measuring Network Sensitivity to Global Shape Configuration Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The current benchmark for evaluating a model’s global shape bias is a set of style-transferred images with the assumption that resistance to the attack of style transfer is related to the development of global structure sensitivity in the model. In this work, we show that networks trained with style-transfer images indeed learn to ignore style, but its shape bias arises primarily from local detail. |
Ziqi Wen; Tianqin Li; Zhi Jing; Tai Sing Lee; | arxiv-cs.CV | 2023-10-11 |
311 | Large Capacity Generative Image Steganography Via Image Style Transfer and Feature-wise Deep Fusion Related Papers Related Patents Related Grants Related Venues Related Experts View |
Youqiang Sun; Jianyi Liu; Ru Zhang; | Applied Intelligence | 2023-10-11 |
312 | Cancellable Biometric Authentication System By Image Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Recently, we have more and more opportunities to identify ourselves through information devices. In this paper, we propose a cancellable biometric authentication system, where … |
Souta Yamamoto; Hiroyuki Inaba; | 2023 IEEE 12th Global Conference on Consumer Electronics … | 2023-10-10 |
313 | Text-Guided Facial Image Manipulation for Wild Images Via Manipulation Direction-Based Loss Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: This paper proposes a novel text-guided facial image manipulation approach to improve robustness against the diversity of input images. Conventional text-guided facial image … |
Yuto Watanabe; Ren Togo; Keisuke Maeda; Takahiro Ogawa; M. Haseyama; | 2023 IEEE International Conference on Image Processing … | 2023-10-08 |
314 | WAIT: Feature Warping for Animation to Illustration Video Translation Using GANs Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we explore a new domain for video-to-video translation. |
SAMET HICSONMEZ et. al. | arxiv-cs.CV | 2023-10-07 |
315 | VTON-IT: Virtual Try-On Using Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we try to produce photo-realistic translated images through semantic segmentation and a generative adversarial architecture-based image translation network. |
Santosh Adhikari; Bishnu Bhusal; Prashant Ghimire; Anil Shrestha; | arxiv-cs.CV | 2023-10-06 |
316 | CineTransfer: Controlling A Robot to Imitate Cinematographic Style from A Single Example Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work presents CineTransfer, an algorithmic framework that drives a robot to record a video sequence that mimics the cinematographic style of an input video. |
Pablo Pueyo; Eduardo Montijano; Ana C. Murillo; Mac Schwager; | arxiv-cs.RO | 2023-10-05 |
317 | FT-Shield: A Watermark Against Unauthorized Fine-tuning in Text-to-Image Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: For example, a malicious user can employ fine-tuning techniques to replicate the style of an artist without consent. In light of this concern, we propose FT-Shield, a watermarking solution tailored for the fine-tuning of text-to-image diffusion models. |
YINGQIAN CUI et. al. | arxiv-cs.CV | 2023-10-03 |
318 | PanoStyle: Semantic, Geometry-Aware and Shading Independent Photorealistic Style Transfer for Indoor Panoramic Scenes Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: While current style transfer models have achieved impressive results for the application of artistic style to generic images, they face challenges in achieving photorealistic … |
MUHAMMAD TUKUR et. al. | 2023 IEEE/CVF International Conference on Computer Vision … | 2023-10-02 |
319 | Color and Texture Dual Pipeline Lightweight Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To solve the problem, we propose a Color and Texture Dual Pipeline Lightweight Style Transfer CTDP method, which employs a dual pipeline method to simultaneously output the results of color and texture transfer. |
ShiQi Jiang; | arxiv-cs.CV | 2023-10-02 |
320 | Neural Style Transfer for 3D Meshes Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Style transfer is a popular research topic in the field of computer vision. In 3D stylization, a mesh model is deformed to achieve a specific geometric style. We explore a general … |
Hongyuan Kang; Xiaopan Dong; Juan Cao; Zhonggui Chen; | Graph. Model. | 2023-10-01 |
321 | Zero-Shot Hate to Non-Hate Text Conversion Using Lexical Constraints Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Systems meant for tackling hate speech have been increasing in demand with the rapid growth of social media platforms. One way of controlling hate speech in texts is to transform … |
Zishan Ahmad; Vinnakota Sai Sujeeth; Asif Ekbal; | IEEE Transactions on Computational Social Systems | 2023-10-01 |
322 | Large-scale Apple Orchard Mapping from Multi-source Data Using The Semantic Segmentation Model with Image- To- Image Translation and Transfer Learning Related Papers Related Patents Related Grants Related Venues Related Experts View |
TINGTING ZHANG et. al. | Comput. Electron. Agric. | 2023-10-01 |
323 | Sketch Assisted Face Image Coding for Human and Machine Vision: A Joint Training Approach Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Image coding is one of the most fundamental techniques and is widely used in image/video processing and multimedia communications. Current image coding methods are mainly … |
Xin Fang; Yiping Duan; Qiyuan Du; Xiaoming Tao; Fan Li; | IEEE Transactions on Circuits and Systems for Video … | 2023-10-01 |
324 | An Easy Zero-shot Learning Combination: Texture Sensitive Semantic Segmentation IceHrNet and Advanced Style Transfer Learning Strategy Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We proposed an easy method of Zero-Shot semantic segmentation by using style transfer. |
ZHIYONG YANG et. al. | arxiv-cs.CV | 2023-09-30 |
325 | Controlling Neural Style Transfer with Deep Reinforcement Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose the first deep Reinforcement Learning (RL) based architecture that splits one-step style transfer into a step-wise process for the NST task. |
CHENGMING FENG et. al. | arxiv-cs.CV | 2023-09-30 |
326 | Two Birds, One Stone: A Unified Framework for Joint Learning of Image and Video Style Transfers Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In order to achieve satisfying image and video style transfers, two different models are inevitably required with separate training processes on image and video domains, respectively. In this paper, we show that this can be precluded by introducing UniST, a Unified Style Transfer framework for both images and videos. |
Bohai Gu; Heng Fan; Libo Zhang; | iccv | 2023-09-27 |
327 | UMFuse: Unified Multi View Fusion for Human Editing Applications Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we explore the utilization of multiple views to minimize the issue of missing information and generate an accurate representation of the underlying human model. |
RISHABH JAIN et. al. | iccv | 2023-09-27 |
328 | Synthetic Latent Fingerprint Generation Using Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a simple and effective approach using style transfer and image blending to synthesize realistic latent fingerprints. |
Amol S. Joshi; Ali Dabouei; Nasser Nasrabadi; Jeremy Dawson; | arxiv-cs.CV | 2023-09-27 |
329 | Bidirectionally Deformable Motion Modulation For Video-based Human Pose Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Considering the difficulties in transferring highly structural patterns on the garments and discontinuous poses, existing methods often generate unsatisfactory results such as distorted textures and flickering artifacts. To address these issues, we propose a novel Deformable Motion Modulation (DMM) that utilizes geometric kernel offset with adaptive weight modulation to simultaneously perform feature alignment and style transfer. |
WING-YIN YU et. al. | iccv | 2023-09-27 |
330 | StyleInV: A Temporal Style Modulated Inversion Network for Unconditional Video Generation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce a novel motion generator design that uses a learning-based inversion network for GAN. |
Yuhan Wang; Liming Jiang; Chen Change Loy; | iccv | 2023-09-27 |
331 | StyleDiffusion: Controllable Disentangled Style Transfer Via Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a new C-S disentangled framework for style transfer without using previous assumptions. |
Zhizhong Wang; Lei Zhao; Wei Xing; | iccv | 2023-09-27 |
332 | Scenimefy: Learning to Craft Anime Scene Via Semi-Supervised Image-to-Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Despite promising attempts, previous efforts are still incompetent in achieving satisfactory results with consistent semantic preservation, evident stylization, and fine details. In this study, we propose Scenimefy, a novel semi-supervised image-to-image translation framework that addresses these challenges. |
Yuxin Jiang; Liming Jiang; Shuai Yang; Chen Change Loy; | iccv | 2023-09-27 |
333 | Cross-modal Latent Space Alignment for Image to Avatar Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a novel method for automatic vectorized avatar generation from a single portrait image. |
MANUEL LADRON DE GUEVARA et. al. | iccv | 2023-09-27 |
334 | Zero-Shot Contrastive Loss for Text-Guided Diffusion Image Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Existing methods require computationally expensive fine-tuning of diffusion models or additional neural network. To address this, here we propose a zero-shot contrastive loss for diffusion models that doesn’t require additional fine-tuning or auxiliary networks. |
Serin Yang; Hyunmin Hwang; Jong Chul Ye; | iccv | 2023-09-27 |
335 | WaveIPT: Joint Attention and Flow Alignment in The Wavelet Domain for Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To leverage the advantages of both attention and flow simultaneously, we propose Wavelet-aware Image-based Pose Transfer (WaveIPT) to fuse the attention and flow in the wavelet domain. |
Liyuan Ma; Tingwei Gao; Haitian Jiang; Haibin Shen; Kejie Huang; | iccv | 2023-09-27 |
336 | Deep Style Transfer for Generation of Photo-realistic Synthetic Images of CNT Forests Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Carbon nanotubes (CNTs) are promising nano-materials with diverse applications in various fields, ranging from electronics and energy storage to biomedical applications. … |
Prashanth Kotha; Minasadat Attari; Matthew R. Maschmann; F. Bunyak; | 2023 IEEE Applied Imagery Pattern Recognition Workshop … | 2023-09-27 |
337 | Collecting The Puzzle Pieces: Disentangled Self-Driven Human Pose Transfer By Permuting Textures Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose Pose Transfer by Permuting Textures, a self-driven human pose transfer approach that disentangles pose from texture at the patch-level. |
Nannan Li; Kevin J Shih; Bryan A. Plummer; | iccv | 2023-09-27 |
338 | AesPA-Net: Aesthetic Pattern-Aware Style Transfer Networks IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce a novel metric, namely pattern repeatability, that quantifies the repetition of patterns in the style image. |
KIBEOM HONG et. al. | iccv | 2023-09-27 |
339 | In-Style: Bridging Text and Uncurated Videos with Style Transfer for Text-Video Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To this end, we propose an approach, In-Style, that learns the style of the text queries and transfers it to uncurated web videos. |
Nina Shvetsova; Anna Kukleva; Bernt Schiele; Hilde Kuehne; | iccv | 2023-09-27 |
340 | Locally Stylized Neural Radiance Fields IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a stylization framework for NeRF based on local style transfer. |
Hong-Wing Pang; Binh-Son Hua; Sai-Kit Yeung; | iccv | 2023-09-27 |
341 | Incorporating Ensemble and Transfer Learning For An End-To-End Auto-Colorized Image Detection Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper presents a novel approach that combines the advantages of transfer and ensemble learning approaches to help reduce training time and resource requirements while proposing a model to classify natural color and computer-colorized images. |
Ahmed Samir Ragab; Shereen Aly Taie; Howida Youssry Abdelnaby; | arxiv-cs.CV | 2023-09-25 |
342 | MM-NeRF: Multimodal-Guided 3D Multi-Style Transfer of Neural Radiance Field Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we reveal that the common training method of stylization with NeRF, which generates stylized multi-view supervision by 2D style transfer models, causes the same object in supervision to show various states (color tone, details, etc.) in different views, leading NeRF to tend to smooth the texture details, further resulting in low-quality rendering for 3D multi-style transfer. |
Zijiang Yang; Zhongwei Qiu; Chang Xu; Dongmei Fu; | arxiv-cs.CV | 2023-09-24 |
343 | MOSAIC: Multi-Object Segmented Arbitrary Stylization Using CLIP Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: On the other hand, diffusion style transfer methods also suffer from the same issue because the regional stylization control over the stylized output is ineffective. To address this problem, We propose a new method Multi-Object Segmented Arbitrary Stylization Using CLIP (MOSAIC), that can apply styles to different objects in the image based on the context extracted from the input prompt. |
PRAJWAL GANUGULA et. al. | arxiv-cs.CV | 2023-09-24 |
344 | Portrait Stylization: Artistic Style Transfer with Auxiliary Networks for Human Face Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes the use of embeddings from an auxiliary pre-trained face recognition model to encourage the algorithm to propagate human face features from the content image to the final stylized result. |
Thiago Ambiel; | arxiv-cs.CV | 2023-09-23 |
345 | Masked Discriminators for Content-Consistent Unpaired Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we show that masking the inputs of a global discriminator for both domains with a content-based mask is sufficient to reduce content inconsistencies significantly. |
Bonifaz Stuhr; Jürgen Brauer; Bernhard Schick; Jordi Gonzàlez; | arxiv-cs.CV | 2023-09-22 |
346 | TextCLIP: Text-Guided Face Image Generation And Manipulation Without Adversarial Training Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose TextCLIP, a unified framework for text-guided image generation and manipulation without adversarial training. |
Xiaozhou You; Jian Zhang; | arxiv-cs.CV | 2023-09-21 |
347 | Boosting SAR Aircraft Detection Performance with Multi-Stage Domain Adaptation Training Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Deep learning has achieved significant success in various synthetic aperture radar (SAR) imagery interpretation tasks. However, automatic aircraft detection is still challenging … |
Wenbo Yu; Jiamu Li; Zijian Wang; Zhongjun Yu; | Remote. Sens. | 2023-09-20 |
348 | Retinex-guided Channel-grouping Based Patch Swap for Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Since the finite features harvested from one single aesthetic style image are inadequate to represent the rich textures of the content natural image, existing techniques treat the full-channel style feature patches as simple signal tensors and create new style feature patches via signal-level fusion, which ignore the implicit diversities existed in style features and thus fail for generating better stylised results. In this paper, we propose a Retinex theory guided, channel-grouping based patch swap technique to solve the above challenges. |
Chang Liu; Yi Niu; Mingming Ma; Fu Li; Guangming Shi; | arxiv-cs.CV | 2023-09-19 |
349 | Instant Photorealistic Style Transfer: A Lightweight and Adaptive Approach Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In this paper, we propose an Instant Photorealistic Style Transfer (IPST) approach, designed to achieve instant photorealistic style transfer on super-resolution inputs without … |
Rong Liu; Enyu Zhao; Zhiyuan Liu; A. Feng; Scott John Easley; | ArXiv | 2023-09-18 |
350 | Universal Photorealistic Style Transfer: A Lightweight and Adaptive Approach Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, existing methods often encounter challenges such as color tone distortions, dependency on pair-wise pre-training, inefficiency with high-resolution inputs, and the need for additional constraints in video style transfer tasks. To address these issues, we propose a Universal Photorealistic Style Transfer (UPST) framework that delivers accurate photorealistic style transfer on high-resolution images and videos without relying on pre-training. |
Rong Liu; Enyu Zhao; Zhiyuan Liu; Andrew Feng; Scott John Easley; | arxiv-cs.CV | 2023-09-18 |
351 | Speech-to-Speech Translation with Discrete-Unit-Based Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: The acoustic language model we introduce for style transfer leverages self-supervised in-context learning, acquiring style transfer ability without relying on any speaker-parallel data, thereby overcoming data scarcity. |
YONGQI WANG et. al. | arxiv-cs.SD | 2023-09-14 |
352 | Nucleus-aware Self-supervised Pretraining Using Unpaired Image-to-image Translation for Histopathology Images Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a novel nucleus-aware self-supervised pretraining framework for histopathology images. |
ZHIYUN SONG et. al. | arxiv-cs.CV | 2023-09-13 |
353 | TSSAT: Two-Stage Statistics-Aware Transformation for Artistic Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Consequently, the stylization results either fail to capture abundant and diversified local style patterns, or contain undesired semantic information of the style image and deviate from the global style distribution. To address this issue, we imitate the drawing process of humans and propose a Two-Stage Statistics-Aware Transformation (TSSAT) module, which first builds the global style foundation by aligning the global statistics of content and style features and then further enriches local style details by swapping the local statistics (instead of local features) in a patch-wise manner, significantly improving the stylization effects. |
Haibo Chen; Lei Zhao; Jun Li; Jian Yang; | arxiv-cs.CV | 2023-09-12 |
354 | PAI-Diffusion: Constructing and Serving A Family of Open Chinese Diffusion Models for Text-to-image Synthesis on The Cloud Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: While existing diffusion models have shown promise in generating images from textual descriptions, they often neglect domain-specific contexts and lack robustness in handling the Chinese language. This paper introduces PAI-Diffusion, a comprehensive framework that addresses these limitations. |
CHENGYU WANG et. al. | arxiv-cs.CL | 2023-09-11 |
355 | MoEController: Instruction-based Arbitrary Image Manipulation with Mixture-of-Expert Controllers Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a method with a mixture-of-expert (MOE) controllers to align the text-guided capacity of diffusion models with different kinds of human instructions, enabling our model to handle various open-domain image manipulation tasks with natural language instructions. |
Sijia Li; Chen Chen; Haonan Lu; | arxiv-cs.CV | 2023-09-08 |
356 | Stroke-based Neural Painting and Stylization with Dynamically Predicted Painting Region IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To solve the problem, we propose Compositional Neural Painter, a novel stroke-based rendering framework which dynamically predicts the next painting region based on the current canvas, instead of dividing the image plane uniformly into painting regions. |
TENG HU et. al. | arxiv-cs.CV | 2023-09-07 |
357 | StyleAdapter: A Unified Stylized Image Generation Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose StyleAdapter, a unified stylized image generation model capable of producing a variety of stylized images that match both the content of a given prompt and the style of reference images, without the need for per-style fine-tuning. |
ZHOUXIA WANG et. al. | arxiv-cs.CV | 2023-09-04 |
358 | Unsupervised Image-to-image Translation in Multi-parametric MRI of Bladder Cancer Related Papers Related Patents Related Grants Related Venues Related Experts View |
ZHIYING CHEN et. al. | Eng. Appl. Artif. Intell. | 2023-09-01 |
359 | Impact of Image Context for Single Deep Learning Face Morphing Attack Detection Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This study investigates the impact of the alignment settings of input images on deep learning face morphing detection performance. |
Joana Pimenta; Iurii Medvedev; Nuno Gonçalves; | arxiv-cs.CV | 2023-09-01 |
360 | Shape-Consistent One-Shot Unsupervised Domain Adaptation for Rail Surface Defect Segmentation IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Deep neural networks have greatly improved the performance of rail surface defect segmentation when the test samples have the same distribution as the training samples. However, … |
SHUAI MA et. al. | IEEE Transactions on Industrial Informatics | 2023-09-01 |
361 | Semantic Image Synthesis Via Class-Adaptive Cross-Attention Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In response, we designed a novel architecture where cross-attention layers are used in place of SPADE for learning shape-style correlations and so conditioning the image generation process. |
Tomaso Fontanini; Claudio Ferrari; Giuseppe Lisanti; Massimo Bertozzi; Andrea Prati; | arxiv-cs.CV | 2023-08-30 |
362 | ParaGuide: Guided Diffusion Paraphrasers for Plug-and-Play Textual Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In contrast, we introduce a novel diffusion-based framework for general-purpose style transfer that can be flexibly adapted to arbitrary target styles at inference time. |
Zachary Horvitz; Ajay Patel; Chris Callison-Burch; Zhou Yu; Kathleen McKeown; | arxiv-cs.CL | 2023-08-29 |
363 | WSAM: Visual Explanations from Style Augmentation As Adversarial Attacker and Their Influence in Image Classification Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: With our augmentation strategy, all models not only present incredible robustness against image stylizing but also outperform all previous methods and surpass the state-of-the-art performance for the STL-10 dataset. |
Felipe Moreno-Vera; Edgar Medina; Jorge Poco; | arxiv-cs.CV | 2023-08-28 |
364 | MagicEdit: High-Fidelity and Temporally Coherent Video Editing IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this report, we present MagicEdit, a surprisingly simple yet effective solution to the text-guided video editing task. |
Jun Hao Liew; Hanshu Yan; Jianfeng Zhang; Zhongcong Xu; Jiashi Feng; | arxiv-cs.CV | 2023-08-28 |
365 | ARF-Plus: Controlling Perceptual Factors in Artistic Radiance Fields for 3D Scene Stylization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present ARF-Plus, a 3D neural style transfer framework offering manageable control over perceptual factors, to systematically explore the perceptual controllability in 3D scene stylization. |
Wenzhao Li; Tianhao Wu; Fangcheng Zhong; Cengiz Oztireli; | arxiv-cs.CV | 2023-08-23 |
366 | TeSTNeRF: Text-Driven 3D Style Transfer Via Cross-Modal Learning IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Simply combining image/video style transfer methods and novel view synthesis methods results in flickering when changing viewpoints, while existing 3D style transfer methods learn styles from images instead of texts. To address this problem, we for the first time design an efficient text-driven model for 3D style transfer, named TeSTNeRF, to stylize the scene using texts via cross-modal learning: we leverage an advanced text encoder to embed the texts in order to control 3D style transfer and align the input text and output stylized images in latent space. |
JIAFU CHEN et. al. | ijcai | 2023-08-23 |
367 | Improving The Transferability of Adversarial Examples with Arbitrary Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Hence, we propose a novel attack method named Style Transfer Method (STM) that utilizes a proposed arbitrary style transfer network to transform the images into different domains. |
ZHIJIN GE et. al. | arxiv-cs.CV | 2023-08-21 |
368 | Regularization By Neural Style Transfer for MRI Field-Transfer Reconstruction with Limited Data Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a regularization by neural style transfer (RNST) method to further leverage the priors from the neural transfer and denoising engine. |
GUOYAO SHEN et. al. | arxiv-cs.CV | 2023-08-21 |
369 | Color Prompting for Data-Free Continual Unsupervised Domain Adaptive Person Re-Identification Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a Color Prompting (CoP) method for data-free continual unsupervised domain adaptive person Re-ID. |
JIANYANG GU et. al. | arxiv-cs.CV | 2023-08-21 |
370 | A White-Box False Positive Adversarial Attack Method on Contrastive Loss Based Offline Handwritten Signature Verification Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we tackle the challenge of white-box false positive adversarial attacks on contrastive loss based offline handwritten signature verification models. |
Zhongliang Guo; Weiye Li; Yifei Qian; Ognjen Arandjelović; Lei Fang; | arxiv-cs.CV | 2023-08-17 |
371 | A White-Box False Positive Adversarial Attack Method on Contrastive Loss-Based Offline Handwritten Signature Verification Models IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In this paper, we tackle the challenge of white-box false positive adversarial attacks on contrastive loss based offline handwritten signature verification models. We propose a … |
Zhongliang Guo; Yifei Qian; Ognjen Arandjelovic; Lei Fang; | ArXiv | 2023-08-17 |
372 | Diff-CAPTCHA: An Image-based CAPTCHA with Security Enhanced By Denoising Diffusion Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, an image-click CAPTCHA scheme called Diff-CAPTCHA is proposed based on denoising diffusion models. |
Ran Jiang; Sanfeng Zhang; Linfeng Liu; Yanbing Peng; | arxiv-cs.CR | 2023-08-16 |
373 | CoDeF: Content Deformation Fields for Temporally Consistent Video Processing IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present the content deformation field CoDeF as a new type of video representation, which consists of a canonical content field aggregating the static contents in the entire video and a temporal deformation field recording the transformations from the canonical image (i.e., rendered from the canonical content field) to each individual frame along the time axis. |
HAO OUYANG et. al. | arxiv-cs.CV | 2023-08-15 |
374 | A Novel Blind Tamper Detection and Localization Scheme for Multiple Faces in Digital Images Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Face image manipulation detection (FIMD) is a research area of great interest, widely applicable in fields requiring data security and authentication. Existing FIMD schemes aim to … |
Rasha Thabit Mohammed; | IET Image Process. | 2023-08-14 |
375 | Hierarchy Flow For High-Fidelity Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose Hierarchy Flow, a novel flow-based model to achieve better content preservation during translation. |
Weichen Fan; Jinghuan Chen; Ziwei Liu; | arxiv-cs.CV | 2023-08-13 |
376 | Zero-shot Text-driven Physically Interpretable Face Editing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes a novel and physically interpretable method for face editing based on arbitrary text prompts. |
YAPENG MENG et. al. | arxiv-cs.CV | 2023-08-11 |
377 | BATINet: Background-Aware Text to Image Synthesis and Manipulation Network Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this study, we analyzed a novel Background-Aware Text2Image (BAT2I) task in which the generated content matches the input background. |
Ryugo Morita; Zhiqiang Zhang; Jinjia Zhou; | arxiv-cs.CV | 2023-08-10 |
378 | A Forensic Methodology for Detecting Image Manipulations Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this study, image file and mobile forensic artifacts analysis were conducted for detecting image manipulation. |
Jiwon Lee; Seungjae Jeon; Yunji Park; Jaehyun Chung; Doowon Jeong; | arxiv-cs.MM | 2023-08-09 |
379 | VAST: Vivify Your Talking Avatar Via Zero-Shot Expressive Facial Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes an unsupervised variational style transfer model (VAST) to vivify the neutral photo-realistic avatars. |
LIYANG CHEN et. al. | arxiv-cs.CV | 2023-08-09 |
380 | A Comparative Study of Image-to-Image Translation Using GANs for Synthetic Child Race Data Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work proposes the utilization of image-to-image transformation to synthesize data of different races and thus adjust the ethnicity of children’s face data. |
Wang Yao; Muhammad Ali Farooq; Joseph Lemley; Peter Corcoran; | arxiv-cs.CV | 2023-08-08 |
381 | DiffSynth: Latent In-Iteration Deflickering for Realistic Video Synthesis Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose DiffSynth, a novel approach that aims to convert image synthesis pipelines to video synthesis pipelines. |
ZHONGJIE DUAN et. al. | arxiv-cs.CV | 2023-08-07 |
382 | Photorealistic and Identity-Preserving Image-Based Emotion Manipulation with Latent Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we investigate the emotion manipulation capabilities of diffusion models with in-the-wild images, a rather unexplored application area relative to the vast and rapidly growing literature for image-to-image translation tasks. |
Ioannis Pikoulis; Panagiotis P. Filntisis; Petros Maragos; | arxiv-cs.CV | 2023-08-06 |
383 | Diving Deeper Into Volume Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The Pando Thrower is a weapon of great importance in Disney’s Strange World. The explosive Pando fuel mixture is expelled with turbulent, fluid-like motion, with branching arcs … |
Mike Navarro; | ACM SIGGRAPH 2023 Talks | 2023-08-06 |
384 | Singed Silhouettes and Feed Forward Flames: Volumetric Neural Style Transfer for Expressive Fire Simulation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: While controlling simulated gaseous volumes remains an ongoing battle when seeking realism in computer graphics, creating appealing characters entirely out of these simulations … |
Paul Kanyuk; V. C. Azevedo; Raphael Ortiz; Jingwei Tang; | ACM SIGGRAPH 2023 Talks | 2023-08-06 |
385 | FASTER: A Font-Agnostic Scene Text Editing and Rendering Framework Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Despite its utility in numerous real-world applications, existing style-transfer-based approaches have shown sub-par editing performance due to (1) complex image backgrounds, (2) diverse font attributes, and (3) varying word lengths within the text. To address such limitations, in this paper, we propose a novel font-agnostic scene text editing and rendering framework, named FASTER, for simultaneously generating text in arbitrary styles and locations while preserving a natural and realistic appearance and structure. |
ALLOY DAS et. al. | arxiv-cs.CV | 2023-08-05 |
386 | Superpixel-Based Style Transfer Method for Single-Temporal Remote Sensing Image Identification in Forest Type Groups Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Forests are the most important carbon reservoirs on land, and forest carbon sinks can effectively reduce atmospheric CO2 concentrations and mitigate climate change. In recent … |
Zhenyu Yu; Jinnian Wang; Xiankun Yang; Juan Ma; | Remote. Sens. | 2023-08-04 |
387 | MSSRNet: Manipulating Sequential Style Representation for Unsupervised Text Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In fact, each token of a text contains different style intensity and makes different contribution to the overall style. Our proposed method addresses this issue by assigning individual style vector to each token in a text, allowing for fine-grained control and manipulation of the style strength. |
Yazheng Yang; Zhou Zhao; Qi Liu; | kdd | 2023-08-04 |
388 | ADS-Cap: A Framework for Accurate and Diverse Stylized Captioning with Unpaired Stylistic Corpora Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a novel framework to generate Accurate and Diverse Stylized Captions (ADS-Cap). |
KANZHI CHENG et. al. | arxiv-cs.CV | 2023-08-02 |
389 | ImageBrush: Learning Visual In-Context Instructions for Exemplar-Based Image Manipulation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel manipulation methodology, dubbed ImageBrush, that learns visual instructions for more accurate image editing. |
YASHENG SUN et. al. | arxiv-cs.CV | 2023-08-01 |
390 | W2GAN: Importance Weight and Wavelet Feature Guided Image-to-Image Translation Under Limited Data Related Papers Related Patents Related Grants Related Venues Related Experts View |
Qiuxia Yang; Yuanyuan Pu; Zhengpeng Zhao; Dan Xu; Siqi Li; | Comput. Graph. | 2023-08-01 |
391 | Joint Image-to-Image Translation for Traffic Monitoring Driver Face Image Enhancement Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The real traffic monitoring driver face (TMDF) images are with complex multiple degradations, which decline face recognition accuracy in real intelligent transportation systems … |
CHANGHUI HU et. al. | IEEE Transactions on Intelligent Transportation Systems | 2023-08-01 |
392 | Controlling Geometric Abstraction and Texture for Artistic Images Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present a novel method for the interactive control of geometric abstraction and texture in artistic images. |
MARTIN BÜSSEMEYER et. al. | arxiv-cs.CV | 2023-07-31 |
393 | InfoStyler: Disentanglement Information Bottleneck for Artistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Although effective, ignoring the clear disentanglement of the content features and the style features from the first beginning, they have difficulty in balancing between content preservation and style transferring. To tackle this problem, we propose a novel information disentanglement method, named InfoStyler, to capture the minimal sufficient information for both content and style representations from the pre-trained encoding network. |
Yueming Lyu; Yue Jiang; Bo Peng; Jing Dong; | arxiv-cs.CV | 2023-07-30 |
394 | The Generation of Articulatory Animations Based on Keypoint Detection and Motion Transfer Combined with Image Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Knowing the correct positioning of the tongue and mouth for pronunciation is crucial for learning English pronunciation correctly. Articulatory animation is an effective way to … |
Xufeng Ling; Yu Zhu; W. Liu; Jingxin Liang; Jie Yang; | Comput. | 2023-07-28 |
395 | CLIP-PAE: Projection-Augmentation Embedding to Extract Relevant Features for A Disentangled, Interpretable and Controllable Text-Guided Face Manipulation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Disentanglement, interpretability, and controllability are also hard to guarantee for manipulation. To alleviate these problems, we propose to define corpus subspaces spanned by relevant prompts to capture specific image characteristics. |
Chenliang Zhou; Fangcheng Zhong; Cengiz Öztireli; | siggraph | 2023-07-26 |
396 | On The Fly Neural Style Smoothing for Risk-Averse Domain Generalization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To enable risk-averse predictions from a DG classifier, we propose a novel inference procedure, Test-Time Neural Style Smoothing (TT-NSS), that uses a style-smoothed version of the DG classifier for prediction at test time. |
Akshay Mehra; Yunbei Zhang; Bhavya Kailkhura; Jihun Hamm; | arxiv-cs.CV | 2023-07-17 |
397 | SAR-UT: A Synthetic-to-Measured SAR Image Translation Network Based on Transformer IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In recent years, some studies used computer to generate synthetic aperture radar (SAR) images. Although the synthetic SAR image looks realistic, there is a domain gap and … |
Hengyi Hu; Z. Cui; Zheng Zhou; Z. Cao; | IGARSS 2023 – 2023 IEEE International Geoscience and Remote … | 2023-07-16 |
398 | Dense Multitask Learning to Reconfigure Comics Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we develop a MultiTask Learning (MTL) model to achieve dense predictions for comics panels to, in turn, facilitate the transfer of comics from one publication channel to another by assisting authors in the task of reconfiguring their narratives. |
Deblina Bhattacharjee; Sabine Süsstrunk; Mathieu Salzmann; | arxiv-cs.CV | 2023-07-16 |
399 | Bidirectionally Deformable Motion Modulation For Video-based Human Pose Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Considering the difficulties in transferring highly structural patterns on the garments and discontinuous poses, existing methods often generate unsatisfactory results such as distorted textures and flickering artifacts. To address these issues, we propose a novel Deformable Motion Modulation (DMM) that utilizes geometric kernel offset with adaptive weight modulation to simultaneously perform feature alignment and style transfer. |
WING-YIN YU et. al. | arxiv-cs.CV | 2023-07-15 |
400 | Sem-CS: Semantic CLIPStyler for Text-Based Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, the ground semantics of objects in the style transfer output is lost due to style spill-over on salient and background objects (content mismatch) or over-stylization. To solve this, we propose Semantic CLIPStyler (Sem-CS), that performs semantic style transfer. |
Chanda Grover Kamra; Indra Deep Mastan; Debayan Gupta; | arxiv-cs.CV | 2023-07-12 |
401 | Substance or Style: What Does Your Image Embedding Know? Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Probes are small networks that predict properties of underlying data from embeddings, and they provide a targeted, effective way to illuminate the information contained in … |
CYRUS RASHTCHIAN et. al. | ArXiv | 2023-07-10 |
402 | DIFF-NST: Diffusion Interleaving For DeFormable Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: With the recent introduction of diffusion models, such as Stable Diffusion, we can access far more powerful image generation techniques, enabling new possibilities. In our work, we propose using this new class of models to perform style transfer while enabling deformable style transfer, an elusive capability in previous models. |
DAN RUTA et. al. | arxiv-cs.CV | 2023-07-09 |
403 | Text Style Transfer Back-Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: For natural inputs, BT brings only slight improvements and sometimes even adverse effects. To address this issue, we propose Text Style Transfer Back Translation (TST BT), which uses a style transfer to modify the source side of BT data. |
DAIMENG WEI et. al. | acl | 2023-07-08 |
404 | PEIT: Bridging The Modality Gap with Pre-trained Models for End-to-End Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose PEIT, an end-to-end image translation framework that bridges the modality gap with pre-trained models. |
Shaolin Zhu; Shangjie Li; Yikun Lei; Deyi Xiong; | acl | 2023-07-08 |
405 | Rendering and Reconstruction Based 3D Portrait Stylization Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Both 2D images and 3D models are vital aspects of portrait applications. Existing style transfer methods principally emphasized 2D images, neglecting the urge for 3D style … |
Shaoxu Li; Ye Pan; | 2023 IEEE International Conference on Multimedia and Expo … | 2023-07-01 |
406 | A Compact Transformer for Adaptive Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Due to the limitation of spatial receptive field, it is challenging for CNN-based style transfer methods to capture rich and long-range semantic concepts in artworks. Though the … |
Yi Li; Xinxiong Xie; Haiyan Fu; Xiangyang Luo; Yanqing Guo; | 2023 IEEE International Conference on Multimedia and Expo … | 2023-07-01 |
407 | PCFN: Progressive Cross-Modal Fusion Network for Human Pose Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The goal of human pose transfer is to transfer the human in the image from the original pose to the desired one. Existing methods utilizing progressive manner have achieved great … |
Wei Yu; Yanping Li; Rui Wang; W. Cao; Wei Xiang; | IEEE Transactions on Circuits and Systems for Video … | 2023-07-01 |
408 | Structure-preserving Image Translation for Multi-source Medical Image Domain Adaptation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
Myeongkyun Kang; Philip Chikontwe; D. Won; Miguel Luna; Sang Hyun Park; | Pattern Recognit. | 2023-07-01 |
409 | Point Cloud-Based Free Viewpoint Artistic Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In recent years, artistic style transfer has gained popularity as a means of creating visually appealing images by injecting style into the content image. Although various methods … |
Eun-Gyeong Bae; Jaekyung Kim; Sanghoon Lee; | 2023 IEEE International Conference on Multimedia and Expo … | 2023-07-01 |
410 | StyleStegan: Leak-free Style Transfer Based on Feature Steganography Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In modern social networks, existing style transfer methods suffer from a serious content leakage issue, which hampers the ability to achieve serial and reversible stylization, thereby hindering the further propagation of stylized images in social networks. To address this problem, we propose a leak-free style transfer method based on feature steganography. |
Xiujian Liang; Bingshan Liu; Qichao Ying; Zhenxing Qian; Xinpeng Zhang; | arxiv-cs.CV | 2023-07-01 |
411 | Ship Detection in Low-Quality SAR Images Via An Unsupervised Domain Adaption Method Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Ship detection in low-quality Synthetic Aperture Radar (SAR) images poses a persistent challenge. Noise signals in complex environments disrupt imaging conditions, hindering SAR … |
Xinyang Pu; He Jia; Yu Xin; Feng Wang; Haipeng Wang; | Remote. Sens. | 2023-06-29 |
412 | SinDDM: A Single Image Denoising Diffusion Model IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Here, we introduce a framework for training a DDM on a single image. |
Vladimir Kulikov; Shahar Yadin; Matan Kleiner; Tomer Michaeli; | icml | 2023-06-27 |
413 | CFFT-GAN: Cross-Domain Feature Fusion Transformer for Exemplar-Based Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a more general learning approach by considering two domain features as a whole and learning both inter-domain correspondence and intra-domain potential information interactions. |
TIANXIANG MA et. al. | aaai | 2023-06-26 |
414 | User-Controllable Arbitrary Style Transfer Via Entropy Regularization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel solution ensuring both efficiency and diversity for generating multiple user-controllable AST results by systematically modulating AST behavior at run-time. |
JIAXIN CHENG et. al. | aaai | 2023-06-26 |
415 | Progressive Energy-Based Cooperative Learning for Multi-Domain Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Since the style generator is represented as an domain-specific distribution of style codes, the translator can provide a one-to-many transformation (i.e., diversified generation) between source domain and target domain. To train our framework, we propose a likelihood-based multi-domain cooperative learning algorithm to jointly train the multi-domain descriptor and the diversified image generator (including translator, style encoder, and style generator modules) via multi-domain MCMC teaching, in which the descriptor guides the diversified image generator to shift its probability density toward the data distribution, while the diversified image generator uses its randomly translated images to initialize the descriptor’s Langevin dynamics process for efficient sampling. |
Weinan Song; Yaxuan Zhu; Lei He; Yingnian Wu; Jianwen Xie; | arxiv-cs.CV | 2023-06-26 |
416 | Preserving Structural Consistency in Arbitrary Artist and Artwork Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: These methods not only homogenize the artist-style of different artworks of the same artist but also lack generalization for the unseen artists. To solve these challenges, we propose a double-style transferring module (DSTM). |
JINGYU WU et. al. | aaai | 2023-06-26 |
417 | CodeStylist: A System for Performing Code Style Transfer Using Neural Networks Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Code style refers to attributes of computer programs that affect their readability, maintainability, and performance. Enterprises consider code style as important and enforce … |
CHIH-KAI TING et. al. | AAAI Conference on Artificial Intelligence | 2023-06-26 |
418 | SHUNIT: Style Harmonization for Unpaired Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a novel solution for unpaired image-to-image (I2I) translation. |
Seokbeom Song; Suhyeon Lee; Hongje Seong; Kyoungwon Min; Euntai Kim; | aaai | 2023-06-26 |
419 | MicroAST: Towards Super-fast Ultra-Resolution Arbitrary Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Despite the recent rapid progress, existing AST methods are either incapable or too slow to run at ultra-resolutions (e.g., 4K) with limited resources, which heavily hinders their further applications. In this paper, we tackle this dilemma by learning a straightforward and lightweight model, dubbed MicroAST. |
ZHIZHONG WANG et. al. | aaai | 2023-06-26 |
420 | Multimodal Deep Generative Models for Remote Medical Applications Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Visible-to-Thermal (VT) face translation is an under-studied problem of image-to-image translation that offers an AI-enabled alternative to traditional thermal sensors. Over three … |
Catherine Ordun; | AAAI Conference on Artificial Intelligence | 2023-06-26 |
421 | Practical Disruption of Image Translation Deepfake Networks Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work we propose Leaking Transferable Perturbations (LTP), an algorithm that significantly reduces the number of queries needed to disrupt an image translation network by dynamically re-purposing previous disruptions into new query efficient disruptions. |
Nataniel Ruiz; Sarah Adel Bargal; Cihang Xie; Stan Sclaroff; | aaai | 2023-06-26 |
422 | Frequency Domain Disentanglement for Arbitrary Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Therefore, these methods always suffer from low-quality results because of the sub-optimal disentanglement. To address such a challenge, this paper proposes the frequency mixer (FreMixer) module that disentangles and re-entangles the frequency spectrum of content and style components in the frequency domain. |
DONGYANG LI et. al. | aaai | 2023-06-26 |
423 | CLIPVG: Text-Guided Image Manipulation Using Differentiable Vector Graphics IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce CLIPVG, a text-guided image manipulation framework using differentiable vector graphics, which is also the first CLIP-based general image manipulation framework that does not require any additional generative models. |
YIREN SONG et. al. | aaai | 2023-06-26 |
424 | AdaCM: Adaptive ColorMLP for Real-Time Universal Photo-Realistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose the Adaptive ColorMLP (AdaCM), an effective and efficient framework for universal photo-realistic style transfer. |
TIANWEI LIN et. al. | aaai | 2023-06-26 |
425 | PP-GAN : Style Transfer from Korean Portraits to ID Photos Using Landmark Extractor with GAN Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Owing to its distinct characteristics from the hair in ID photos, transferring the Gat is challenging. To address this issue, this study proposes a deep learning network that can perform style transfer, including the Gat, while preserving the identity of the face. |
Jongwook Si; Sungyoung Kim; | arxiv-cs.CV | 2023-06-23 |
426 | What to Learn: Features, Image Transformations, or Both? Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose to combine an image transformation network and a feature-learning network to improve long-term localization performance. |
Yuxuan Chen; Binbin Xu; Frederike Dümbgen; Timothy D. Barfoot; | arxiv-cs.RO | 2023-06-22 |
427 | LisaCLIP: Locally Incremental Semantics Adaptation Towards Zero-shot Text-driven Image Synthesis Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The automatic transfer of a plain photo into a desired synthetic style has attracted numerous users in the application fields of photo editing, visual art, and entertainment. By … |
An Cao; Yilin Zhou; Gang Shen; | 2023 International Joint Conference on Neural Networks … | 2023-06-18 |
428 | ArtFusion: Controllable Arbitrary Style Transfer Using Dual Conditional Latent Diffusion Models Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: Arbitrary Style Transfer (AST) aims to transform images by adopting the style from any selected artwork. Nonetheless, the need to accommodate diverse and subjective user … |
Da Chen; | ArXiv | 2023-06-15 |
429 | Motion Capture Dataset for Practical Use of AI-based Motion Editing and Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we proposed a new style-diverse dataset for the domain of motion style transfer. |
Makito Kobayashi; Chen-Chieh Liao; Keito Inoue; Sentaro Yojima; Masafumi Takahashi; | arxiv-cs.CV | 2023-06-15 |
430 | ArtFusion: Arbitrary Style Transfer Using Dual Conditional Latent Diffusion Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a new approach, ArtFusion, which provides a flexible balance between content and style. |
Dar-Yen Chen; | arxiv-cs.CV | 2023-06-15 |
431 | GBSD: Generative Bokeh with Stage Diffusion Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present GBSD, the first generative text-to-image model that synthesizes photorealistic images with a bokeh style. |
Jieren Deng; Xin Zhou; Hao Tian; Zhihong Pan; Derek Aguiar; | arxiv-cs.CV | 2023-06-14 |
432 | GP-UNIT: Generative Prior for Versatile Unsupervised Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we introduce a novel versatile framework, Generative Prior-guided UNsupervised Image-to-image Translation (GP-UNIT), that improves the quality, applicability and controllability of the existing translation models. |
Shuai Yang; Liming Jiang; Ziwei Liu; Chen Change Loy; | arxiv-cs.CV | 2023-06-07 |
433 | Improving Diffusion-based Image Translation Using Asymmetric Gradient Guidance Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Yet, these methods often require computationally intense fine-tuning of diffusion models or additional neural networks. To address these challenges, here we present an approach that guides the reverse process of diffusion sampling by applying asymmetric gradient guidance. |
Gihyun Kwon; Jong Chul Ye; | arxiv-cs.CV | 2023-06-07 |
434 | Interpretable Style Transfer for Text-to-Speech with ControlVAE and Diffusion Bridge Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a new TTS system that can perform style transfer with interpretability and high fidelity. |
WENHAO GUAN et. al. | arxiv-cs.SD | 2023-06-07 |
435 | A Conditional GAN Architecture for Colorization of Thermal Infrared Images Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: The applicability of visible spectrum cameras is limited to nighttime and extreme weather conditions. To overcome these limitations, infrared (IR) cameras were introduced, but … |
Ekaagra Dubey; N. Singh; Prateek Joshi; R. Prasad; | 2023 IEEE World AI IoT Congress (AIIoT) | 2023-06-07 |
436 | Instruct-Video2Avatar: Video-to-Avatar Generation with Instructions Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose a method for synthesizing edited photo-realistic digital avatars with text instructions. |
Shaoxu Li; | arxiv-cs.CV | 2023-06-05 |
437 | Identifying The Style By A Qualified Reader on A Short Fragment of Generated Poetry Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: I used 3 character-based LSTM-models to work with style reproducing assessment. |
Boris Orekhov; | arxiv-cs.CL | 2023-06-05 |
438 | Diffusion-Enhanced PatchMatch: A Framework for Arbitrary Style Transfer with Diffusion Models IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Diffusion models have gained immense popularity in recent years due to their impressive ability to generate high-quality images. The opportunities that diffusion models provide … |
Mark Hamazaspyan; Shant Navasardyan; | 2023 IEEE/CVF Conference on Computer Vision and Pattern … | 2023-06-01 |
439 | Unsupervised Bidirectional Style Transfer Network Using Local Feature Transform Module Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In this paper, we propose a bidirectional style transfer method by exchanging the style of inputs while preserving the structural information. The proposed bidirectional style … |
K. Bae; Hyungil Kim; Y. Kwon; Jinyoung Moon; | 2023 IEEE/CVF Conference on Computer Vision and Pattern … | 2023-06-01 |
440 | DeSRF: Deformable Stylized Radiance Field Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: When stylizing 3D scenes, current methods need to render the full-resolution images from different views and use the style loss, which is proposed for 2D style transfer and needs … |
Shiyao Xu; Lingzhi Li; Li Shen; Z. Lian; | 2023 IEEE/CVF Conference on Computer Vision and Pattern … | 2023-06-01 |
441 | Gatha: Relational Loss for Enhancing Text-based Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Text-based style transfer is a promising area of research that enables the generation of stylistic images from plain text descriptions. However, the existing text-based style … |
Surgan Jandial; Shripad Deshmukh; Abhinav Java; Simra Shahid; Balaji Krishnamurthy; | 2023 IEEE/CVF Conference on Computer Vision and Pattern … | 2023-06-01 |
442 | Name Your Style: Text-guided Artistic Style Transfer Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Image style transfer has attracted widespread attention in the past years. Despite its remarkable results, it requires additional style images available as references, making it … |
Zhi-Song Liu; Li-Wen Wang; W. Siu; Vicky Kalogeiton; | 2023 IEEE/CVF Conference on Computer Vision and Pattern … | 2023-06-01 |
443 | Image Reference-guided Fashion Design with Structure-aware Transfer By Diffusion Models IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Image-based fashion design with AI techniques has attracted increasing attention in recent years. We focus on a new fashion design task, where we aim to transfer a reference … |
Shidong Cao; Wenhao Chai; Shengyu Hao; Gaoang Wang; | 2023 IEEE/CVF Conference on Computer Vision and Pattern … | 2023-06-01 |
444 | Simulation-Aided Deep Learning for Laser Ultrasonic Visualization Testing Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In recent years, laser ultrasonic visualization testing (LUVT) has attracted much attention because of its ability to efficiently perform non-contact ultrasonic non-destructive … |
Miya Nakajima; T. Saitoh; Tsuyoshi Kato; | ArXiv | 2023-05-30 |
445 | SAVE: Spectral-Shift-Aware Adaptation of Image Diffusion Models for Text-guided Video Editing Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Text-to-Image (T2I) diffusion models have achieved remarkable success in synthesizing high-quality images conditioned on text prompts. Recent methods have tried to replicate the … |
Nazmul Karim; Umar Khalid; M. Joneidi; Chen Chen; N. Rahnavard; | ArXiv | 2023-05-30 |
446 | Context-Preserving Two-Stage Video Domain Translation for Portrait Stylization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To address the issue, we propose a novel two-stage video translation framework with an objective function which enforces a model to generate a temporally coherent stylized video while preserving context in the source video. |
DOYEON KIM et. al. | arxiv-cs.CV | 2023-05-30 |
447 | Improving Open Set Domain Adaptation Using Image-to-Image Translation and Instance-Weighted Adversarial Learning Related Papers Related Patents Related Grants Related Venues Related Experts View |
Hongjie Zhang; Ang Li; Jie Guo; Yanwen Guo; | Journal of Computer Science and Technology | 2023-05-30 |
448 | SAVE: Spectral-Shift-Aware Adaptation of Image Diffusion Models for Text-driven Video Editing Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Although the latter is computationally less expensive, it still takes a significant amount of time for per-video adaption. To address this issue, we propose SAVE, a novel spectral-shift-aware adaptation framework, in which we fine-tune the spectral shift of the parameter space instead of the parameters themselves. |
Nazmul Karim; Umar Khalid; Mohsen Joneidi; Chen Chen; Nazanin Rahnavard; | arxiv-cs.CV | 2023-05-29 |
449 | Conditional Score Guidance for Text-Driven Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We present a novel algorithm for text-driven image-to-image translation based on a pretrained text-to-image diffusion model. |
Hyunsoo Lee; Minsoo Kang; Bohyung Han; | arxiv-cs.CV | 2023-05-29 |
450 | StyleS2ST: Zero-shot Style Transfer for Direct Speech-to-speech Translation Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Direct speech-to-speech translation (S2ST) has gradually become popular as it has many advantages compared with cascade S2ST. However, current research mainly focuses on the … |
KUN SONG et. al. | arxiv-cs.SD | 2023-05-28 |
451 | Image Style Transfer Based on Cyclegan Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Generative Adversarial Networks (GAN) have powerful adversarial learning capabilities and are currently being used by more and more researchers. The style transfer of images is an … |
Lisha Yao; Qiaoqiao Feng; | 2023 IEEE 3rd International Conference on Electronic … | 2023-05-26 |
452 | CLIP3Dstyler: Language Guided 3D Arbitrary Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel language-guided 3D arbitrary neural style transfer method (CLIP3Dstyler). |
MING GAO et. al. | arxiv-cs.CV | 2023-05-25 |
453 | SAMScore: A Semantic Structural Similarity Metric for Image Translation Evaluation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Traditional image-level similarity metrics are of limited use, since the semantics of an image are high-level, and not strongly governed by pixel-wise faithfulness to an original image. Towards filling this gap, we introduce SAMScore, a generic semantic structural similarity metric for evaluating the faithfulness of image translation models. |
YUNXIANG LI et. al. | arxiv-cs.CV | 2023-05-24 |
454 | Balancing Effect of Training Dataset Distribution of Multiple Styles for Multi-Style Text Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Through quantitative analysis, we explore the impact of multiple style distributions in training data on style-transferred output. |
Debarati Das; David Ma; Dongyeop Kang; | arxiv-cs.CL | 2023-05-24 |
455 | Variational Bayesian Framework for Advanced Image Generation with Domain-Related Variables Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, it remains challenging for existing methods to address advanced conditional generative problems without annotations, which can enable multiple applications like image-to-image translation and image editing. We present a unified Bayesian framework for such problems, which introduces an inference stage on latent variables within the learning process. |
Yuxiao Li; Santiago Mazuelas; Yuan Shen; | arxiv-cs.CV | 2023-05-23 |
456 | InstructVid2Vid: Controllable Video Editing with Natural Language Instructions IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce InstructVid2Vid, an end-to-end diffusion-based methodology for video editing guided by human language instructions. |
Bosheng Qin; Juncheng Li; Siliang Tang; Tat-Seng Chua; Yueting Zhuang; | arxiv-cs.CV | 2023-05-20 |
457 | Brain Captioning: Decoding Human Brain Activity Into Images and Text IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Recent breakthroughs in functional magnetic resonance imaging (fMRI) have enabled scientists to extract visual information from human brain activity patterns. In this study, we present an innovative method for decoding brain activity into meaningful images and captions, with a specific focus on brain captioning due to its enhanced flexibility as compared to brain decoding into images. |
Matteo Ferrante; Furkan Ozcelik; Tommaso Boccato; Rufin VanRullen; Nicola Toschi; | arxiv-cs.CV | 2023-05-19 |
458 | Drag Your GAN: Interactive Point-based Manipulation on The Generative Image Manifold IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we study a powerful yet much less explored way of controlling GANs, that is, to drag any points of the image to precisely reach target points in a user-interactive manner, as shown in Fig.1. |
XINGANG PAN et. al. | arxiv-cs.CV | 2023-05-18 |
459 | Domain Adaptive Sim-to-Real Segmentation of Oropharyngeal Organs Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose a domain adaptive Sim-to-Real framework called IoU-Ranking Blend-ArtFlow (IRB-AF) for image segmentation of oropharyngeal organs. |
Guankun Wang; Tian-Ao Ren; Jiewen Lai; Long Bai; Hongliang Ren; | arxiv-cs.AI | 2023-05-18 |
460 | Imagic: Text-Based Real Image Editing With Diffusion Models IF:7 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper we demonstrate, for the very first time, the ability to apply complex (e.g., non-rigid) text-based semantic edits to a single real image. |
BAHJAT KAWAR et. al. | cvpr | 2023-05-17 |
461 | Modernizing Old Photos Using Multiple References Via Photorealistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In order to modernize old photos, we propose a novel multi-reference-based old photo modernization (MROPM) framework consisting of a network MROPM-Net and a novel synthetic data generation scheme. |
Agus Gunawan; Soo Ye Kim; Hyeonjun Sim; Jae-Ho Lee; Munchurl Kim; | cvpr | 2023-05-17 |
462 | Masked and Adaptive Transformer for Exemplar Based Image Translation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present a novel framework for exemplar based image translation. |
CHANG JIANG et. al. | cvpr | 2023-05-17 |
463 | Plug-and-Play Diffusion Features for Text-Driven Image-to-Image Translation IF:6 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a new framework that takes text-to-image synthesis to the realm of image-to-image translation — given a guidance image and a target text prompt as input, our method harnesses the power of a pre-trained text-to-image diffusion model to generate a new image that complies with the target text, while preserving the semantic layout of the guidance image. |
Narek Tumanyan; Michal Geyer; Shai Bagon; Tali Dekel; | cvpr | 2023-05-17 |
464 | Tunable Convolutions With Parametric Multi-Loss Optimization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this work, we propose to optimize a parametric tunable convolutional layer, which includes a number of different kernels, using a parametric multi-loss, which includes an equal number of objectives. |
Matteo Maggioni; Thomas Tanay; Francesca Babiloni; Steven McDonagh; Aleš Leonardis; | cvpr | 2023-05-17 |
465 | Inversion-Based Style Transfer With Diffusion Models IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Specifically, we perceive style as a learnable textual description of a painting.We propose an inversion-based style transfer method (InST), which can efficiently and accurately learn the key information of an image, thus capturing and transferring the artistic style of a painting. |
YUXIN ZHANG et. al. | cvpr | 2023-05-17 |
466 | Neural Preset for Color Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a Neural Preset technique to address the limitations of existing color style transfer methods, including visual artifacts, vast memory requirement, and slow style switching speed. |
Zhanghan Ke; Yuhao Liu; Lei Zhu; Nanxuan Zhao; Rynson W.H. Lau; | cvpr | 2023-05-17 |
467 | Transforming Radiance Field With Lipschitz Network for Photorealistic 3D Scene Stylization IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Simply coupling NeRF with photorealistic style transfer (PST) will result in cross-view inconsistency and degradation of stylized view syntheses. Through a thorough analysis, we demonstrate that this non-trivial task can be simplified in a new light: When transforming the appearance representation of a pre-trained NeRF with Lipschitz mapping, the consistency and photorealism across source views will be seamlessly encoded into the syntheses. |
ZICHENG ZHANG et. al. | cvpr | 2023-05-17 |
468 | StyleRF: Zero-Shot 3D Style Transfer of Neural Radiance Fields IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose StyleRF (Style Radiance Fields), an innovative 3D style transfer technique that resolves the three-way dilemma by performing style transformation within the feature space of a radiance field. |
KUNHAO LIU et. al. | cvpr | 2023-05-17 |
469 | Learning Dynamic Style Kernels for Artistic Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To further enhance the flexibility of our style transfer method, we propose a Style Alignment Encoding (SAE) module complemented with a Content-based Gating Modulation (CGM) module for learning the dynamic style kernels in focusing regions. |
Wenju Xu; Chengjiang Long; Yongwei Nie; | cvpr | 2023-05-17 |
470 | Master: Meta Style Transformer for Controllable Zero-Shot and Few-Shot Artistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we devise a novel Transformer model termed as Master specifically for style transfer. |
HAO TANG et. al. | cvpr | 2023-05-17 |
471 | EDICT: Exact Diffusion Inversion Via Coupled Transformations IF:4 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, DDIM inversion for real images is unstable as it relies on local linearization assumptions, which result in the propagation of errors, leading to incorrect image reconstruction and loss of content. To alleviate these problems, we propose Exact Diffusion Inversion via Coupled Transformations (EDICT), an inversion method that draws inspiration from affine coupling layers. |
Bram Wallace; Akash Gokul; Nikhil Naik; | cvpr | 2023-05-17 |
472 | CAP-VSTNet: Content Affinity Preserved Versatile Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes a new framework named CAP-VSTNet, which consists of a new reversible residual network and an unbiased linear transform module, for versatile style transfer. |
Linfeng Wen; Chengying Gao; Changqing Zou; | cvpr | 2023-05-17 |
473 | Ref-NPR: Reference-Based Non-Photorealistic Radiance Fields for Controllable Scene Stylization IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Current 3D scene stylization methods transfer textures and colors as styles using arbitrary style references, lacking meaningful semantic correspondences. We introduce Reference-Based Non-Photorealistic Radiance Fields (Ref-NPR) to address this limitation. |
Yuechen Zhang; Zexin He; Jinbo Xing; Xufeng Yao; Jiaya Jia; | cvpr | 2023-05-17 |
474 | BBDM: Image-to-Image Translation With Brownian Bridge Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, a novel image-to-image translation method based on the Brownian Bridge Diffusion Model(BBDM) is proposed, which models image-to-image translation as a stochastic Brownian Bridge process, and learns the translation between two domains directly through the bidirectional diffusion process rather than a conditional generation process. |
Bo Li; Kaitao Xue; Bin Liu; Yu-Kun Lai; | cvpr | 2023-05-17 |
475 | Style Transfer Enabled Sim2Real Framework for Efficient Learning of Robotic Ultrasound Image Analysis Using Simulated Data Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This work presents a Sim2Real framework to efficiently learn robotic US image analysis tasks based only on simulated data for real-world deployment. |
KEYU LI et. al. | arxiv-cs.RO | 2023-05-16 |
476 | Wavelet-based Unsupervised Label-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: State-of-the-art conditional Generative Adversarial Networks (GANs) need a huge amount of paired data to accomplish this task while generic unpaired image-to-image translation frameworks underperform in comparison, because they color-code semantic layouts and learn correspondences in appearance instead of semantic content. Starting from the assumption that a high quality generated image should be segmented back to its semantic layout, we propose a new Unsupervised paradigm for SIS (USIS) that makes use of a self-supervised segmentation loss and whole image wavelet based discrimination. |
George Eskandar; Mohamed Abdelsamad; Karim Armanious; Shuai Zhang; Bin Yang; | arxiv-cs.CV | 2023-05-16 |
477 | Realization RGBD Image Stylization Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel method that incorporates the depth map and a heatmap of the RGB image to generate more realistic style transfer results. |
Bhavya Sehgal; Vaishnavi Mendu; Aparna Mendu; | arxiv-cs.CV | 2023-05-11 |
478 | Analysis of Adversarial Image Manipulations Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper investigates how simple, accessible image manipulation techniques affect the accuracy of facial recognition software in identifying an individual’s various face images based on one unique image. |
Ahsi Lo; Gabriella Pangelinan; Michael C. King; | arxiv-cs.CV | 2023-05-10 |
479 | Style-A-Video: Agile Diffusion for Arbitrary Text-based Video Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: This paper proposes a zero-shot video stylization method named Style-A-Video, which utilizes a generative pre-trained transformer with an image latent diffusion model to achieve a concise text-controlled video stylization. |
Nisha Huang; Yuxin Zhang; Weiming Dong; | arxiv-cs.CV | 2023-05-09 |
480 | Joint Multiscale Cross-Lingual Speaking Style Transfer With Bidirectional Attention Mechanism for Automatic Dubbing Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Automatic dubbing, which generates a corresponding version of the input speech in another language, can be widely utilized in many real-world scenarios, such as video and game … |
JINGBEI LI et. al. | IEEE/ACM Transactions on Audio, Speech, and Language … | 2023-05-09 |
481 | Multi-Teacher Knowledge Distillation For Text Image Machine Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a novel Multi-Teacher Knowledge Distillation (MTKD) method to effectively distillate knowledge into the end-to-end TIMT model from the pipeline model. |
CONG MA et. al. | arxiv-cs.CL | 2023-05-09 |
482 | Joint Multi-scale Cross-lingual Speaking Style Transfer with Bidirectional Attention Mechanism for Automatic Dubbing Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a joint multi-scale cross-lingual speaking style transfer framework to simultaneously model the bidirectional speaking style transfer between languages at both global (i.e. utterance level) and local (i.e. word level) scales. |
JINGBEI LI et. al. | arxiv-cs.SD | 2023-05-09 |
483 | Image Neural Style Transfer: A Review IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
Qiang Cai; Mengxu Ma; Chen Wang; Haisheng Li; | Comput. Electr. Eng. | 2023-05-01 |
484 | HRInversion: High-Resolution GAN Inversion for Cross-Domain Image Synthesis IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: We investigate GAN inversion problems of using pre-trained GANs to reconstruct real images. Recent methods for such problems typically employ a VGG perceptual loss to measure the … |
Peng Zhou; Lingxi Xie; Bingbing Ni; Lin Liu; Qi Tian; | IEEE Transactions on Circuits and Systems for Video … | 2023-05-01 |
485 | Physical Model and Image Translation Fused Network for Single-image Dehazing IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
Yanzhao Su; Chuan He; Zhigao Cui; A. Li; Nian Wang; | Pattern Recognit. | 2023-05-01 |
486 | Image-to-Image Translation for Data Augmentation on Multimodal Medical Images Related Papers Related Patents Related Grants Related Venues Related Experts View |
Yue Peng; Zuqiang Meng; Lina Yang; | IEICE Trans. Inf. Syst. | 2023-05-01 |
487 | Duetcs: Code Style Transfer Through Generation and Retrieval Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Coding style has direct impact on code comprehension. Automatically transferring code style to user’s preference or consistency can facilitate project cooperation and maintenance, … |
Binger Chen; Ziawasch Abedjan; | 2023 IEEE/ACM 45th International Conference on Software … | 2023-05-01 |
488 | Transplayer: Timbre Style Transfer with Flexible Timbre Control Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Inspired by the practice in voice conversion, we propose TransPlayer, which uses an autoencoder model with one-hot representations of instruments as the condition, and a Diffwave model trained especially for music synthesis. |
Y. Wu; Y. He; X. Liu; Y. Wang; R. B. Dannenberg; | icassp | 2023-04-27 |
489 | MSNet: A Deep Architecture Using Multi-Sentiment Semantics for Sentiment-Aware Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To incorporate the sentiment information into the image style transfer task for better sentiment-aware performance, we introduce a new task named sentiment-aware image style transfer. |
S. Sun; J. Jia; H. Wu; Z. Ye; J. Xing; | icassp | 2023-04-27 |
490 | CPD-GAN: Cascaded Pyramid Deformation GAN for Pose Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Existing work often failed to transfer complex textures to generated images well. To solve this problem, we propose a novel network for this task. |
Y. Huang; Y. Tang; X. Zheng; J. Tang; | icassp | 2023-04-27 |
491 | OSAGGAN: One-shot Unsupervised Image-to-image Translation Using Attention-guided Generative Adversarial Networks Related Papers Related Patents Related Grants Related Venues Related Experts View |
Xiaofei Huo; Bin Jiang; Haotian Hu; Xinjiao Zhou; Bolin Zhang; | International Journal of Machine Learning and Cybernetics | 2023-04-27 |
492 | Multidimensional Evaluation for Text Style Transfer Using ChatGPT IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We perform a comprehensive correlation analysis for two transfer directions (and overall) at different levels. |
Huiyuan Lai; Antonio Toral; Malvina Nissim; | arxiv-cs.CL | 2023-04-26 |
493 | Hierarchical Diffusion Autoencoders and Disentangled Image Manipulation IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: To mitigate those limitations, we propose Hierarchical Diffusion Autoencoders (HDAE) that exploit the fine-grained-to-abstract and lowlevel-to-high-level feature hierarchy for the latent space of diffusion models. |
ZEYU LU et. al. | arxiv-cs.CV | 2023-04-24 |
494 | Unsupervised Style-based Explicit 3D Face Reconstruction from Single Image Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this work, we propose a general adversarial learning framework for solving Unsupervised 2D to Explicit 3D Style Transfer (UE3DST). |
Heng Yu; Zoltan A. Milacski; Laszlo A. Jeni; | arxiv-cs.CV | 2023-04-24 |
495 | InvolutionGAN: Lightweight GAN with Involution for Unsupervised Image-to-image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View |
Haipeng Deng; Qiuxia Wu; Han Huang; Xiaowei Yang; Zhiyong Wang; | Neural Computing and Applications | 2023-04-24 |
496 | Aesthetic Style Transferring Method Based on Deep Neural Network Between Chinese Landscape Painting and Classical Private Garden’s Virtual Scenario Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Most of the existing virtual scenarios built for the digital protection of Chinese classical private gardens are too modern in expression style to show the aesthetic significance … |
SHUAI HONG et. al. | International Journal of Digital Earth | 2023-04-23 |
497 | Spectral Normalization and Dual Contrastive Regularization for Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we propose a new unpaired I2I translation framework based on dual contrastive regularization and spectral normalization, namely SN-DCR. |
Chen Zhao; Wei-Ling Cai; Zheng Yuan; | arxiv-cs.CV | 2023-04-22 |
498 | Arbitrary Style Transfer with Multiple Self-Attention Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Style transfer aims to transfer the style information of a given style image to the other images, but most existing methods cannot transfer the texture details in style images … |
Yuzhu Song; Li Liu; Huaxiang Zhang; Dongmei Liu; Hongzhen Li; | Proceedings of the 2023 8th International Conference on … | 2023-04-21 |
499 | A Plug-and-Play Defensive Perturbation for Copyright Protection of DNN-based Applications Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose a novel plug-and-play invisible copyright protection method based on defensive perturbation for DNN-based applications (i.e., style transfer). |
DONGHUA WANG et. al. | arxiv-cs.CV | 2023-04-20 |
500 | Any-to-Any Style Transfer: Making Picasso and Da Vinci Collaborate IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In either case, only one result can be generated for a specific pair of content and style images, which therefore lacks flexibility and is hard to satisfy different users with different preferences. We propose here a novel strategy termed Any-to-Any Style Transfer to address this drawback, which enables users to interactively select styles of regions in the style image and apply them to the prescribed content regions. |
Songhua Liu; Jingwen Ye; Xinchao Wang; | arxiv-cs.CV | 2023-04-19 |
501 | UPGPT: Universal Diffusion Model for Person Image Generation, Editing and Pose Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: The appearance inconsistency makes T2I unsuitable for pose transfer. We address this by proposing a multimodal diffusion model that accepts text, pose, and visual prompting. |
Soon Yau Cheong; Armin Mustafa; Andrew Gilbert; | arxiv-cs.CV | 2023-04-18 |
502 | ALADIN-NST: Self-supervised Disentangled Representation Learning of Artistic Style Through Neural Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Our paper aims to learn a representation of visual artistic style more strongly disentangled from the semantic content depicted in an image. |
Dan Ruta; Gemma Canet Tarres; Alexander Black; Andrew Gilbert; John Collomosse; | arxiv-cs.CV | 2023-04-12 |
503 | Improving Diffusion Models for Scene Text Editing with Dual Encoders IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, our empirical analysis reveals that state-of-the-art diffusion models struggle with rendering correct text and controlling text style. To address these problems, we propose DIFFSTE to improve pre-trained diffusion models with a dual encoder design, which includes a character encoder for better text legibility and an instruction encoder for better style control. |
JIABAO JI et. al. | arxiv-cs.CV | 2023-04-11 |
504 | Panoramic Image-to-Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we tackle the challenging task of Panoramic Image-to-Image translation (Pano-I2I) for the first time. |
SOOHYUN KIM et. al. | arxiv-cs.CV | 2023-04-11 |
505 | NeAT: Neural Artistic Tracing for Beautiful Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present NeAT, a new state-of-the art feed-forward style transfer method. |
Dan Ruta; Andrew Gilbert; John Collomosse; Eli Shechtman; Nicholas Kolkin; | arxiv-cs.CV | 2023-04-11 |
506 | DDRF: Denoising Diffusion Model for Remote Sensing Image Fusion IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Denosing diffusion model, as a generative model, has received a lot of attention in the field of image generation recently, thanks to its powerful generation capability. However, … |
ZIHAN CAO et. al. | ArXiv | 2023-04-10 |
507 | ITportrait: Image-Text Coupled 3D Portrait Domain Adaptation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we propose an Image-Text multi-modal framework, namely Image and Text portrait (ITportrait), for 3D portrait domain adaptation. |
XIANGWEN DENG et. al. | arxiv-cs.MM | 2023-04-09 |
508 | SAM-GAN: Supervised Learning-Based Aerial Image-to-Map Translation Via Generative Adversarial Networks Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Accurate translation of aerial imagery to maps is a direction of great value and challenge in mapping, a method of generating maps that does not require using vector data as … |
Jian Xu; Xiaowen Zhou; Chaolin Han; Bing Dong; Hongwei Li; | ISPRS Int. J. Geo Inf. | 2023-04-07 |
509 | A CNN Inference Accelerator on FPGA With Compression and Layer-Chaining Techniques for Style Transfer Applications Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Recently, convolutional neural networks (CNNs) have actively been applied to computer vision applications such as style transfer that changes the style of a content image into … |
SUCHANG KIM et. al. | IEEE Transactions on Circuits and Systems I: Regular Papers | 2023-04-01 |
510 | Fake Colorized Image Detection Based on Special Image Representation and Transfer Learning Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Nowadays, images have become one of the most popular forms of communication as image editing tools have evolved. Image manipulation, particularly image colorization, has become … |
Khalid A. Salman; Khalid Shaker; Sufyan T. Faraj Al-Janabi; | Int. J. Comput. Intell. Appl. | 2023-04-01 |
511 | Towards Spatially Disentangled Manipulation of Face Images With Pre-Trained StyleGANs IF:3 Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Generative Adversarial Networks with style-based generators could successfully synthesize realistic images from input latent code. Moreover, recent studies have revealed that … |
Yunfan Liu; Qi Li; Qiyao Deng; Zhenan Sun; | IEEE Transactions on Circuits and Systems for Video … | 2023-04-01 |
512 | Unpaired Image-to-image Translation of Structural Damage IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts View |
Subin Varghese; Vedhus Hoskere; | Adv. Eng. Informatics | 2023-04-01 |
513 | One-shot Unsupervised Domain Adaptation with Personalized Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Departing from the common notion of transferring only the target “texture” information, we leverage text-to-image diffusion models (e.g., Stable Diffusion) to generate a synthetic target dataset with photo-realistic images that not only faithfully depict the style of the target domain, but are also characterized by novel scenes in diverse contexts. |
Yasser Benigmim; Subhankar Roy; Slim Essid; Vicky Kalogeiton; Stéphane Lathuilière; | arxiv-cs.CV | 2023-03-31 |
514 | Semantic Image Translation for Repairing The Texture Defects of Building Models Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In order to preserve fine details and regular structures, we propose a regularity-aware multi-domain method that capitalizes on frequency information and corner maps. |
QISEN SHANG et. al. | arxiv-cs.CV | 2023-03-30 |
515 | Instant Photorealistic Neural Radiance Fields Stylization Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present Instant Neural Radiance Fields Stylization, a novel approach for multi-view image stylization for the 3D scene. |
Shaoxu Li; Ye Pan; | arxiv-cs.CV | 2023-03-29 |
516 | Depth-Aware Neural Style Transfer for Videos Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Temporal consistency and content preservation are the prominent challenges in artistic video style transfer. To address these challenges, we present a technique that utilizes … |
E. Ioannou; S. Maddock; | Comput. | 2023-03-27 |
517 | Linear-ResNet GAN-based Anime Style Transfer of Face Images Related Papers Related Patents Related Grants Related Venues Related Experts View |
Mingxi Chen; Hansen Dai; Shijie Wei; Zhenzhen Hu; | Signal, Image and Video Processing | 2023-03-23 |
518 | Neural Preset for Color Style Transfer IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a Neural Preset technique to address the limitations of existing color style transfer methods, including visual artifacts, vast memory requirement, and slow style switching speed. |
Zhanghan Ke; Yuhao Liu; Lei Zhu; Nanxuan Zhao; Rynson W. H. Lau; | arxiv-cs.CV | 2023-03-23 |
519 | Open-World Pose Transfer Via Sequential Test-Time Adaption Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: A typical pose transfer framework usually employs representative datasets to train a discriminative model, which is often violated by out-of-distribution (OOD) instances. |
JUNYANG CHEN et. al. | arxiv-cs.CV | 2023-03-20 |
520 | StyleRF: Zero-shot 3D Style Transfer of Neural Radiance Fields IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose StyleRF (Style Radiance Fields), an innovative 3D style transfer technique that resolves the three-way dilemma by performing style transformation within the feature space of a radiance field. |
KUNHAO LIU et. al. | arxiv-cs.CV | 2023-03-19 |
521 | Style Transfer for 2D Talking Head Animation Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present a new method to generate talking head animation with learnable style references. |
TRONG-THANG PHAM et. al. | arxiv-cs.CV | 2023-03-17 |
522 | DialogPaint: A Dialog-based Image Editing Model Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We introduce DialogPaint, a novel framework that bridges conversational interactions with image editing, enabling users to modify images through natural dialogue. |
Jingxuan Wei; Shiyu Wu; Xin Jiang; Yequan Wang; | arxiv-cs.CV | 2023-03-17 |
523 | NLUT: Neural-based 3D Lookup Tables for Video Photorealistic Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, existing methods obtain stylized video sequences by performing frame-by-frame photorealistic style transfer, which is inefficient and does not ensure the temporal consistency of the stylized video. To address this issue, we use neural network-based 3D Lookup Tables (LUTs) for the photorealistic transfer of videos, achieving a balance between efficiency and effectiveness. |
YAOSEN CHEN et. al. | arxiv-cs.CV | 2023-03-16 |
524 | SpectralCLIP: Preventing Artifacts in Text-Guided Style Transfer from A Spectral Perspective Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We propose SpectralCLIP, which implements a spectral filtering layer on top of the CLIP vision encoder, to alleviate the artifact issue. |
Zipeng Xu; Songlong Xing; Enver Sangineto; Nicu Sebe; | arxiv-cs.CV | 2023-03-16 |
525 | StylerDALLE: Language-Guided Style Transfer Using A Vector-Quantized Tokenizer of A Large-Scale Generative Model IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, these abstract semantics can be captured by models like DALL-E or CLIP, which have been trained using huge datasets of images and textual documents. In this paper, we propose StylerDALLE, a style transfer method that exploits both of these models and uses natural language to describe abstract art styles. |
Zipeng Xu; Enver Sangineto; Nicu Sebe; | arxiv-cs.CV | 2023-03-16 |
526 | Class-Guided Image-to-Image Diffusion: Cell Painting from Brightfield Images with Class Labels Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We introduce and implement a model which combines image-to-image and class-guided denoising diffusion probabilistic models. |
JAN OSCAR CROSS-ZAMIRSKI et. al. | arxiv-cs.CV | 2023-03-15 |
527 | 3D Face Arbitrary Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, previous methods mainly use images of artistic faces for style transfer while ignoring arbitrary style images such as abstract paintings. To solve this problem, we propose a novel method, namely Face-guided Dual Style Transfer (FDST). |
XIANGWEN DENG et. al. | arxiv-cs.CV | 2023-03-14 |
528 | PADAAV: Enhancing Perception Systems Using GAN-generated Adversarial Augmented Domains for Autonomous Vehicles Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: In the field of autonomous vehicles (AV), it is crucial for the perceptual systems of the AVs to learn inter-domain adaptations in the absence of paired examples for detecting … |
Oshin Rawlley; Shashank Gupta; | 2023 IEEE International Conference on Pervasive Computing … | 2023-03-13 |
529 | SEM-CS: Semantic CLIPStyler for Text-Based Image Style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, the ground semantics of objects in style transfer output is lost due to style spillover on salient and background objects (content mismatch) or over-stylization. To solve this, we propose Semantic CLIPStyler (Sem-CS) that performs semantic style transfer. |
Chanda G Kamra; Indra Deep Mastan; Debayan Gupta; | arxiv-cs.CV | 2023-03-11 |
530 | AptSim2Real: Approximately-Paired Sim-to-Real Image Translation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Unpaired image translation, while more suitable for sim-to-real transfer, is still challenging to learn for complex natural scenes. To address these challenges, we propose a third category: approximately-paired sim-to-real translation, where the source and target images do not need to be exactly paired. |
Charles Y Zhang; Ashish Shrivastava; | arxiv-cs.CV | 2023-03-09 |
531 | Classification of Crop Leaf Diseases Using Image to Image Translation with Deep-dream Related Papers Related Patents Related Grants Related Venues Related Experts View |
Priyanka Sahu; A. Chug; Ashutosh Kumar Singh; Dinesh Singh; | Multimedia Tools and Applications | 2023-03-08 |
532 | A Unified Arbitrary Style Transfer Framework Via Adaptive Contrastive Learning IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: We present Unified Contrastive Arbitrary Style Transfer (UCAST), a novel style representation learning and transfer framework, which can fit in most existing arbitrary image style transfer models, e.g., CNN-based, ViT-based, and flow-based methods. |
YUXIN ZHANG et. al. | arxiv-cs.CV | 2023-03-08 |
533 | End-to-end Face-swapping Via Adaptive Latent Representation Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This paper proposes a novel and end-to-end integrated framework for high resolution and attribute preservation face swapping via Adaptive Latent Representation Learning. |
Chenhao Lin; Pengbin Hu; Chao Shen; Qian Li; | arxiv-cs.CV | 2023-03-07 |
534 | Neural Style Transfer for Vector Graphics Summary Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Abstract: Neural style transfer draws researchers’ attention, but the interest focuses on bitmap images. Various models have been developed for bitmap image generation both online and … |
V. Efimova; Artyom Chebykin; Ivan Jarsky; Evgenii Prosvirnin; A. Filchenkov; | ArXiv | 2023-03-06 |
535 | Guided Image-to-Image Translation By Discriminator-Generator Communication Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: This formulation illustrates the information insufficiency in the GAN training. To mitigate this problem, we propose to add a communication channel between discriminators and generators. |
Yuanjiang Cao; Lina Yao; Le Pan; Quan Z. Sheng; Xiaojun Chang; | arxiv-cs.CV | 2023-03-06 |
536 | Image-to-image Translation Using An Offset-based Multi-scale Codes GAN Encoder Related Papers Related Patents Related Grants Related Venues Related Experts View |
Zihao Guo; Mingwen Shao; Shunhang Li; | The Visual Computer | 2023-03-04 |
537 | Stylized Image Denoising Via Noise Style Transfer and Quasi Siamese Network Related Papers Related Patents Related Grants Related Venues Related Experts View |
Jikang Cheng; Zhen Han; Zhongyuan Wang; | Signal Process. Image Commun. | 2023-03-01 |
538 | Enhanced Controllability of Diffusion Models Via Feature Disentanglement and Realism-Enhanced Sampling Methods Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: In this paper, we present a training framework for feature disentanglement of Diffusion Models (FDiff). |
WONWOONG CHO et. al. | arxiv-cs.CV | 2023-02-28 |
539 | Cross-modal Face- and Voice-style Transfer Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a cross-modal style transfer framework called XFaVoT that jointly learns four tasks: image translation and voice conversion tasks with audio or image guidance, which enables the generation of “face that matches given voice and “voice that matches given face, and intra-modality translation tasks with a single framework. |
Naoya Takahashi; Mayank K. Singh; Yuki Mitsufuji; | arxiv-cs.CV | 2023-02-27 |
540 | Multi-Modal Multi-Stage Underwater Side-Scan Sonar Target Recognition Based on Synthetic Images Summary Related Papers Related Patents Related Grants Related Venues Related Experts View Abstract: Due to the small sample size of underwater acoustic data and the strong noise interference caused by seabed reverberation, recognizing underwater targets in Side-Scan Sonar (SSS) … |
Jian Wang; Haisen S. Li; Guanying Huo; Chao Li; Yuhang Wei; | Remote. Sens. | 2023-02-26 |
541 | ACE: Zero-Shot Image to Image Translation Via Pretrained Auto-Contrastive-Encoder Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: However, learning such mapping between domains is challenging because data from different domains can be highly unbalanced in terms of both quality and quantity. To address this problem, we propose a new approach to extract image features by learning the similarities and differences of samples within the same data distribution via a novel contrastive learning framework, which we call Auto-Contrastive-Encoder (ACE). |
Sihan Xu; Zelong Jiang; Ruisi Liu; Kaikai Yang; Zhijie Huang; | arxiv-cs.CV | 2023-02-22 |
542 | Paint It Black: Generating Paintings from Text Descriptions Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: However, the intersection of these two, i.e., generating paintings from a given caption, is a relatively unexplored area with little data available. In this paper, we have explored two distinct strategies and have integrated them together. |
Mahnoor Shahid; Mark Koch; Niklas Schneider; | arxiv-cs.CV | 2023-02-17 |
543 | Conversation Style Transfer Using Few-Shot Learning Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: We propose a novel in-context learning approach to solve the task with style-free dialogues as a pivot. |
SHAMIK ROY et. al. | arxiv-cs.CL | 2023-02-16 |
544 | DiffFashion: Reference-based Fashion Design with Structure-aware Transfer By Diffusion Models IF:3 Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: Although diffusion-based image translation or neural style transfer (NST) has enabled flexible style transfer, it is often difficult to maintain the original structure of the image realistically during the reverse diffusion, especially when the referenced appearance image greatly differs from the common clothing appearance. To tackle this issue, we present a novel diffusion model-based unsupervised structure-aware transfer method to semantically generate new clothes from a given clothing image and a reference appearance image. |
SHIDONG CAO et. al. | arxiv-cs.CV | 2023-02-13 |
545 | Unified Vision-Language Representation Modeling for E-Commerce Same-Style Products Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Common methods use the image as the detected object, but they only consider the visual features and overlook the attribute information contained in the textual descriptions, and perform weakly for products in image less important industries like machinery, hardware tools and electronic component, even if an additional text matching module is added. In this paper, we propose a unified vision-language modeling method for e-commerce same-style products retrieval, which is designed to represent one product with its textual descriptions and visual contents. |
BEN CHEN et. al. | arxiv-cs.IR | 2023-02-10 |
546 | Neural Artistic Style Transfer with Conditional Adversaria Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: In this paper, we present two methods that step toward the style image independent neural style transfer model. |
P. N. Deelaka; | arxiv-cs.CV | 2023-02-07 |
547 | ReDi: Efficient Learning-Free Diffusion Inference Via Trajectory Retrieval Related Papers Related Patents Related Grants Related Venues Related Experts Related Code View Highlight: To accelerate the inference, we propose ReDi, a simple yet learning-free Retrieval-based Diffusion sampling framework. |
Kexun Zhang; Xianjun Yang; William Yang Wang; Lei Li; | arxiv-cs.CV | 2023-02-04 |
548 | Design Booster: A Text-Guided Diffusion Model for Image Translation with Spatial Layout Preservation Related Papers Related Patents Related Grants Related Venues Related Experts View Highlight: Besides, existing methods are mainly based on test-time optimization or fine-tuning model for each input image, which are extremely time-consuming for practical applications. To address these issues, we propose a new approach for flexible image translation by learning a layout-aware image condition together with a text condition. |
Shiqi Sun; Shancheng Fang; Qian He; Wei Liu; | arxiv-cs.CV | 2023-02-04 |
549 | Multi-domain Image Generation and Translation with Identifiability Guarantees IF:3 Related Papers Related Patents Related Grants |