StyleDomain VS DeceiveD

Compare StyleDomain vs DeceiveD and see what are their differences.

StyleDomain

Official Implementation for "StyleDomain: Efficient and Lightweight Parameterizations of StyleGAN for One-shot and Few-shot Domain Adaptation" (ICCV 2023) (by AIRI-Institute)

DeceiveD

[NeurIPS 2021] Deceive D: Adaptive Pseudo Augmentation for GAN Training with Limited Data (by EndlessSora)
InfluxDB - Power Real-Time Data Analytics at Scale
Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
www.influxdata.com
featured
SaaSHub - Software Alternatives and Reviews
SaaSHub helps you find the best software and product alternatives
www.saashub.com
featured
StyleDomain DeceiveD
1 1
23 249
- -
6.4 0.0
5 months ago over 2 years ago
Python Python
- GNU General Public License v3.0 or later
The number of mentions indicates the total number of mentions that we've tracked plus the number of user suggested alternatives.
Stars - the number of stars that a project has on GitHub. Growth - month over month growth in stars.
Activity is a relative number indicating how actively a project is being developed. Recent commits have higher weight than older ones.
For example, an activity of 9.0 indicates that a project is amongst the top 10% of the most actively developed projects that we are tracking.

StyleDomain

Posts with mentions or reviews of StyleDomain. We have used some of these posts to build our list of alternatives and similar projects.
  • [Research] Exciting New Paper on StyleGAN Domain Adaptation: StyleDomain - ICCV 2023
    1 project | /r/MachineLearning | 30 Sep 2023
    Abstract: Domain adaptation of GANs is a problem of fine-tuning GAN models pretrained on a large dataset (e.g., StyleGAN) to a specific domain with few samples (e.g., painting faces, sketches, etc.). While there are many methods that tackle this problem in different ways, there are still many important questions that remain unanswered. In this paper, we provide a systematic and in-depth analysis of the domain adaptation problem of GANs, focusing on the StyleGAN model. We perform a detailed exploration of the most important parts of StyleGAN that are responsible for adapting the generator to a new domain depending on the similarity between the source and target domains. As a result of this study, we propose new efficient and lightweight parameterizations of StyleGAN for domain adaptation. Particularly, we show that there exist directions in StyleSpace (StyleDomain directions) that are sufficient for adapting to similar domains. For dissimilar domains, we propose Affine+ and AffineLight+ parameterizations that allow us to outperform existing baselines in few-shot adaptation while having significantly fewer training parameters. Finally, we examine StyleDomain directions and discover their many surprising properties that we apply for domain mixing and cross-domain image morphing. Source code can be found at GitHub.

DeceiveD

Posts with mentions or reviews of DeceiveD. We have used some of these posts to build our list of alternatives and similar projects.

What are some alternatives?

When comparing StyleDomain and DeceiveD you can also consider the following projects:

Transfer-Learning-Library - Transfer Learning Library for Domain Adaptation, Task Adaptation, and Domain Generalization

pytorch-CycleGAN-and-pix2pix - Image-to-Image Translation in PyTorch

MotionBERT - [ICCV 2023] PyTorch Implementation of "MotionBERT: A Unified Perspective on Learning Human Motion Representations"

ArtGAN - ArtGAN + WikiArt: This work presents a series of new approaches to improve GAN for conditional image synthesis and we name the proposed model as “ArtGAN”.

ALAE - [CVPR2020] Adversarial Latent Autoencoders

anycost-gan - [CVPR 2021] Anycost GANs for Interactive Image Synthesis and Editing

MobileStyleGAN.pytorch - An official implementation of MobileStyleGAN in PyTorch

RelayDiffusion - The official implementation of "Relay Diffusion: Unifying diffusion process across resolutions for image synthesis" [ICLR 2024 Spotlight]