forked from POSTECH-CVLab/PyTorch-StudioGAN
-
Notifications
You must be signed in to change notification settings - Fork 0
/
StyleGAN2-D2DCE-ADA.yaml
67 lines (66 loc) · 1.72 KB
/
StyleGAN2-D2DCE-ADA.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
# Guidelines for StyleGAN2-D2DCE-ADA config
# g_cond_mtd can be changed btw ["W/O", "cAdaIN"]
# d_cond_mtd can be changed btw ["W/O", "AC", "PD", "MH", "MD", "2C", "D2DCE", "SPD"]
# z_dim, w_dim should be fixed to 512 regardless of image size.
# apply_g_ema should be true for stable results.
# apply_r1_reg should be true.
# g_reg_interval, d_reg_interval is fixed to 4, 16 regardless of image size.
# pl_reg is disabled only for cifar10.
# d_architecture is 'orig' for cifar10.
# style_mixing_p should be 0.9 for all settings except for cifar10 (0)
# for total_step, batch_size, d_epilogue_mbstd_group_size, g/d_lr, r1_lambda, g_ema_kimg
# g_ema_rampup, mapping_network, check StyleGAN2 section in src/config.py
DATA:
name: "AFHQ"
img_size: 512
num_classes: 3
MODEL:
backbone: "stylegan2"
g_cond_mtd: "cAdaIN"
d_cond_mtd: "D2DCE"
normalize_d_embed: True
d_embed_dim: 1024
g_act_fn: "Auto"
d_act_fn: "Auto"
z_prior: "gaussian"
z_dim: 512
w_dim: 512
g_conv_dim: "N/A"
d_conv_dim: "N/A"
apply_g_ema: True
LOSS:
adv_loss: "logistic"
cond_lambda: 0.5
m_p: 0.95
temperature: 0.5
apply_r1_reg: True
r1_lambda: 0.5
OPTIMIZATION:
# These values will be slightly changed if lazy regularization is applied.
batch_size: 64
acml_steps: 1
g_lr: 0.0025
d_lr: 0.0025
beta1: 0
beta2: 0.99
g_updates_per_step: 1
d_updates_per_step: 2
total_steps: 200000
AUG:
apply_ada: True
ada_aug_type: "bgc"
ada_initial_augment_p: 0
ada_target: 0.8
ada_kimg: 500
ada_interval: 4
STYLEGAN2:
g_reg_interval: 4
d_reg_interval: 16
mapping_network: 8
style_mixing_p: 0.9
g_ema_kimg: 20
g_ema_rampup: "N/A"
apply_pl_reg: True
pl_weight: 2
d_architecture: "resnet"
d_epilogue_mbstd_group_size: 8