UPstud commited on
Commit
85ab3f2
Β·
verified Β·
1 Parent(s): e05b8ed

Upload 2 files

Browse files
models/cldm_v15_inpainting_infer.yaml ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ target: cldm.cldm.ControlLDM
3
+ params:
4
+ linear_start: 0.00085
5
+ linear_end: 0.0120
6
+ num_timesteps_cond: 1
7
+ log_every_t: 200
8
+ timesteps: 1000
9
+ first_stage_key: "jpg"
10
+ cond_stage_key: "txt"
11
+ control_key: "hint"
12
+ masked_image: "mask_img"
13
+ mask: "mask"
14
+ image_size: 64
15
+ channels: 4
16
+ cond_stage_trainable: false
17
+ conditioning_key: crossattn
18
+ monitor: val/loss_simple_ema
19
+ scale_factor: 0.18215
20
+ use_ema: False
21
+ only_mid_control: False
22
+ load_loss: False
23
+
24
+ control_stage_config:
25
+ target: cldm.cldm.ControlNet
26
+ params:
27
+ image_size: 32 # unused
28
+ in_channels: 4
29
+ hint_channels: 3
30
+ model_channels: 320
31
+ attention_resolutions: [ 4, 2, 1 ]
32
+ num_res_blocks: 2
33
+ channel_mult: [ 1, 2, 4, 4 ]
34
+ num_heads: 8
35
+ use_spatial_transformer: True
36
+ transformer_depth: 1
37
+ context_dim: 768
38
+ use_checkpoint: True
39
+ legacy: False
40
+
41
+ unet_config:
42
+ target: cldm.cldm.ControlledUnetModel
43
+ params:
44
+ image_size: 32 # unused
45
+ in_channels: 9
46
+ out_channels: 4
47
+ model_channels: 320
48
+ attention_resolutions: [ 4, 2, 1 ]
49
+ num_res_blocks: 2
50
+ channel_mult: [ 1, 2, 4, 4 ]
51
+ num_heads: 8
52
+ use_spatial_transformer: True
53
+ transformer_depth: 1
54
+ context_dim: 768
55
+ use_checkpoint: True
56
+ legacy: False
57
+
58
+ first_stage_config:
59
+ target: ldm.models.autoencoder.AutoencoderKL
60
+ params:
61
+ embed_dim: 4
62
+ monitor: val/rec_loss
63
+ ddconfig:
64
+ double_z: true
65
+ z_channels: 4
66
+ resolution: 256
67
+ in_channels: 3
68
+ out_ch: 3
69
+ ch: 128
70
+ ch_mult:
71
+ - 1
72
+ - 2
73
+ - 4
74
+ - 4
75
+ num_res_blocks: 2
76
+ attn_resolutions: []
77
+ dropout: 0.0
78
+ lossconfig:
79
+ target: torch.nn.Identity
80
+
81
+ contextual_stage_config:
82
+ target: models_deep_exp.NonlocalNet.VGG19_pytorch
83
+
84
+ cond_stage_config:
85
+ # target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
86
+ target: ldm.modules.encoders.modules.FrozenCLIPDualEmbedder
87
+ #ldm.modules.encoders.modules.FrozenCLIPDualEmbedder
models/cldm_v15_inpainting_infer1.yaml ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ target: cldm.cldm.ControlLDM
3
+ params:
4
+ linear_start: 0.00085
5
+ linear_end: 0.0120
6
+ num_timesteps_cond: 1
7
+ log_every_t: 200
8
+ timesteps: 1000
9
+ first_stage_key: "jpg"
10
+ cond_stage_key: "txt"
11
+ control_key: "hint"
12
+ masked_image: "mask_img"
13
+ mask: "mask"
14
+ image_size: 64
15
+ channels: 4
16
+ cond_stage_trainable: false
17
+ conditioning_key: crossattn
18
+ monitor: val/loss_simple_ema
19
+ scale_factor: 0.18215
20
+ use_ema: False
21
+ only_mid_control: False
22
+ load_loss: False
23
+
24
+ control_stage_config:
25
+ target: cldm.cldm.ControlNet
26
+ params:
27
+ image_size: 32 # unused
28
+ in_channels: 4
29
+ hint_channels: 3
30
+ model_channels: 320
31
+ attention_resolutions: [ 4, 2, 1 ]
32
+ num_res_blocks: 2
33
+ channel_mult: [ 1, 2, 4, 4 ]
34
+ num_heads: 8
35
+ use_spatial_transformer: True
36
+ transformer_depth: 1
37
+ context_dim: 768
38
+ use_checkpoint: True
39
+ legacy: False
40
+
41
+ unet_config:
42
+ target: cldm.cldm.ControlledUnetModel
43
+ params:
44
+ image_size: 32 # unused
45
+ in_channels: 9
46
+ out_channels: 4
47
+ model_channels: 320
48
+ attention_resolutions: [ 4, 2, 1 ]
49
+ num_res_blocks: 2
50
+ channel_mult: [ 1, 2, 4, 4 ]
51
+ num_heads: 8
52
+ use_spatial_transformer: True
53
+ transformer_depth: 1
54
+ context_dim: 768
55
+ use_checkpoint: True
56
+ legacy: False
57
+
58
+ first_stage_config:
59
+ target: ldm.models.autoencoder.AutoencoderKL
60
+ params:
61
+ embed_dim: 4
62
+ monitor: val/rec_loss
63
+ ddconfig:
64
+ double_z: true
65
+ z_channels: 4
66
+ resolution: 256
67
+ in_channels: 3
68
+ out_ch: 3
69
+ ch: 128
70
+ ch_mult:
71
+ - 1
72
+ - 2
73
+ - 4
74
+ - 4
75
+ num_res_blocks: 2
76
+ attn_resolutions: []
77
+ dropout: 0.0
78
+ lossconfig:
79
+ target: torch.nn.Identity
80
+
81
+ contextual_stage_config:
82
+ target: models_deep_exp.NonlocalNet.VGG19_pytorch
83
+
84
+ cond_stage_config:
85
+ target: ldm.modules.encoders.modules.FrozenCLIPEmbedder
86
+ # target: ldm.modules.encoders.modules.FrozenCLIPDualEmbedder
87
+ #ldm.modules.encoders.modules.FrozenCLIPDualEmbedder