Skip to content

Commit d0bee35

Browse files
authored
Merge pull request #1 from vicgalle/dev
Initial functionality
2 parents 768517b + b248e2f commit d0bee35

28 files changed

+481
-250
lines changed

README.md

+115-155
Large diffs are not rendered by default.

aesthetic_embeddings/aivazovsky.pt

2.23 KB
Binary file not shown.

aesthetic_embeddings/cloudcore.pt

2.23 KB
Binary file not shown.

aesthetic_embeddings/gloomcore.pt

2.23 KB
Binary file not shown.

aesthetic_embeddings/glowwave.pt

2.23 KB
Binary file not shown.

aesthetic_embeddings/laion_7plus.pt

3.73 KB
Binary file not shown.

aesthetic_embeddings/sac_8plus.pt

3.73 KB
Binary file not shown.

assets/grid-0033.png

1.64 MB
Loading

assets/grid-0035.png

1.59 MB
Loading

assets/grid-0043.png

1.9 MB
Loading

assets/grid-0047.png

1.78 MB
Loading

assets/grid-0072.png

3.14 MB
Loading

assets/grid-0073.png

3.29 MB
Loading

assets/grid-0075.png

3.1 MB
Loading

assets/grid-0089.png

2.91 MB
Loading

assets/grid-0091.png

2.93 MB
Loading

assets/grid-0096.png

2.93 MB
Loading

assets/grid-0097.png

3.07 MB
Loading

assets/grid-0131.png

4.43 MB
Loading

assets/grid-0133.png

4.75 MB
Loading

assets/grid-0135.png

4.91 MB
Loading
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,74 @@
1+
model:
2+
base_learning_rate: 1.0e-04
3+
target: ldm.models.diffusion.ddpm.LatentDiffusion
4+
params:
5+
linear_start: 0.00085
6+
linear_end: 0.0120
7+
num_timesteps_cond: 1
8+
log_every_t: 200
9+
timesteps: 1000
10+
first_stage_key: "jpg"
11+
cond_stage_key: "txt"
12+
image_size: 64
13+
channels: 4
14+
cond_stage_trainable: false # Note: different from the one we trained before
15+
conditioning_key: crossattn
16+
monitor: val/loss_simple_ema
17+
scale_factor: 0.18215
18+
use_ema: False
19+
20+
scheduler_config: # 10000 warmup steps
21+
target: ldm.lr_scheduler.LambdaLinearScheduler
22+
params:
23+
warm_up_steps: [10000]
24+
cycle_lengths: [10000000000000] # incredibly large number to prevent corner cases
25+
f_start: [1.e-6]
26+
f_max: [1.]
27+
f_min: [1.]
28+
29+
unet_config:
30+
target: ldm.modules.diffusionmodules.openaimodel.UNetModel
31+
params:
32+
image_size: 32 # unused
33+
in_channels: 4
34+
out_channels: 4
35+
model_channels: 320
36+
attention_resolutions: [4, 2, 1]
37+
num_res_blocks: 2
38+
channel_mult: [1, 2, 4, 4]
39+
num_heads: 8
40+
use_spatial_transformer: True
41+
transformer_depth: 1
42+
context_dim: 768
43+
use_checkpoint: True
44+
legacy: False
45+
46+
first_stage_config:
47+
target: ldm.models.autoencoder.AutoencoderKL
48+
params:
49+
embed_dim: 4
50+
monitor: val/rec_loss
51+
ddconfig:
52+
double_z: true
53+
z_channels: 4
54+
resolution: 256
55+
in_channels: 3
56+
out_ch: 3
57+
ch: 128
58+
ch_mult:
59+
- 1
60+
- 2
61+
- 4
62+
- 4
63+
num_res_blocks: 2
64+
attn_resolutions: []
65+
dropout: 0.0
66+
lossconfig:
67+
target: torch.nn.Identity
68+
69+
cond_stage_config:
70+
target: ldm.modules.encoders.modules.PersonalizedCLIPEmbedder
71+
params:
72+
aesthetic_embedding_path: "aesthetic_embeddings/sac_8plus.pt"
73+
T: 3
74+
lr: 0.0001

0 commit comments

Comments
 (0)