diff options
| author | AUTOMATIC <16777216c@gmail.com> | 2022-12-31 18:06:35 +0300 | 
|---|---|---|
| committer | AUTOMATIC <16777216c@gmail.com> | 2022-12-31 18:06:35 +0300 | 
| commit | f34c7341720fb2059992926c9f9ae6ff25f7385b (patch) | |
| tree | be719a629f8754c206d891b1850f0b5eaf186e2e /configs/altdiffusion/ad-inference.yaml | |
| parent | 3f401cdb644066fd43abf6642d2e53be53c73668 (diff) | |
| download | stable-diffusion-webui-gfx803-f34c7341720fb2059992926c9f9ae6ff25f7385b.tar.gz | |
alt-diffusion integration
Diffstat (limited to 'configs/altdiffusion/ad-inference.yaml')
| -rw-r--r-- | configs/altdiffusion/ad-inference.yaml | 72 | 
1 files changed, 0 insertions, 72 deletions
| diff --git a/configs/altdiffusion/ad-inference.yaml b/configs/altdiffusion/ad-inference.yaml deleted file mode 100644 index cfbee72d..00000000 --- a/configs/altdiffusion/ad-inference.yaml +++ /dev/null @@ -1,72 +0,0 @@ -model: -  base_learning_rate: 1.0e-04 -  target: ldm.models.diffusion.ddpm.LatentDiffusion -  params: -    linear_start: 0.00085 -    linear_end: 0.0120 -    num_timesteps_cond: 1 -    log_every_t: 200 -    timesteps: 1000 -    first_stage_key: "jpg" -    cond_stage_key: "txt" -    image_size: 64 -    channels: 4 -    cond_stage_trainable: false   # Note: different from the one we trained before -    conditioning_key: crossattn -    monitor: val/loss_simple_ema -    scale_factor: 0.18215 -    use_ema: False - -    scheduler_config: # 10000 warmup steps -      target: ldm.lr_scheduler.LambdaLinearScheduler -      params: -        warm_up_steps: [ 10000 ] -        cycle_lengths: [ 10000000000000 ] # incredibly large number to prevent corner cases -        f_start: [ 1.e-6 ] -        f_max: [ 1. ] -        f_min: [ 1. ] - -    unet_config: -      target: ldm.modules.diffusionmodules.openaimodel.UNetModel -      params: -        image_size: 32 # unused -        in_channels: 4 -        out_channels: 4 -        model_channels: 320 -        attention_resolutions: [ 4, 2, 1 ] -        num_res_blocks: 2 -        channel_mult: [ 1, 2, 4, 4 ] -        num_heads: 8 -        use_spatial_transformer: True -        transformer_depth: 1 -        context_dim: 768 -        use_checkpoint: True -        legacy: False - -    first_stage_config: -      target: ldm.models.autoencoder.AutoencoderKL -      params: -        embed_dim: 4 -        monitor: val/rec_loss -        ddconfig: -          double_z: true -          z_channels: 4 -          resolution: 256 -          in_channels: 3 -          out_ch: 3 -          ch: 128 -          ch_mult: -          - 1 -          - 2 -          - 4 -          - 4 -          num_res_blocks: 2 -          attn_resolutions: [] -          dropout: 0.0 -        lossconfig: -          target: torch.nn.Identity - -    cond_stage_config: -      target: modules.xlmr.BertSeriesModelWithTransformation -      params: -        name: "XLMR-Large"
\ No newline at end of file | 
