diff options
author | Zac Liu <liuguang@baai.ac.cn> | 2022-11-30 03:14:04 +0000 |
---|---|---|
committer | GitHub <noreply@github.com> | 2022-11-30 03:14:04 +0000 |
commit | a39a57cb1f5964d9af2b541f7b352576adeeac0f (patch) | |
tree | ebae98ea40ecc5b34497424bee19310e9fac4068 /ldm/modules/ema.py | |
parent | 4b3c5bc24bffdf429c463a465763b3077fe55eb8 (diff) | |
parent | 0831ab476c626eb796b609acf8771177692bfab7 (diff) | |
download | stable-diffusion-webui-gfx803-a39a57cb1f5964d9af2b541f7b352576adeeac0f.tar.gz stable-diffusion-webui-gfx803-a39a57cb1f5964d9af2b541f7b352576adeeac0f.tar.bz2 stable-diffusion-webui-gfx803-a39a57cb1f5964d9af2b541f7b352576adeeac0f.zip |
Merge pull request #1 from 920232796/master
Add AltDiffusion
Diffstat (limited to 'ldm/modules/ema.py')
-rw-r--r-- | ldm/modules/ema.py | 76 |
1 files changed, 76 insertions, 0 deletions
diff --git a/ldm/modules/ema.py b/ldm/modules/ema.py new file mode 100644 index 00000000..c8c75af4 --- /dev/null +++ b/ldm/modules/ema.py @@ -0,0 +1,76 @@ +import torch +from torch import nn + + +class LitEma(nn.Module): + def __init__(self, model, decay=0.9999, use_num_upates=True): + super().__init__() + if decay < 0.0 or decay > 1.0: + raise ValueError('Decay must be between 0 and 1') + + self.m_name2s_name = {} + self.register_buffer('decay', torch.tensor(decay, dtype=torch.float32)) + self.register_buffer('num_updates', torch.tensor(0,dtype=torch.int) if use_num_upates + else torch.tensor(-1,dtype=torch.int)) + + for name, p in model.named_parameters(): + if p.requires_grad: + #remove as '.'-character is not allowed in buffers + s_name = name.replace('.','') + self.m_name2s_name.update({name:s_name}) + self.register_buffer(s_name,p.clone().detach().data) + + self.collected_params = [] + + def forward(self,model): + decay = self.decay + + if self.num_updates >= 0: + self.num_updates += 1 + decay = min(self.decay,(1 + self.num_updates) / (10 + self.num_updates)) + + one_minus_decay = 1.0 - decay + + with torch.no_grad(): + m_param = dict(model.named_parameters()) + shadow_params = dict(self.named_buffers()) + + for key in m_param: + if m_param[key].requires_grad: + sname = self.m_name2s_name[key] + shadow_params[sname] = shadow_params[sname].type_as(m_param[key]) + shadow_params[sname].sub_(one_minus_decay * (shadow_params[sname] - m_param[key])) + else: + assert not key in self.m_name2s_name + + def copy_to(self, model): + m_param = dict(model.named_parameters()) + shadow_params = dict(self.named_buffers()) + for key in m_param: + if m_param[key].requires_grad: + m_param[key].data.copy_(shadow_params[self.m_name2s_name[key]].data) + else: + assert not key in self.m_name2s_name + + def store(self, parameters): + """ + Save the current parameters for restoring later. + Args: + parameters: Iterable of `torch.nn.Parameter`; the parameters to be + temporarily stored. + """ + self.collected_params = [param.clone() for param in parameters] + + def restore(self, parameters): + """ + Restore the parameters stored with the `store` method. + Useful to validate the model with EMA parameters without affecting the + original optimization process. Store the parameters before the + `copy_to` method. After validation (or model saving), use this to + restore the former parameters. + Args: + parameters: Iterable of `torch.nn.Parameter`; the parameters to be + updated with the stored parameters. + """ + for c_param, param in zip(self.collected_params, parameters): + param.data.copy_(c_param.data) |