fix to use zero for initial latent

This commit is contained in:
Kohya S
2024-06-24 23:12:48 +09:00
parent d53ea22b2a
commit 0fe4eafac9

View File

@@ -64,7 +64,8 @@ def do_sample(
device: str,
):
if initial_latent is None:
latent = torch.ones(1, 16, height // 8, width // 8, device=device) * 0.0609
# latent = torch.ones(1, 16, height // 8, width // 8, device=device) * 0.0609
latent = torch.zeros(1, 16, height // 8, width // 8, device=device)
else:
latent = initial_latent