You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for creating this repository, it is really helpful. So I trained a decoder for my custom task and was generating images using it and I found out that it was generating random noise images. A sample image generated by the model is added below-
The inference code of the decoder model is as follows-
image_decoder = torch.load(image_decoder_path).cuda()
image_generated = image_decoder.sample(image_embed = image_proj, cond_scale = 2.)
The image_proj is a processed embedding from a trained CLIP.
This is the Unet decoder setting I am using while training
Hi,
Thank you for creating this repository, it is really helpful. So I trained a decoder for my custom task and was generating images using it and I found out that it was generating random noise images. A sample image generated by the model is added below-
The inference code of the decoder model is as follows-
image_decoder = torch.load(image_decoder_path).cuda()
image_generated = image_decoder.sample(image_embed = image_proj, cond_scale = 2.)
The image_proj is a processed embedding from a trained CLIP.
This is the Unet decoder setting I am using while training
unet = Unet(
dim = 8,
image_embed_dim = 256,
cond_dim = 128,
channels = 3,
dim_mults=(1, 2, 4, 8)
).cuda()
decoder = Decoder(
unet = unet,
image_size = 224,
timesteps = 100,
image_cond_drop_prob = 0.1,
text_cond_drop_prob = 0.5,
learned_variance=False
).cuda()
The training loss is coming out to be 0.051178544054353195
Any help would be really useful...Thank You!!!
The text was updated successfully, but these errors were encountered: