Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

image gen experiments #23

Open
zzsi opened this issue Apr 20, 2023 · 1 comment
Open

image gen experiments #23

zzsi opened this issue Apr 20, 2023 · 1 comment

Comments

@zzsi
Copy link
Collaborator

zzsi commented Apr 20, 2023

@zzsi
Copy link
Collaborator Author

zzsi commented Apr 29, 2023

Notes about the RRR paper above:

  • The jax implementation on toy models is informative: https://github.com/yilundu/reduce_reuse_recycle/blob/main/notebooks/simple_distributions.ipynb
  • There is the diffusion model, whose forward(x, t) function computes the gradient.
  • There is the ebm wrapped around the diffusion model. The ebm has an additional neg_logp_unnorm function that computes the negative log prob. ebm also has a forward function that computes the gradient, but this gradient requires a backprop through the diffusion model's network.
  • ebm models can be composed using their neg_log_unnorm functions. ProductEBMDiffusionModel is the product of 2 distributions, which means the addition of neg_log_unnorm outputs.
  • Either the diffusion model or the ebm can be used to construct a PortableDiffusionModel (ddpm). ddpm is required for training. If ddpm holds an ebm, it can compute p_energy. The behavior of the loss function is the same, and does not use ebm's neg_log_unnorm.
  • There are samplers that don't need energy: AnnealedULASampler, AnnealedUHASampler
  • There are samplers that require energy: AnnealedMALASampler, AnnealedMUHASampler

Questions:

  • What's the sample quality if, we use a pretrained diffusion model, without finetune, be wrapped as an EBM and with an energy-based sampler like AnnealedMUHASampler?
  • Does further EBM training/fine-tune help with the sample quality?
  • If the diffusion is on latents (e.g. stable diffusion), do the conclusions to the above questions change? What tweaks are needed for using latents?
  • Besides composition of distributions, what other benefits do EBM models have against score-based models?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant