Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The possible contradiction for disentangled cfg between the paper and train code #44

Open
takesukeDS opened this issue Aug 15, 2023 · 1 comment

Comments

@takesukeDS
Copy link

Hello authors, your work is impressive. Thanks for sharing the code base.

I want to clarify about your disentangled cfg.
The paper mentions that you omitted the pose condition and the style condition with 0.1 probability.
However, this code(train.py) seems to omit only the style condition. Since, invocation of unet in GaussianDiffusion.training_losses()

model_output = model(x = torch.cat([x_t, target_pose],1), t = self._scale_timesteps(t), x_cond = img, prob = prob)

passes both target_pose as concatenated input and img(style) as condition along with prob.
Although the x_cond is masked with the probability given in the forward function of the unet BeatGANsAutoencModel.forward(), the argumentx is used without any modification.

Could you clarify how you train your model for disentangled cfg?

Excuse me if I overlooked something.
Best regards.

@YanzuoLu
Copy link

YanzuoLu commented Sep 5, 2023

Same question about this.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants