-
-
Notifications
You must be signed in to change notification settings - Fork 167
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Weight data type float16 produce error #402
Comments
This is possibly a problem with using a cache that was created in fp32. I noticed your config has clear_cache_before_training: false. Try clearing the cache and see if that fixes your problem. |
Clearing cache didn't helped, but I notice that error is gone when I turn on masked training. It seems that without masking batch['latent_mask'] has dtype float32 as it comes from GenerateImageLike mgds node. So produces float32 tensor and unet with float16 gives error |
I also can confirm this issue (with any model). but I ignored it as I usually go with bf16. Also clear cache would not help. |
Also similar errors arise in debug process and sampling process. It can be seen after fixing GenerateImageLike node |
During sampling: Traceback (most recent call last): | 0/1 [00:00<?, ?it/s] during debug process: |
What happened?
Weight data type float16 produce error
config attached
default.json
model used - majicmixRealistic_v7-inpainting.safetensors fp16
https://civitai.com/models/43331?modelVersionId=221343
Think it's because of some places where dtype is explicitly set to .float()
What did you expect would happen?
That float16 training would work)
Relevant log output
Output of
pip freeze
No response
The text was updated successfully, but these errors were encountered: