Conversation
|
Thank you for this! I am working on Anima based on ComfyUI code, but the GPL 3.0 license is a problem. This is based on ASL 2.0, which is great. Please note that there is a possibility that some changes will be made after merging to align with other models. |
|
If you need any help, feel free to reach out. I’d be happy to help. |
|
I have a few questions.
|
|
Yes, I noticed these issues as well when I reviewed them. However, these configurations come from the diffusion-pipe repository, and the owner of this repository is also the creator of this model. |
Thank you for the explanation. I understood. So keeping the config in the repository sounds good. |
|
I also noticed one other thing: ComfyUI's inference code doesn't seem to use an attention mask in its Text Encoder calls, but diffusion-pipe does seem to use one. |
|
I’m also very confused about this, but when I tested the model with and without an attention mask, I observed that using the attention mask makes the model more stable and produces better results. |
|
Hmm, the attention mask is used for Text Encoder, but then the embedding where mask=0 is set to zero, and a fixed-length embedding is passed to DiT. That's interesting... |
|
I'm not entirely sure about the author's intention, but here is my understanding (I could be wrong):
|
|
Thank you for the detailed explanation. I think I have a general understanding of Anima's training and inference. Please understand that I will make some changes after merging. I would also appreciate any feedback or additional pull requests after the merge. Thank you again for this great PR! |
|
working well. Thank you @kohya-ss, @duongve13112002 |
Hi @kohya-ss
Here is the description of my change for support Anima model in sd-scripts
Summary
Add Anima model support for LoRA and full-finetune training
Changes
Anima Model Support
Note: No existing files modified, all changes are new files only. Follows the same patterns as SD3/Lumina/Flux. Tested with both LoRA and full finetune training. If you have any question feel free to ask.