Skip to content

question about distill_dcm_wan_detail_expert #24

@zhuamaaa

Description

@zhuamaaa

I noticed that this line was commented out before your LoraConfig to transformer:

transformer.requires_grad_(False)

Image

Doesn't this go against the original intention of your detail expert for fine-tuning the semantic expert?
Total num of the training parameters is about 1.2 billion

btw im trying to train the detail expert but OOM problem happended. The problem was found in computing gan_g_loss. wondering if you could give me some suggestions. looking forward to your reply

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions