You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Yes, we have attempted training with lora but encountered performance degradation with default ranks. Therefore, we did not end up having a lora finetuned version.
That said, it still might be possible to finetune with lora, although many hyperparameters (e.g., rank, types of attention weights) likely need to be carefully experimented specifically with our setup, which makes attention map itself an objective to optimize.
Hi, I wonder if you have tried using LoRA to finetune the model? It should need less gpu memory than the current full finetuning strategy.
The text was updated successfully, but these errors were encountered: