You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The original textual inversion support using more than one vector to represent the learnt concept.
For the current implementation, if we just extend the learned vocab and CLIP token embedding then it would use only one vector for it.
What could be the best way to support this? cc @patil-suraj