You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Branching off from the issue which @aflah02 opened a few weeks ago, #39:
Is the KerasNLP team interested in implementing adversarial attacks? We could start off with simple attacks on classification models.
I understand if this is a bit broad, and the team may want to integrate it later to the repository, especially because we may need some augmentation APIs. For example, some adversarial attacks may want to perturb only those words which are assigned a higher importance score by the model. For perturbation, we can leverage the augmentation APIs.
@abheesht17 Thanks for opening this feature request!
Yes, having an adversarial attack system would be nice for model evaluation. Our current problem is that we do not have pretrained model available. When you start working on this, would you mind sharing a colab so that we can do some early reviews on the interface? Thanks!
Branching off from the issue which @aflah02 opened a few weeks ago, #39:
Is the KerasNLP team interested in implementing adversarial attacks? We could start off with simple attacks on classification models.
I understand if this is a bit broad, and the team may want to integrate it later to the repository, especially because we may need some augmentation APIs. For example, some adversarial attacks may want to perturb only those words which are assigned a higher importance score by the model. For perturbation, we can leverage the augmentation APIs.
A good resource is https://github.com/QData/TextAttack.
The text was updated successfully, but these errors were encountered: