Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Text Preprocessing #56

Open
someshsingh22 opened this issue May 13, 2020 · 2 comments
Open

Text Preprocessing #56

someshsingh22 opened this issue May 13, 2020 · 2 comments
Labels
Priority: Medium question Further information is requested

Comments

@someshsingh22
Copy link
Member

To implement a common black box we need text loading, extraction of words to be attacked, perturbations, distance metrics, models.

Text Loading needs to be very uniform and universal, it should encapsulate all common practices including embedding, tokenizers, batch_loaders, and should support commonly used libraries like nltk, spacy, BERT etc.

We need to think about how we should design this before our first attack.

@someshsingh22 someshsingh22 added the question Further information is requested label May 13, 2020
@someshsingh22 someshsingh22 added this to In progress in Character level attacks May 13, 2020
@someshsingh22 someshsingh22 moved this from In progress to To do in Character level attacks May 13, 2020
@parantak
Copy link
Contributor

@someshsingh22 , I believe we could start by creating a common class for this inside decepticonlp/preprocess, and then start by implementing a class for each of the separate practices. Does that sound good, or do you want to take a different approach to this?

@parantak
Copy link
Contributor

Refer to #75

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Priority: Medium question Further information is requested
Projects
Development

No branches or pull requests

2 participants