Skip to content

Conversation

mattdangerw
Copy link
Member

@mattdangerw mattdangerw commented Sep 12, 2024

For some reason, all our causal lm models had a preprocessing layer base class that packed inputs without setting things up for causal language modeling. With our new base classes, CausalLMPreprocessor and Seq2SeqLMPreprocessor, these are no longer used at all.

These are public, but I did a search for them on GitHub and the only one I could find with any usage is GPT2Preprocessor. The others are complete unused, and not generally very useful.

Let's remove these before anyone does depend on them so that we can keep preprocessing one-to-one with a task. E.g. it should always look like this:

BertTextClassifier
BertTextClassifierPreprocesssor

For some reason, all our causal lm models had a preprocessing
layer base class that packed inputs without setting things up for
the causal language model loss.

With our new base classes, `CausalLMPreprocessor` and
`Seq2SeqLMPreprocessor`, these are no longer used at all.

These are public, but I did a search for them on GitHub and the
only one I could find with any usage is `GPT2Preprocessor`. The
others are complete unused, and not generally very useful.

Let's remove them before anyone does depend on them so that we can
keep preprocessing one-to-one paired with a task. E.g.

```
BertTextClassifier
BertTextClassifierPreprocesssor
```
@mattdangerw mattdangerw changed the title Remove preprocessing base classes we no longer use Remove preprocessing layers we no longer use Sep 12, 2024
Copy link
Member

@SamanehSaadat SamanehSaadat left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks, Matt!

@mattdangerw mattdangerw merged commit dd0651e into keras-team:master Sep 12, 2024
11 of 13 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants