Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Compression techniques #37

Open
erickdp opened this issue Mar 21, 2024 · 2 comments
Open

Compression techniques #37

erickdp opened this issue Mar 21, 2024 · 2 comments
Labels
enhancement New feature or request

Comments

@erickdp
Copy link

erickdp commented Mar 21, 2024

I find it truly fascinating! Have you come across any methods similar to pruning, distillation, or quantization that could be applied to this model? While I'm aware of some size options, it would be truly remarkable if we could utilize compression techniques for more efficient processing and deployment on edge devices.

@urchade
Copy link
Owner

urchade commented Mar 22, 2024

Hi @erickdp thanks

Indeed it would be interesting, however I am not really familiar with this field. Do you have any idea how we could do that?

@erickdp
Copy link
Author

erickdp commented Mar 22, 2024

I could recommend you Knowledge Distillation method, that consists in fitting "students" models from "teachers" since it is mentioned that GLiNER handles a BERT-like architecture, I have used it to distill sentiment classification models and the result is really efficient, from the computational process needed, model size and accuracy.

However, excellent contribution.

References:
https://neptune.ai/blog/knowledge-distillation
https://github.com/huggingface/transformers/tree/main/examples/research_projects/distillation
https://huggingface.co/lxyuan/distilbert-base-multilingual-cased-sentiments-student

@urchade urchade added the enhancement New feature or request label Apr 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants