Skip to content
#

low-rank-approximation

Here are 22 public repositories matching this topic...

Deep learning models have become state of the art for natural language processing (NLP) tasks, however deploying these models in production system poses significant memory constraints. Existing compression methods are either lossy or introduce significant latency. We propose a compression method that leverages low rank matrix factorization durin…

  • Updated Sep 14, 2021
  • Python

Improve this page

Add a description, image, and links to the low-rank-approximation topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the low-rank-approximation topic, visit your repo's landing page and select "manage topics."

Learn more