Code repository for the research paper "Space Efficient Transformer Neural Network"
-
Updated
Aug 6, 2023 - C++
Code repository for the research paper "Space Efficient Transformer Neural Network"
Method for searching relevant podcast segments from transcripts using transformer models
Dutch/Indonesian BERT-NER setup.
A tutorial of building tensorflow serving service from scratch
optimized BERT transformer inference on NVIDIA GPU. https://arxiv.org/abs/2210.03052
Running BERT without Padding
Fast implementation of BERT inference directly on NVIDIA (CUDA, CUBLAS) and Intel MKL
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
LightSeq: A High Performance Library for Sequence Processing and Generation
该仓库主要记录 NLP 算法工程师相关的顶会论文研读笔记
Transformer related optimization, including BERT, GPT
Add a description, image, and links to the bert topic page so that developers can more easily learn about it.
To associate your repository with the bert topic, visit your repo's landing page and select "manage topics."