Skip to content

Transformer related optimization, including BERT, GPT

License

Notifications You must be signed in to change notification settings

THUDM/FasterTransformer

 
 

Repository files navigation

FasterTransformer

This repository is based on FasterTransformer adapted to GLM-130B, for FasterTransformer, please read the original project.

Quick Start

Read inference-with-fastertransformer.

About

Transformer related optimization, including BERT, GPT

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • C++ 64.3%
  • Cuda 30.9%
  • CMake 2.5%
  • Python 1.4%
  • Shell 0.8%
  • C 0.1%