Skip to content

sujitvasanth/streaming-LLM-chat

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 
 
 

Repository files navigation

Streaming-LLM-chat

samplechat

This is a transformers library application that allows you to choose a local LLM and run streaming inference on GPU.

it uses:

  • Python: 3.8.10
  • transformers library: 4.36.2
  • transformers_stream_generator library

the models are assumed to be in oogabooga textgeneration ui folder

the openchat model is available at https://huggingface.co/

TheBloke/openchat-3.5-0106-GPTQ

sujitvasanth/TheBloke-openchat-3.5-0106-GPTQ

About

transformers based streaming chat for GPTQ models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages