Skip to content

codepongo/benchmark

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

benchmark

Test suites for serving & offline benchmark

Benchmark structure for llm

Inference

  • serving

    work with serving task

    cd inference/serving/llm

    suppose you are trying to benchmark glm-4-9b with configs:

     1. 'inp: 64; out: 64; concurrent: 4'
     2. 'inp: 64; out: 64; concurrent: 8'
     3. 'inp: 128; out: 64; concurrent: 4'
     4. 'inp: 128; out: 64; concurrent: 8'
    

    code

    python benchmark_llm.py --config config.yaml --model glm-4-9b --device musa --input "64,128" --output "64" --concurrent "4,8"
    
  • offline

About

Test suites for serving & offline benchmark

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages

  • Python 79.3%
  • Shell 14.7%
  • Batchfile 6.0%