Issues
- 0
Error while using pip
#70 opened by avi7611 - 0
Contributions guidelines
#69 opened by miloszwatroba - 5
Divide by zero: request_metrics[common_metrics.REQ_OUTPUT_THROUGHPUT] = num_output_tokens / request_metrics[common_metrics.E2E_LAT]
#55 opened by yaronr - 0
Azure OpenAi Endpoint Support
#67 opened by gujju62 - 1
Since it does not have a 'setup.py' nor a 'setup.cfg', it cannot be installed in editable mode PEP 660
#47 opened by focusunsink - 0
When max-num-completed-requests is not divisible by num-concurrent-requests, an error will occur.
#63 opened by huangdi614 - 1
- 5
Subsequent requests cannot be sent until 'num_concurrent_requests' requests have all finished
#56 opened by llsj14 - 1
- 0
Sagemaker client issue
#53 opened by SuchethaChintha - 0
- 0
Vertex AI API needs to be updated.
#48 opened by Durga2Dash - 1
unable to get the benchmark result
#44 opened by dipshirajput - 0
HUGGINGFACE set
#38 opened by capyun - 0
bug of counting output tokens
#35 opened by irasin - 0
llmperf not working for concurrent users
#34 opened by nkanike07 - 0
Bug: Hugging Face TGI not working
#33 opened by ptrmayer - 0
Concurrency level is not handled properly
#32 opened by alexeykudinkin - 0
Add memory bandwidth utilization metric
#31 opened by mmcclean-aws - 0
- 0
usage for local models .
#29 opened by Akash08naik - 0
Basic usage issue
#24 opened by wangxingjun778 - 5
Are benchmark results released somewhere
#9 opened by ogencoglu - 0
litellm serializable issue ?
#21 opened by ishaan-jaff - 0
- 3
Does llmperf support measuring local disk models? What's the meaning of framework in llmperf.py line 355?
#11 opened by zhangjiawei5911