Issues
- 0
ROCm support missing from Dockerfile
#3380 opened by jakki-amd - 0
Clarify interactions between TorchServe and KServe
#3378 opened by kimminw00 - 1
Getting started guide client samples broken ?
#3348 opened by nikste - 0
Fix missing system metrics for Apple M1
#3376 opened by jakki-amd - 3
- 1
503 InternalServerException, prediction failed
#3375 opened by Jax29 - 1
- 0
vLLM ZeroDivisionError
#3373 opened by mfahadyousaf - 2
Incorrect Metric Type for HPA Scaling
#3286 opened by liaddrori1 - 0
- 0
- 0
Trying to find a doc explaining how the scaling works (min_worker to max_worker)
#3362 opened by lschaupp - 0
Issue when sending parallel requests
#3361 opened by lschaupp - 0
`weights_only` default flip for `torch.load`
#3360 opened by mikaylagawarecki - 0
- 0
Model & Instance scaling
#3358 opened by markcNewell - 0
413 Request Entity Too Large
#3357 opened by pengxin233 - 1
GPU not detected inside torchserve docker container
#3352 opened by dummyuser-123 - 0
Allow to provide custom model_service_worker.py
#3353 opened by racinmat - 0
Make TorchServe usable without torch installed
#3350 opened by racinmat - 10
Torchserve not starting for diffusers example
#3345 opened by dummyuser-123 - 0
- 1
- 7
io.netty.channel.unix.Errors$NativeIoException: readAddress(..) failed: Connection reset by peer
#3279 opened by KD1994 - 2
- 2
- 0
UnboundLocalError: local variable 'model_snapshot_path' referenced before assignment
#3335 opened by johnathanchiu - 0
Clarification on minWorkers and maxWorkers parameters
#3339 opened by krzwaraksa - 1
Kserve MNIST CI failure
#3308 opened by maaquib - 2
Allow .tar.gz models to be loaded using load_models=all
#3320 opened by m10an - 0
Kserve management api for registering new models
#3325 opened by matej14086 - 10
Examples for authorization model
#3260 opened by RyanKadri - 1
- 0
Kserve TorchserveModel can't handle torch's auth
#3301 opened by AntPeixe - 3
when to use ArgumentParser, raise "unrecognized arguments: --sock-type unix --sock-name /tmp/.ts.sock.9000"
#3299 opened by james-joobs - 1
model_yaml_config usage is not explained well enough
#3290 opened by Foundsheep - 1
- 1
Ability to return Pydantic Models
#3255 opened by mhashas - 5
Startup timeout should be configurable separately
#3261 opened by Isalia20 - 1
Model handler response structure enforced by ts.service.Service class but should be defined by model
#3284 opened by stf976 - 3
Websocket Support
#3252 opened by tiefucai - 1
Support for a CUDA 12.5.x
#3278 opened by jasonsmithio - 1
StreamPredictions2 gPRC method execution produces a server error while changing from version *0.10.0* to *0.11.1*
#3264 opened by ferugit - 1
Problem when run custom Docker image with GPU
#3258 opened by hungtrieu07 - 1
GPU memory not released after inference
#3253 opened by Di-Gu - 1
TorchServe docker/kserve nightly image size increase
#3244 opened by agunapal - 0
- 0
TorchServe docker image with vllm, trt-llm dependencies
#3247 opened by agunapal - 0
Alarming log(false alarm) about xpu-smi
#3236 opened by agunapal - 2
misaligned argument between `args.disable_token` and `--disable_token-auth` in llm_launcher.py
#3229 opened by jinyichao