Lightning-AI/LitServe
Lightning-fast serving engine for any AI model of any size. Flexible. Easy. Enterprise-scale.
PythonApache-2.0
Issues
- 1
Improve the debugging experience of LitServe
#344 opened by tchaton - 2
how to use Litserve to serve vllm engine
#382 opened by GhostXu11 - 0
[Windows] OSError when multiple workers
#384 opened by FrsECM - 0
`max_tokens` in OpenAI spec is deprecated
#373 opened by zcasanova - 7
- 2
[Windows] Server Hangs while closing
#372 opened by FrsECM - 2
Info route
#366 opened by lorenzomassimiani - 15
Embedding model support with openai spec
#305 opened by riyajatar37003 - 1
- 3
Raising HTTPexception on single request from encode_request or decode_request raise same for all request in batch
#340 opened by bimarshak7 - 1
terminate early if accelerator is missing
#292 opened by Borda - 0
Default Input-Output schema
#354 opened by aniketmaurya - 17
- 2
Add custom path option for healthcheck
#341 opened by akuma12 - 3
Support for async requests or webhooks
#348 opened by brian316 - 6
bug: OpenAI-Spec should send text/event-stream
#359 opened by fabigr8 - 6
How to route /docs path in litserve behind a proxy?
#352 opened by Mayurji - 3
- 5
GET Request
#355 opened by gururaser - 3
Can't start the demo example without bug
#350 opened by anonym-ai - 5
Prometheus logger is not pickable + monitoring metrics set via self.log are not tracked
#339 opened by miguelalba96 - 2
Add dockerize example in the README
#345 opened by tchaton - 1
- 6
How to load models from S3
#329 opened by AbhishekBose - 4
Example in documentation on how to setup an OpenAI-spec API with LlamaIndex-RAG
#286 opened by PierreMesure - 4
Websocket Support for Streaming Input and Output
#320 opened by ChenghaoMou - 2
- 12
- 1
- 5
May I ask if it is possible to deploy complex pipelines such as ControlNet and IP Adapter in Stable Diffusion? Do you have any examples
#236 opened by libai-lab - 2
CUDA Multiprocessing Issue
#321 opened by floschne - 2
Pyright issues [Name mismatches]
#310 opened by grumpyp - 1
dry run after server started
#293 opened by Borda - 3
Custom HTTPException Not Working as Expected
#301 opened by randydl - 5
unexpected output for HF model with matching
#294 opened by Borda - 0
Handle case when Logger.process is stuck
#289 opened by aniketmaurya - 5
Feature Request: Customize FastAPI Metadata
#270 opened by bhimrazy - 2
Can't appear to load the server
#273 opened by B-M-S-West - 4
Supporting query params alongside file upload
#272 opened by energydrink9 - 2
Add Support to Huggingface Diffusers!
#275 opened by KaifAhmad1 - 6
- 2
Setup step is not awaited
#263 opened by andreieuganox - 8
If `workers_per_device` is set to 1, how can I set a model instance to be able to concurrency?
#267 opened by ALEXuH - 2
The current openai spec is only compatible with version openai==0.28.0 and cannot recognize newer versions
#245 opened by GhostXu11 - 5
Feature: Pass custom middlewares to the app
#228 opened by lorenzomassimiani - 1
Replace uvicorn server with gunicorn
#243 opened by vrdn-23 - 2
- 2
client.py and fastapi /docs "try it out" both are not working with BERT example
#230 opened by sebastianschramm - 1
Decorator design for `LitServer`
#218 opened by baggiponte - 2
Add support for `response_format` in OpenAI Spec
#196 opened by bhimrazy