awslabs/multi-model-server
Multi Model Server is a tool for serving neural net models for inference
JavaApache-2.0
Issues
- 1
- 1
Streaming support on MMS
#1012 opened by rauldiaz - 4
Issue: Memory Leak when serving multiple models
#999 opened by pratikluitel - 0
MMS Server getting stuck while registering the model and says "worker pid is not available yet"
#1025 opened by suchith-sixsense - 1
How to handle invalid input!
#1022 opened by xuweidongkobe - 0
Improve MMS model loading exception handling
#1010 opened by namannandan - 0
Overriding the model routing logic
#1017 opened by James-UnlikelyAI - 1
readAddress(..) failed: Connection reset by peer
#1016 opened by xuweidongkobe - 0
Update documentation to establish the difference between backend time and backend response time
#1014 opened by sachanub - 1
Config option to increase or disable model load timeout
#1009 opened by svenkata9 - 1
Permission denied when loading model
#949 opened by akulk314 - 0
Invalid Response Headers Set for Non MME Inference Scenario (scikit learn container)
#1007 opened by Grassycup - 1
python version
#1003 opened by n0thing233 - 0
Validate MMS process during start
#1005 opened by nikhil-sk - 0
Model specific custom python package installation
#1004 opened by salvadiswar02 - 0
- 1
- 0
Is it supporting Apple M1 Chip ?
#1000 opened by jaiswalvineet - 0
mms-gpu - cuda error - No kernel image is available for execution on the device
#998 opened by kaushal-idx - 0
how to change url format
#997 opened by xuweidongkobe - 4
- 1
command not found: multi-model-server
#993 opened by nimafo - 2
Process run on a single CPU core
#960 opened by ngoanpv - 0
log4j2 metrics JsonLayout / QLogLayout logger broken
#994 opened by lxning - 0
`com.amazonaws.ml.mms.metrics.MetricCollector - java.io.IOException: Broken pipe` and `error while loading shared libraries: libpython3.7m.so.1.0`
#992 opened by llorenzo-matterport - 3
[Q] GPU support
#938 opened by oonisim - 1
The example in /examples/mxnet_vision/ does not work
#991 opened by kylehh - 4
Default log level change from 1.1.4 -> 1.1.6
#987 opened by kastman - 0
Upgrade log4j version to 2.17.1
#989 opened by glc-froussel - 0
Remove ineffective log4j 1 references from code
#988 opened by nikhil-sk - 1
memory utilization increment after every request, worker died, memory issue
#974 opened by n0thing233 - 0
- 0
Custom plug-in
#976 opened by HuryanKliashchouTR - 0
Change plugins logic
#975 opened by UsernameJava - 3
Allow custom HTTP status in mms.service.Service
#961 opened by jcsaaddupuy - 1
For multithreaded inferencing on GPU machine, with preload_model=True and default_workers_per_model=2 getting the following error
#959 opened by msameedkhan - 1
- 0
big file request will not release memory
#967 opened by yangjian1218 - 0
how to build several apps and functions
#966 opened by yangjian1218 - 0
- 0
- 0
Dependencies not installed in docker.
#958 opened by bahar3474 - 1
define inference proto
#951 opened by lxning - 0
s3 path in tutorial is not available
#956 opened by HahTK - 0
bad links in the Model Zoo page blocking our pipelines
#955 opened by junpuf - 0
Slack link not working.
#954 opened by azmaktr - 0
tensor_gpu-inl.h:35: Check failed: e == cudaSuccess: CUDA: initialization error
#953 opened by wdh234 - 0
Detect video
#947 opened by wdh234 - 1
Is it possible to implement multi-stage inferences?
#939 opened by bedilbek - 6
ONNX to .MAR converter test case fails.
#936 opened by quantum-fusion