Material for the talk at ETIM 2024
The slides can be found here.
- to set up the local Kubernetes cluster that is required to run the demo, follow the instructions on https://github.com/grothesk/kserve-on-minikube or https://github.com/deepshore/kserve-locally/tree/docker-desktop - or use a regular cluster and adapt the Makefile accordingly.
- ensure the MinIO client has been installed: https://min.io/docs/minio/linux/reference/minio-mc.html
- install the required Python packages via
pipenv
:
pipenv sync
- download the model files like this:
make download-model-files
For running the demo, perform the following steps:
- Package model files (MAR) so that
torchserve
as model serving runtime can work with the model:
make create-mar
- Create the required directory structure on the object storage and upload the MAR file and the
config.properties
:
make create-model-buckets upload-model-files
- Create an
InferenceService
resource including references to the location of the model files and the model serving runtime to be used:
make deploy-inference-service
- Prepare the image of the good boy and classify it via HTTP requests:
IMAGE=dog make convert-jpg-to-json classify-picture-of-cute-animal
- Do the same for another cute animal:
IMAGE=shark make convert-jpg-to-json classify-picture-of-cute-animal