Pinned Repositories
kubeflow
Machine Learning Toolkit for Kubernetes
dcgm-exporter
NVIDIA GPU metrics exporter for Prometheus leveraging DCGM
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
citicern
A platform for crowd-sourcing data for scientific research.
Elixir_SymWorld
This is a simulation project for an Elixir Actor World
profile
Profile website source repository
seldon-core
Machine Learning Deployment for Kubernetes
MLServer
An inference server for your machine learning models, including support for multiple frameworks, multi-model serving and more
seldon-core
An MLOps framework to package, deploy, monitor and manage thousands of production machine learning models
python_backend
Triton backend that enables pre-process, post-processing and other logic to be implemented in Python.
SachinVarghese's Repositories
SachinVarghese/seldon-core
Machine Learning Deployment for Kubernetes
SachinVarghese/profile
Profile website source repository
SachinVarghese/SachinVarghese
SachinVarghese/telma
Toolkit Evaluator for Language Model Agents
SachinVarghese/alibi-detect
Algorithms for outlier, adversarial and drift detection
SachinVarghese/pgamber
Data observability for postgreSQL using alibi-detect
SachinVarghese/aws-saa-code
SachinVarghese/beatoven-public-api
SachinVarghese/dcgm-exporter
NVIDIA GPU metrics exporter for Prometheus leveraging DCGM
SachinVarghese/edge-cloud-inference
SachinVarghese/etc3
etc3 is the controller that powers Iter8, the AI-driven Kubernetes experimentation platform. etc3 stands for Extensible Thin Controller with Composable CRDs.
SachinVarghese/foundation-model-benchmarking-tool
Foundation model benchmarking tool. Run any model on any AWS platform and benchmark for performance across instance type and serving stack options.
SachinVarghese/genai-landscape
SachinVarghese/genar8t
SachinVarghese/kserve
Serverless Inferencing on Kubernetes
SachinVarghese/langchain
⚡ Building applications with LLMs through composability ⚡
SachinVarghese/manifests
A repository for Kustomize manifests
SachinVarghese/ml-prediction-schema
SachinVarghese/MLServer
An inference server for your machine learning models, including support for multiple frameworks, multi-model serving and more
SachinVarghese/nuclear
SachinVarghese/pinot-docs
Apache Pinot Documentation
SachinVarghese/python_backend
Triton backend that enables pre-process, post-processing and other logic to be implemented in Python.
SachinVarghese/seldon-deploy-sdk
SachinVarghese/seldon-gitops
SachinVarghese/sns-infinity
SachinVarghese/story-score
Automatic background scores from text inputs
SachinVarghese/tempo
MLOps Python Library
SachinVarghese/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
SachinVarghese/text-generation-inference
Large Language Model Text Generation Inference
SachinVarghese/virtual-web-museum