inference-api
There are 85 repositories under inference-api topic.
roboflow/inference
Turn any computer or edge device into a command center for your computer vision projects.
basetenlabs/truss
The simplest way to serve AI/ML models in production
quic/ai-hub-models
The Qualcomm® AI Hub Models are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
quic/ai-hub-apps
The Qualcomm® AI Hub apps are a collection of state-of-the-art machine learning models optimized for performance (latency, memory etc.) and ready to deploy on Qualcomm® devices.
Michael-OvO/Yolov7-Flask
A Beautiful Flask Web API for Yolov7 (and custom) models
mustafamerttunali/deep-learning-training-gui
Train and predict your model on pre-trained deep learning models through the GUI (web app). No more many parameters, no more data preprocessing.
pszemraj/textsum
CLI & Python API to easily summarize text-based files with transformers
BMW-InnovationLab/BMW-Classification-Training-GUI
This repository allows you to get started with training a State-of-the-art Deep Learning model with little to no configuration needed! You provide your labeled dataset and you can start the training right away. You can even test your model with our built-in Inference REST API. Training classification models with GluonCV has never been so easy.
inference-gateway/inference-gateway
An open-source, cloud-native, high-performance gateway unifying multiple LLM providers, from local solutions like Ollama to major cloud providers such as OpenAI, Groq, Cohere, Anthropic, Cloudflare and DeepSeek.
intelligencedev/eternal
Eternal is an experimental platform for machine learning models and workflows.
Kardbord/hfapigo
Unofficial (Golang) Go bindings for the Hugging Face Inference API
hupe1980/go-huggingface
🤗 Hugging Face Inference Client written in Go
BMW-InnovationLab/BMW-Classification-Inference-GPU-CPU
This is a repository for an image classification inference API using the Gluoncv framework. The inference REST API works on CPU/GPU. It's supported on Windows and Linux Operating systems. Models trained using our Gluoncv Classification training repository can be deployed in this API. Several models can be loaded and used at the same time.
Prismadic/magnet
the small distributed language model toolkit; fine-tune state-of-the-art LLMs anywhere, rapidly
TimMikeladze/huggingface
Typescript wrapper for the Hugging Face Inference API.
TommyLemon/CVAuto
👁 零代码零标注 CV AI 自动化测试工具 🚀 免除大量人工画框和打标签等,直接零代码快速自动化测试 CV 计算机视觉 AI 人工智能图像识别算法:行人检测、动植物分类、人脸识别、OCR 车牌识别、旋转校正、舞蹈姿态、抠图分割 等,还可一键 下载测试报告、导出训练和测试数据集
stephanj/Llama3JavaChatCompletionService
Llama3.java Inference engine with OpenAI Chat Completion REST API/
decisionfacts/semantic-ai
An open source framework for Retrieval-Augmented System (RAG) uses semantic search helps to retrieve the expected results and generate human readable conversational response with the help of LLM (Large Language Model).
yas-sim/openvino-ep-enabled-onnxruntime
Describing How to Enable OpenVINO Execution Provider for ONNX Runtime
BorjaOteroFerreira/IALab-Suite
Tool for test diferents large language models without code.
RageAgainstThePixel/com.rest.huggingface
A Non-Official HuggingFace Rest Client for Unity (UPM)
SaeedNajafi/infer-pytorch-pyspark
Coupling PySpark with PyTorch Models
kyryl-opens-ml/ml-in-production-practice
Practice for Machine Learning in Production course
jparkerweb/bedrock-proxy-endpoint
🔀 Bedrock Proxy Endpoint ⇢ Spin up your own custom OpenAI API server endpoint for easy AWS Bedrock inference (using standard baseUrl, and apiKey params)
shivamMg/stable-diffusion-on-azureml
REST APIs for StableDiffusion. Inferencing support on AzureML
antoninoLorenzo/Ollama-on-Colab-with-ngrok
Notebook to run Ollama on Google Colab
BMW-InnovationLab/BMW-TensorFlow-Training-GUI
This repository allows you to get started with a gui based training a State-of-the-art Deep Learning model with little to no configuration needed! NoCode training with TensorFlow has never been so easy.
gmkung/Cheemera
A Node.js backend that exposes a Typescript implementation of the deCheem inference engine for LLMs/ChatGPT.
LM4eu/goinfer
Local LLM proxy, DevOps friendly
pandruszkow/whisper-inference-server
A networked inference server for Whisper speech recognition
PromptOn/prompton
Chat prompt template evaluation and inference monitoring
ingyuseong/rabbitmq-inference
A message queue based server architecture to asynchronously handle resource-intensive tasks (e.g., ML inference)
pchandrasekaran1595/Computer-Vision-API
Computer VIsion API built using FastAPI and pretrained models converted to ONNX format
yas-sim/OpenVINO_Asynchronous_API_Performance_Demo
This project demonstrates the high performance of OpenVINO asynchronous inference API
YAV-AI/NodeJS-Stable-Diffusion-XL-Base-1.0-Hugging-Face-Inference-API
A simple node.js example that generates an image using StableDiffusion via Hugging Face Inference API.