Best Open-source Model Serving & Inference Platforms tools

Explore curated open-source tools in the Model Serving & Inference Platforms category. Compare technologies, see alternatives, and find the right solution for your workflow.

10+ projects · Page 1 of 1

SGLang logo

SGLang

High‑performance serving framework for LLMs and vision‑language models.

Stars
25,010
License
Apache-2.0
Last commit
1 hour ago
PythonActive
Triton Inference Server logo

Triton Inference Server

Unified AI model serving across clouds, edge, and GPUs

Stars
10,466
License
BSD-3-Clause
Last commit
3 hours ago
PythonActive
BentoML logo

BentoML

Unified Python framework for building high‑performance AI inference APIs

Stars
8,542
License
Apache-2.0
Last commit
11 hours ago
PythonActive
KServe logo

KServe

Unified AI inference platform for generative and predictive workloads on Kubernetes

Stars
5,262
License
Apache-2.0
Last commit
1 day ago
GoActive
NanoFlow logo

NanoFlow

High‑throughput LLM serving with intra‑device parallelism and asynchronous CPU scheduling

Stars
951
License
Last commit
4 months ago
Jupyter NotebookStable
FEDML logo

FEDML

Unified ML library for scalable training, serving, and federated learning.

Stars
4,021
License
Apache-2.0
Last commit
4 months ago
PythonStable